var/home/core/zuul-output/0000755000175000017500000000000015111302303014512 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015111313410015457 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000004761007315111313402017674 0ustar rootrootNov 25 10:32:08 crc systemd[1]: Starting Kubernetes Kubelet... Nov 25 10:32:08 crc restorecon[4678]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 10:32:08 crc restorecon[4678]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 10:32:09 crc restorecon[4678]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 10:32:09 crc restorecon[4678]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Nov 25 10:32:09 crc kubenswrapper[4821]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 25 10:32:09 crc kubenswrapper[4821]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Nov 25 10:32:09 crc kubenswrapper[4821]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 25 10:32:09 crc kubenswrapper[4821]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 25 10:32:09 crc kubenswrapper[4821]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Nov 25 10:32:09 crc kubenswrapper[4821]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.902116 4821 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.910903 4821 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.910932 4821 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.910939 4821 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.910944 4821 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.910949 4821 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.910954 4821 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.910966 4821 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.910971 4821 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.910976 4821 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.910981 4821 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.910986 4821 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.910991 4821 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.910996 4821 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.911001 4821 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.911007 4821 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.911012 4821 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.911016 4821 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.911022 4821 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.911027 4821 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.911032 4821 feature_gate.go:330] unrecognized feature gate: Example Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.911037 4821 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.911042 4821 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.911048 4821 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.911053 4821 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.911058 4821 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.911063 4821 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.911068 4821 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.911072 4821 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.911077 4821 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.911085 4821 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.911091 4821 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.911096 4821 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.911101 4821 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.911107 4821 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.911112 4821 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.911117 4821 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.911123 4821 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.911130 4821 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.911134 4821 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.911140 4821 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.911146 4821 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.911151 4821 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.911172 4821 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.911178 4821 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.911184 4821 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.911189 4821 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.911194 4821 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.911199 4821 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.911204 4821 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.911209 4821 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.911214 4821 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.911219 4821 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.911224 4821 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.911232 4821 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.911239 4821 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.911246 4821 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.911252 4821 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.911258 4821 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.911263 4821 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.911268 4821 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.911273 4821 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.911278 4821 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.911283 4821 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.911289 4821 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.911294 4821 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.911303 4821 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.911309 4821 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.911315 4821 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.911321 4821 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.911326 4821 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.911331 4821 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911448 4821 flags.go:64] FLAG: --address="0.0.0.0" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911460 4821 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911470 4821 flags.go:64] FLAG: --anonymous-auth="true" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911477 4821 flags.go:64] FLAG: --application-metrics-count-limit="100" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911520 4821 flags.go:64] FLAG: --authentication-token-webhook="false" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911527 4821 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911535 4821 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911542 4821 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911548 4821 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911554 4821 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911560 4821 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911567 4821 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911573 4821 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911579 4821 flags.go:64] FLAG: --cgroup-root="" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911585 4821 flags.go:64] FLAG: --cgroups-per-qos="true" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911591 4821 flags.go:64] FLAG: --client-ca-file="" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911596 4821 flags.go:64] FLAG: --cloud-config="" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911602 4821 flags.go:64] FLAG: --cloud-provider="" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911607 4821 flags.go:64] FLAG: --cluster-dns="[]" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911616 4821 flags.go:64] FLAG: --cluster-domain="" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911621 4821 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911627 4821 flags.go:64] FLAG: --config-dir="" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911633 4821 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911640 4821 flags.go:64] FLAG: --container-log-max-files="5" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911649 4821 flags.go:64] FLAG: --container-log-max-size="10Mi" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911656 4821 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911662 4821 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911668 4821 flags.go:64] FLAG: --containerd-namespace="k8s.io" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911674 4821 flags.go:64] FLAG: --contention-profiling="false" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911680 4821 flags.go:64] FLAG: --cpu-cfs-quota="true" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911687 4821 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911693 4821 flags.go:64] FLAG: --cpu-manager-policy="none" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911700 4821 flags.go:64] FLAG: --cpu-manager-policy-options="" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911713 4821 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911719 4821 flags.go:64] FLAG: --enable-controller-attach-detach="true" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911725 4821 flags.go:64] FLAG: --enable-debugging-handlers="true" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911731 4821 flags.go:64] FLAG: --enable-load-reader="false" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911737 4821 flags.go:64] FLAG: --enable-server="true" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911743 4821 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911751 4821 flags.go:64] FLAG: --event-burst="100" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911757 4821 flags.go:64] FLAG: --event-qps="50" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911763 4821 flags.go:64] FLAG: --event-storage-age-limit="default=0" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911769 4821 flags.go:64] FLAG: --event-storage-event-limit="default=0" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911774 4821 flags.go:64] FLAG: --eviction-hard="" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911781 4821 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911787 4821 flags.go:64] FLAG: --eviction-minimum-reclaim="" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911793 4821 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911799 4821 flags.go:64] FLAG: --eviction-soft="" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911805 4821 flags.go:64] FLAG: --eviction-soft-grace-period="" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911810 4821 flags.go:64] FLAG: --exit-on-lock-contention="false" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911816 4821 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911822 4821 flags.go:64] FLAG: --experimental-mounter-path="" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911828 4821 flags.go:64] FLAG: --fail-cgroupv1="false" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911834 4821 flags.go:64] FLAG: --fail-swap-on="true" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911839 4821 flags.go:64] FLAG: --feature-gates="" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911846 4821 flags.go:64] FLAG: --file-check-frequency="20s" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911853 4821 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911859 4821 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911866 4821 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911872 4821 flags.go:64] FLAG: --healthz-port="10248" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911878 4821 flags.go:64] FLAG: --help="false" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911884 4821 flags.go:64] FLAG: --hostname-override="" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911889 4821 flags.go:64] FLAG: --housekeeping-interval="10s" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911896 4821 flags.go:64] FLAG: --http-check-frequency="20s" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911902 4821 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911908 4821 flags.go:64] FLAG: --image-credential-provider-config="" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911913 4821 flags.go:64] FLAG: --image-gc-high-threshold="85" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911919 4821 flags.go:64] FLAG: --image-gc-low-threshold="80" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911925 4821 flags.go:64] FLAG: --image-service-endpoint="" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911931 4821 flags.go:64] FLAG: --kernel-memcg-notification="false" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911937 4821 flags.go:64] FLAG: --kube-api-burst="100" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911942 4821 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911949 4821 flags.go:64] FLAG: --kube-api-qps="50" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911955 4821 flags.go:64] FLAG: --kube-reserved="" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911961 4821 flags.go:64] FLAG: --kube-reserved-cgroup="" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911966 4821 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911973 4821 flags.go:64] FLAG: --kubelet-cgroups="" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911978 4821 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911985 4821 flags.go:64] FLAG: --lock-file="" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911990 4821 flags.go:64] FLAG: --log-cadvisor-usage="false" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.911996 4821 flags.go:64] FLAG: --log-flush-frequency="5s" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.912001 4821 flags.go:64] FLAG: --log-json-info-buffer-size="0" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.912010 4821 flags.go:64] FLAG: --log-json-split-stream="false" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.912016 4821 flags.go:64] FLAG: --log-text-info-buffer-size="0" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.912022 4821 flags.go:64] FLAG: --log-text-split-stream="false" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.912027 4821 flags.go:64] FLAG: --logging-format="text" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.912033 4821 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.912040 4821 flags.go:64] FLAG: --make-iptables-util-chains="true" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.912045 4821 flags.go:64] FLAG: --manifest-url="" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.912051 4821 flags.go:64] FLAG: --manifest-url-header="" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.912059 4821 flags.go:64] FLAG: --max-housekeeping-interval="15s" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.912064 4821 flags.go:64] FLAG: --max-open-files="1000000" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.912071 4821 flags.go:64] FLAG: --max-pods="110" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.912077 4821 flags.go:64] FLAG: --maximum-dead-containers="-1" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.912083 4821 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.912089 4821 flags.go:64] FLAG: --memory-manager-policy="None" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.912094 4821 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.912101 4821 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.912107 4821 flags.go:64] FLAG: --node-ip="192.168.126.11" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.912113 4821 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.912127 4821 flags.go:64] FLAG: --node-status-max-images="50" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.912132 4821 flags.go:64] FLAG: --node-status-update-frequency="10s" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.912138 4821 flags.go:64] FLAG: --oom-score-adj="-999" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.912144 4821 flags.go:64] FLAG: --pod-cidr="" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.912150 4821 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.912175 4821 flags.go:64] FLAG: --pod-manifest-path="" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.912182 4821 flags.go:64] FLAG: --pod-max-pids="-1" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.912188 4821 flags.go:64] FLAG: --pods-per-core="0" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.912194 4821 flags.go:64] FLAG: --port="10250" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.912200 4821 flags.go:64] FLAG: --protect-kernel-defaults="false" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.912205 4821 flags.go:64] FLAG: --provider-id="" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.912211 4821 flags.go:64] FLAG: --qos-reserved="" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.912216 4821 flags.go:64] FLAG: --read-only-port="10255" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.912223 4821 flags.go:64] FLAG: --register-node="true" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.912229 4821 flags.go:64] FLAG: --register-schedulable="true" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.912234 4821 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.912245 4821 flags.go:64] FLAG: --registry-burst="10" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.912250 4821 flags.go:64] FLAG: --registry-qps="5" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.912256 4821 flags.go:64] FLAG: --reserved-cpus="" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.912262 4821 flags.go:64] FLAG: --reserved-memory="" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.912269 4821 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.912275 4821 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.912281 4821 flags.go:64] FLAG: --rotate-certificates="false" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.912287 4821 flags.go:64] FLAG: --rotate-server-certificates="false" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.912293 4821 flags.go:64] FLAG: --runonce="false" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.912298 4821 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.912304 4821 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.912310 4821 flags.go:64] FLAG: --seccomp-default="false" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.912316 4821 flags.go:64] FLAG: --serialize-image-pulls="true" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.912322 4821 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.912328 4821 flags.go:64] FLAG: --storage-driver-db="cadvisor" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.912334 4821 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.912340 4821 flags.go:64] FLAG: --storage-driver-password="root" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.912346 4821 flags.go:64] FLAG: --storage-driver-secure="false" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.912351 4821 flags.go:64] FLAG: --storage-driver-table="stats" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.912357 4821 flags.go:64] FLAG: --storage-driver-user="root" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.912363 4821 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.912369 4821 flags.go:64] FLAG: --sync-frequency="1m0s" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.912375 4821 flags.go:64] FLAG: --system-cgroups="" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.912380 4821 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.912392 4821 flags.go:64] FLAG: --system-reserved-cgroup="" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.912397 4821 flags.go:64] FLAG: --tls-cert-file="" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.912403 4821 flags.go:64] FLAG: --tls-cipher-suites="[]" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.912410 4821 flags.go:64] FLAG: --tls-min-version="" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.912416 4821 flags.go:64] FLAG: --tls-private-key-file="" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.912422 4821 flags.go:64] FLAG: --topology-manager-policy="none" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.912428 4821 flags.go:64] FLAG: --topology-manager-policy-options="" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.912433 4821 flags.go:64] FLAG: --topology-manager-scope="container" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.912439 4821 flags.go:64] FLAG: --v="2" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.912447 4821 flags.go:64] FLAG: --version="false" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.912454 4821 flags.go:64] FLAG: --vmodule="" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.912462 4821 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.912468 4821 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.912603 4821 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.912611 4821 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.912618 4821 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.912624 4821 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.912629 4821 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.912635 4821 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.912641 4821 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.912647 4821 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.912653 4821 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.912659 4821 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.912664 4821 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.912669 4821 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.912674 4821 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.912679 4821 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.912686 4821 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.912692 4821 feature_gate.go:330] unrecognized feature gate: Example Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.912698 4821 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.912703 4821 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.912709 4821 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.912713 4821 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.912719 4821 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.912723 4821 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.912728 4821 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.912734 4821 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.912739 4821 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.912744 4821 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.912749 4821 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.912754 4821 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.912759 4821 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.912764 4821 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.912769 4821 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.912774 4821 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.912779 4821 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.912784 4821 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.912789 4821 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.912793 4821 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.912798 4821 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.912803 4821 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.912808 4821 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.912813 4821 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.912824 4821 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.912828 4821 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.912833 4821 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.912838 4821 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.912843 4821 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.912848 4821 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.912852 4821 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.912857 4821 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.912862 4821 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.912867 4821 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.912872 4821 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.912876 4821 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.912881 4821 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.912886 4821 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.912890 4821 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.912896 4821 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.912900 4821 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.912905 4821 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.912910 4821 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.912922 4821 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.912929 4821 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.912934 4821 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.912940 4821 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.912945 4821 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.912949 4821 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.912955 4821 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.912959 4821 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.912964 4821 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.912969 4821 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.912974 4821 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.912979 4821 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.912996 4821 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.922748 4821 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.922793 4821 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.922870 4821 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.922884 4821 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.922890 4821 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.922895 4821 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.922899 4821 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.922904 4821 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.922908 4821 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.922913 4821 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.922918 4821 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.922923 4821 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.922928 4821 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.922933 4821 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.922937 4821 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.922940 4821 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.922944 4821 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.922948 4821 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.922951 4821 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.922954 4821 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.922958 4821 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.922961 4821 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.922965 4821 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.922969 4821 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.922973 4821 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.922977 4821 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.922980 4821 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.922983 4821 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.922987 4821 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.922992 4821 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.922996 4821 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.922999 4821 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923003 4821 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923006 4821 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923011 4821 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923016 4821 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923024 4821 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923027 4821 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923031 4821 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923034 4821 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923039 4821 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923042 4821 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923045 4821 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923049 4821 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923053 4821 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923056 4821 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923060 4821 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923064 4821 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923067 4821 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923071 4821 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923074 4821 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923078 4821 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923081 4821 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923085 4821 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923088 4821 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923092 4821 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923095 4821 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923099 4821 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923102 4821 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923105 4821 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923109 4821 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923113 4821 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923116 4821 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923120 4821 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923123 4821 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923127 4821 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923130 4821 feature_gate.go:330] unrecognized feature gate: Example Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923134 4821 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923137 4821 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923142 4821 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923146 4821 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923150 4821 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923155 4821 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.923178 4821 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923296 4821 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923303 4821 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923308 4821 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923313 4821 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923317 4821 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923321 4821 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923325 4821 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923329 4821 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923333 4821 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923337 4821 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923340 4821 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923344 4821 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923348 4821 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923352 4821 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923355 4821 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923359 4821 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923362 4821 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923366 4821 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923369 4821 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923373 4821 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923377 4821 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923380 4821 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923384 4821 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923387 4821 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923390 4821 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923394 4821 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923397 4821 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923401 4821 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923405 4821 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923410 4821 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923414 4821 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923418 4821 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923423 4821 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923427 4821 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923431 4821 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923436 4821 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923439 4821 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923443 4821 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923447 4821 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923450 4821 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923454 4821 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923457 4821 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923460 4821 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923464 4821 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923467 4821 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923508 4821 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923513 4821 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923517 4821 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923521 4821 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923525 4821 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923528 4821 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923532 4821 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923535 4821 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923539 4821 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923542 4821 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923546 4821 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923549 4821 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923553 4821 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923557 4821 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923560 4821 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923563 4821 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923567 4821 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923570 4821 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923574 4821 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923577 4821 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923582 4821 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923586 4821 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923592 4821 feature_gate.go:330] unrecognized feature gate: Example Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923596 4821 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923599 4821 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 25 10:32:09 crc kubenswrapper[4821]: W1125 10:32:09.923604 4821 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.923610 4821 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.924815 4821 server.go:940] "Client rotation is on, will bootstrap in background" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.928892 4821 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.928974 4821 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.930290 4821 server.go:997] "Starting client certificate rotation" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.930315 4821 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.930536 4821 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-30 09:42:07.69318108 +0000 UTC Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.930634 4821 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 839h9m57.762550101s for next certificate rotation Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.970141 4821 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.972398 4821 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 25 10:32:09 crc kubenswrapper[4821]: I1125 10:32:09.991597 4821 log.go:25] "Validated CRI v1 runtime API" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.022840 4821 log.go:25] "Validated CRI v1 image API" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.024634 4821 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.029429 4821 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-11-25-10-27-17-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.029455 4821 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.042184 4821 manager.go:217] Machine: {Timestamp:2025-11-25 10:32:10.040600705 +0000 UTC m=+0.576920572 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654124544 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:ea86e80f-41b7-4190-ad7e-d8c549d535de BootID:a5b2b070-39b9-412c-90f1-c7ed2e9e0950 Filesystems:[{Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108169 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827060224 Type:vfs Inodes:4108169 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:64:f0:a1 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:64:f0:a1 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:29:39:a2 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:67:1e:14 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:fb:f0:8a Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:3a:c3:42 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:d2:02:12:af:71:fc Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:6e:4d:34:f6:8b:64 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654124544 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.042368 4821 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.042476 4821 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.043999 4821 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.044254 4821 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.044295 4821 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.044528 4821 topology_manager.go:138] "Creating topology manager with none policy" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.044541 4821 container_manager_linux.go:303] "Creating device plugin manager" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.045054 4821 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.045086 4821 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.045930 4821 state_mem.go:36] "Initialized new in-memory state store" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.046013 4821 server.go:1245] "Using root directory" path="/var/lib/kubelet" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.049411 4821 kubelet.go:418] "Attempting to sync node with API server" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.049433 4821 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.049462 4821 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.049479 4821 kubelet.go:324] "Adding apiserver pod source" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.049490 4821 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.053306 4821 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.054274 4821 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Nov 25 10:32:10 crc kubenswrapper[4821]: W1125 10:32:10.055400 4821 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.129.56.252:6443: connect: connection refused Nov 25 10:32:10 crc kubenswrapper[4821]: E1125 10:32:10.055468 4821 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.129.56.252:6443: connect: connection refused" logger="UnhandledError" Nov 25 10:32:10 crc kubenswrapper[4821]: W1125 10:32:10.055487 4821 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.129.56.252:6443: connect: connection refused Nov 25 10:32:10 crc kubenswrapper[4821]: E1125 10:32:10.055578 4821 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.129.56.252:6443: connect: connection refused" logger="UnhandledError" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.056556 4821 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.059607 4821 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.059651 4821 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.059661 4821 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.059674 4821 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.059689 4821 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.059738 4821 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.059751 4821 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.059768 4821 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.059779 4821 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.059789 4821 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.059804 4821 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.059813 4821 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.061647 4821 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.062118 4821 server.go:1280] "Started kubelet" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.062344 4821 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.062305 4821 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.063097 4821 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Nov 25 10:32:10 crc systemd[1]: Started Kubernetes Kubelet. Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.066578 4821 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.252:6443: connect: connection refused Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.067981 4821 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.068030 4821 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.068468 4821 volume_manager.go:287] "The desired_state_of_world populator starts" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.068507 4821 volume_manager.go:289] "Starting Kubelet Volume Manager" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.068644 4821 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.068612 4821 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-07 22:08:42.889855155 +0000 UTC Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.068668 4821 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 299h36m32.82118971s for next certificate rotation Nov 25 10:32:10 crc kubenswrapper[4821]: E1125 10:32:10.068501 4821 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 25 10:32:10 crc kubenswrapper[4821]: W1125 10:32:10.072971 4821 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.129.56.252:6443: connect: connection refused Nov 25 10:32:10 crc kubenswrapper[4821]: E1125 10:32:10.073057 4821 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.129.56.252:6443: connect: connection refused" logger="UnhandledError" Nov 25 10:32:10 crc kubenswrapper[4821]: E1125 10:32:10.073045 4821 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.252:6443: connect: connection refused" interval="200ms" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.073471 4821 factory.go:153] Registering CRI-O factory Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.073503 4821 factory.go:221] Registration of the crio container factory successfully Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.073587 4821 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.073603 4821 factory.go:55] Registering systemd factory Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.073614 4821 factory.go:221] Registration of the systemd container factory successfully Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.073637 4821 factory.go:103] Registering Raw factory Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.073652 4821 manager.go:1196] Started watching for new ooms in manager Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.073705 4821 server.go:460] "Adding debug handlers to kubelet server" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.074554 4821 manager.go:319] Starting recovery of all containers Nov 25 10:32:10 crc kubenswrapper[4821]: E1125 10:32:10.078286 4821 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.129.56.252:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187b3958356f76c3 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-25 10:32:10.062083779 +0000 UTC m=+0.598403626,LastTimestamp:2025-11-25 10:32:10.062083779 +0000 UTC m=+0.598403626,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.085276 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.085379 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.085394 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.085514 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.085542 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.085557 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.085664 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.085681 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.085702 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.085715 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.085731 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.085741 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.085754 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.085773 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.085789 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.085800 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.085812 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.086801 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.086903 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.086926 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.086953 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.086966 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.086981 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.086999 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.087014 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.087058 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.087105 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.087125 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.087136 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.087150 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.087175 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.087185 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.087198 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.087209 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.087235 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.087244 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.087255 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.087277 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.087315 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.087639 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.087668 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.087711 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.088589 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.088649 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.088663 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.088676 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.088687 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.088698 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.088708 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.088720 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.088731 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.088741 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.088758 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.088769 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.088781 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.088794 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.088804 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.088814 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.088824 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.088834 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.088845 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.088854 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.088869 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.088879 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.088889 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.088898 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.088909 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.088917 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.088927 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.088940 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.088978 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.088994 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089008 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089022 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089031 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089042 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089051 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089062 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089072 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089096 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089105 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089116 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089128 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089144 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089154 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089178 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089187 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089196 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089206 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089216 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089226 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089238 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089247 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089258 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089271 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089281 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089291 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089301 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089315 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089329 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089348 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089359 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089368 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089377 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089392 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089403 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089415 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089425 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089436 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089446 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089457 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089467 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089476 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089486 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089498 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089507 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089516 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089524 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089532 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089542 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089553 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089562 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089572 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089580 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089588 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089597 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089605 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089613 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089622 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089631 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089641 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089650 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089660 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089668 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089681 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089690 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089699 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089709 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089718 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089727 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089736 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089744 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089752 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089760 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089768 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089776 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089786 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089802 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.089811 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.091380 4821 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.091410 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.091427 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.091437 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.091446 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.091457 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.091468 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.091480 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.091490 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.091498 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.091507 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.091516 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.091523 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.091534 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.091544 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.091557 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.091568 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.091578 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.091588 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.091597 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.091608 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.091618 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.091628 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.091641 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.091653 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.091664 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.091677 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.091688 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.091697 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.091707 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.091716 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.091727 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.091741 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.091753 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.091767 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.091779 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.091789 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.091799 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.091811 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.091823 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.091835 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.091858 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.091872 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.091885 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.091909 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.091923 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.091932 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.091942 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.091951 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.091961 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.091971 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.091981 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.091992 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.092002 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.092012 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.092022 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.092033 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.092043 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.092052 4821 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.092062 4821 reconstruct.go:97] "Volume reconstruction finished" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.092069 4821 reconciler.go:26] "Reconciler: start to sync state" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.095268 4821 manager.go:324] Recovery completed Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.105402 4821 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.106747 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.106777 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.106786 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.107440 4821 cpu_manager.go:225] "Starting CPU manager" policy="none" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.107453 4821 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.107473 4821 state_mem.go:36] "Initialized new in-memory state store" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.111142 4821 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.112954 4821 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.112996 4821 status_manager.go:217] "Starting to sync pod status with apiserver" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.113032 4821 kubelet.go:2335] "Starting kubelet main sync loop" Nov 25 10:32:10 crc kubenswrapper[4821]: E1125 10:32:10.113080 4821 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Nov 25 10:32:10 crc kubenswrapper[4821]: W1125 10:32:10.113864 4821 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.129.56.252:6443: connect: connection refused Nov 25 10:32:10 crc kubenswrapper[4821]: E1125 10:32:10.114012 4821 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.129.56.252:6443: connect: connection refused" logger="UnhandledError" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.126004 4821 policy_none.go:49] "None policy: Start" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.126819 4821 memory_manager.go:170] "Starting memorymanager" policy="None" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.126851 4821 state_mem.go:35] "Initializing new in-memory state store" Nov 25 10:32:10 crc kubenswrapper[4821]: E1125 10:32:10.169144 4821 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.181415 4821 manager.go:334] "Starting Device Plugin manager" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.181474 4821 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.181486 4821 server.go:79] "Starting device plugin registration server" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.181907 4821 eviction_manager.go:189] "Eviction manager: starting control loop" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.181923 4821 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.182149 4821 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.182236 4821 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.182243 4821 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Nov 25 10:32:10 crc kubenswrapper[4821]: E1125 10:32:10.188899 4821 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.213583 4821 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.213675 4821 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.214809 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.214839 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.214847 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.214957 4821 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.215270 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.215320 4821 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.215641 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.215688 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.215699 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.215786 4821 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.215917 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.215945 4821 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.216198 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.216235 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.216246 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.216437 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.216457 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.216468 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.216594 4821 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.216608 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.216638 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.216648 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.216714 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.216744 4821 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.217940 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.217959 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.217967 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.218084 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.218111 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.218120 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.218223 4821 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.218331 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.218370 4821 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.218745 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.218764 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.218772 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.218942 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.218964 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.218975 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.219176 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.219200 4821 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.219888 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.219916 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.219955 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:10 crc kubenswrapper[4821]: E1125 10:32:10.273958 4821 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.252:6443: connect: connection refused" interval="400ms" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.283136 4821 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.284573 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.284605 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.284614 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.284635 4821 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 10:32:10 crc kubenswrapper[4821]: E1125 10:32:10.285022 4821 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.252:6443: connect: connection refused" node="crc" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.294214 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.294300 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.294349 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.294375 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.294417 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.294459 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.294487 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.294513 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.294534 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.294558 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.294581 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.294632 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.294724 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.294800 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.294859 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.395898 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.395946 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.395981 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.395999 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.396015 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.396030 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.396044 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.396065 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.396101 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.396071 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.396132 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.396134 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.396147 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.396178 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.396185 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.396193 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.396192 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.396235 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.396208 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.396221 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.396273 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.396255 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.396227 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.396293 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.396310 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.396314 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.396353 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.396357 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.396382 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.396703 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.485325 4821 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.486434 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.486476 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.486488 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.486516 4821 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 10:32:10 crc kubenswrapper[4821]: E1125 10:32:10.486967 4821 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.252:6443: connect: connection refused" node="crc" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.535847 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.555581 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.560937 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:32:10 crc kubenswrapper[4821]: W1125 10:32:10.573097 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-3c0b8e027eded1ab26bc265ebed3d9a7ed5b35f56118a5b57c0fb87eb86d24e0 WatchSource:0}: Error finding container 3c0b8e027eded1ab26bc265ebed3d9a7ed5b35f56118a5b57c0fb87eb86d24e0: Status 404 returned error can't find the container with id 3c0b8e027eded1ab26bc265ebed3d9a7ed5b35f56118a5b57c0fb87eb86d24e0 Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.579256 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.583678 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 10:32:10 crc kubenswrapper[4821]: W1125 10:32:10.591348 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-85895f47bfa31dec186e0d7d59abd8593988c15f376f9127397c2d2fd06386b6 WatchSource:0}: Error finding container 85895f47bfa31dec186e0d7d59abd8593988c15f376f9127397c2d2fd06386b6: Status 404 returned error can't find the container with id 85895f47bfa31dec186e0d7d59abd8593988c15f376f9127397c2d2fd06386b6 Nov 25 10:32:10 crc kubenswrapper[4821]: W1125 10:32:10.603639 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-83b9ff6f77defb2c1e6a6540993f25401020eae361ff950c26f407fe96eb569c WatchSource:0}: Error finding container 83b9ff6f77defb2c1e6a6540993f25401020eae361ff950c26f407fe96eb569c: Status 404 returned error can't find the container with id 83b9ff6f77defb2c1e6a6540993f25401020eae361ff950c26f407fe96eb569c Nov 25 10:32:10 crc kubenswrapper[4821]: W1125 10:32:10.604893 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-17398d47cac9be8326b2f0e9e1d6d05febe8789f9c05091778df4b9c38c7a1e6 WatchSource:0}: Error finding container 17398d47cac9be8326b2f0e9e1d6d05febe8789f9c05091778df4b9c38c7a1e6: Status 404 returned error can't find the container with id 17398d47cac9be8326b2f0e9e1d6d05febe8789f9c05091778df4b9c38c7a1e6 Nov 25 10:32:10 crc kubenswrapper[4821]: W1125 10:32:10.607697 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-4154cb1406d22e26946f08a6c6d4fe9cf43855e92a67d213edb9689bacb379e7 WatchSource:0}: Error finding container 4154cb1406d22e26946f08a6c6d4fe9cf43855e92a67d213edb9689bacb379e7: Status 404 returned error can't find the container with id 4154cb1406d22e26946f08a6c6d4fe9cf43855e92a67d213edb9689bacb379e7 Nov 25 10:32:10 crc kubenswrapper[4821]: E1125 10:32:10.675934 4821 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.252:6443: connect: connection refused" interval="800ms" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.887362 4821 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.888340 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.888375 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.888388 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:10 crc kubenswrapper[4821]: I1125 10:32:10.888415 4821 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 10:32:10 crc kubenswrapper[4821]: E1125 10:32:10.889017 4821 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.252:6443: connect: connection refused" node="crc" Nov 25 10:32:10 crc kubenswrapper[4821]: W1125 10:32:10.914952 4821 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.129.56.252:6443: connect: connection refused Nov 25 10:32:10 crc kubenswrapper[4821]: E1125 10:32:10.915098 4821 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.129.56.252:6443: connect: connection refused" logger="UnhandledError" Nov 25 10:32:10 crc kubenswrapper[4821]: W1125 10:32:10.983565 4821 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.129.56.252:6443: connect: connection refused Nov 25 10:32:10 crc kubenswrapper[4821]: E1125 10:32:10.983653 4821 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.129.56.252:6443: connect: connection refused" logger="UnhandledError" Nov 25 10:32:11 crc kubenswrapper[4821]: I1125 10:32:11.067724 4821 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.252:6443: connect: connection refused Nov 25 10:32:11 crc kubenswrapper[4821]: I1125 10:32:11.116831 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"3c0b8e027eded1ab26bc265ebed3d9a7ed5b35f56118a5b57c0fb87eb86d24e0"} Nov 25 10:32:11 crc kubenswrapper[4821]: I1125 10:32:11.117629 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"17398d47cac9be8326b2f0e9e1d6d05febe8789f9c05091778df4b9c38c7a1e6"} Nov 25 10:32:11 crc kubenswrapper[4821]: I1125 10:32:11.118385 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"4154cb1406d22e26946f08a6c6d4fe9cf43855e92a67d213edb9689bacb379e7"} Nov 25 10:32:11 crc kubenswrapper[4821]: I1125 10:32:11.119428 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"83b9ff6f77defb2c1e6a6540993f25401020eae361ff950c26f407fe96eb569c"} Nov 25 10:32:11 crc kubenswrapper[4821]: I1125 10:32:11.120273 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"85895f47bfa31dec186e0d7d59abd8593988c15f376f9127397c2d2fd06386b6"} Nov 25 10:32:11 crc kubenswrapper[4821]: W1125 10:32:11.453965 4821 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.129.56.252:6443: connect: connection refused Nov 25 10:32:11 crc kubenswrapper[4821]: E1125 10:32:11.454371 4821 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.129.56.252:6443: connect: connection refused" logger="UnhandledError" Nov 25 10:32:11 crc kubenswrapper[4821]: E1125 10:32:11.477079 4821 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.252:6443: connect: connection refused" interval="1.6s" Nov 25 10:32:11 crc kubenswrapper[4821]: W1125 10:32:11.579046 4821 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.129.56.252:6443: connect: connection refused Nov 25 10:32:11 crc kubenswrapper[4821]: E1125 10:32:11.579128 4821 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.129.56.252:6443: connect: connection refused" logger="UnhandledError" Nov 25 10:32:11 crc kubenswrapper[4821]: I1125 10:32:11.690179 4821 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:32:11 crc kubenswrapper[4821]: I1125 10:32:11.692808 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:11 crc kubenswrapper[4821]: I1125 10:32:11.692865 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:11 crc kubenswrapper[4821]: I1125 10:32:11.692886 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:11 crc kubenswrapper[4821]: I1125 10:32:11.692924 4821 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 10:32:11 crc kubenswrapper[4821]: E1125 10:32:11.693782 4821 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.252:6443: connect: connection refused" node="crc" Nov 25 10:32:12 crc kubenswrapper[4821]: I1125 10:32:12.067797 4821 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.252:6443: connect: connection refused Nov 25 10:32:12 crc kubenswrapper[4821]: I1125 10:32:12.127464 4821 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3" exitCode=0 Nov 25 10:32:12 crc kubenswrapper[4821]: I1125 10:32:12.127573 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3"} Nov 25 10:32:12 crc kubenswrapper[4821]: I1125 10:32:12.127688 4821 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:32:12 crc kubenswrapper[4821]: I1125 10:32:12.129197 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:12 crc kubenswrapper[4821]: I1125 10:32:12.129251 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:12 crc kubenswrapper[4821]: I1125 10:32:12.129267 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:12 crc kubenswrapper[4821]: I1125 10:32:12.130638 4821 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12" exitCode=0 Nov 25 10:32:12 crc kubenswrapper[4821]: I1125 10:32:12.130730 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12"} Nov 25 10:32:12 crc kubenswrapper[4821]: I1125 10:32:12.130852 4821 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:32:12 crc kubenswrapper[4821]: I1125 10:32:12.132225 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:12 crc kubenswrapper[4821]: I1125 10:32:12.132258 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:12 crc kubenswrapper[4821]: I1125 10:32:12.132269 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:12 crc kubenswrapper[4821]: I1125 10:32:12.132398 4821 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:32:12 crc kubenswrapper[4821]: I1125 10:32:12.133222 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:12 crc kubenswrapper[4821]: I1125 10:32:12.133267 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:12 crc kubenswrapper[4821]: I1125 10:32:12.133281 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:12 crc kubenswrapper[4821]: I1125 10:32:12.134189 4821 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="733440f5c7493b5e1f4597ebb8f9d3d79507df0a1bc33393fb4534b5a2f2f077" exitCode=0 Nov 25 10:32:12 crc kubenswrapper[4821]: I1125 10:32:12.134271 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"733440f5c7493b5e1f4597ebb8f9d3d79507df0a1bc33393fb4534b5a2f2f077"} Nov 25 10:32:12 crc kubenswrapper[4821]: I1125 10:32:12.134324 4821 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:32:12 crc kubenswrapper[4821]: I1125 10:32:12.135345 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:12 crc kubenswrapper[4821]: I1125 10:32:12.135374 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:12 crc kubenswrapper[4821]: I1125 10:32:12.135385 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:12 crc kubenswrapper[4821]: I1125 10:32:12.137149 4821 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="5e48c0e3b58359c6c0a493d58f7d2aec6011747791a3c0d7788102e1fec18865" exitCode=0 Nov 25 10:32:12 crc kubenswrapper[4821]: I1125 10:32:12.137259 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"5e48c0e3b58359c6c0a493d58f7d2aec6011747791a3c0d7788102e1fec18865"} Nov 25 10:32:12 crc kubenswrapper[4821]: I1125 10:32:12.137269 4821 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:32:12 crc kubenswrapper[4821]: I1125 10:32:12.137903 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:12 crc kubenswrapper[4821]: I1125 10:32:12.137928 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:12 crc kubenswrapper[4821]: I1125 10:32:12.137937 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:12 crc kubenswrapper[4821]: I1125 10:32:12.144095 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"ad5c205e0eeecfd43cbdf0eb1a334e3e8faa5e9a9e8dd202c397d5cbb36e081e"} Nov 25 10:32:12 crc kubenswrapper[4821]: I1125 10:32:12.144145 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"0251519fd4262f2ea5c64959db49433afdafe36b7ed2404d660eeb7c365d2746"} Nov 25 10:32:12 crc kubenswrapper[4821]: I1125 10:32:12.144183 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"45d4b2d8db4d13392d9789e161fd4b8f6eef983b79128ee584d72e4c06565bc9"} Nov 25 10:32:12 crc kubenswrapper[4821]: I1125 10:32:12.144200 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"cf780f791771ea43a8c8176d02fdfb5204406e023207299369ae46ceb342baaf"} Nov 25 10:32:12 crc kubenswrapper[4821]: I1125 10:32:12.144226 4821 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:32:12 crc kubenswrapper[4821]: I1125 10:32:12.145923 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:12 crc kubenswrapper[4821]: I1125 10:32:12.145953 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:12 crc kubenswrapper[4821]: I1125 10:32:12.145965 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:12 crc kubenswrapper[4821]: I1125 10:32:12.199893 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 10:32:13 crc kubenswrapper[4821]: W1125 10:32:13.049997 4821 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.129.56.252:6443: connect: connection refused Nov 25 10:32:13 crc kubenswrapper[4821]: E1125 10:32:13.050115 4821 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.129.56.252:6443: connect: connection refused" logger="UnhandledError" Nov 25 10:32:13 crc kubenswrapper[4821]: I1125 10:32:13.067279 4821 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.252:6443: connect: connection refused Nov 25 10:32:13 crc kubenswrapper[4821]: E1125 10:32:13.067316 4821 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.129.56.252:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187b3958356f76c3 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-25 10:32:10.062083779 +0000 UTC m=+0.598403626,LastTimestamp:2025-11-25 10:32:10.062083779 +0000 UTC m=+0.598403626,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 25 10:32:13 crc kubenswrapper[4821]: E1125 10:32:13.078366 4821 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.252:6443: connect: connection refused" interval="3.2s" Nov 25 10:32:13 crc kubenswrapper[4821]: I1125 10:32:13.149836 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"269ced131bc0777dc6752070470c95b4bedfe4b4240940eabd769500ddef0404"} Nov 25 10:32:13 crc kubenswrapper[4821]: I1125 10:32:13.149884 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"67df7c609d5a5d658be4c3e873ec3052bd070ad9469f8ff44c29a9571e3e5d95"} Nov 25 10:32:13 crc kubenswrapper[4821]: I1125 10:32:13.149889 4821 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:32:13 crc kubenswrapper[4821]: I1125 10:32:13.149904 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"a68a57fa7fba8f351f264e427a50bed23f57775644e9c268aa7305efb6916794"} Nov 25 10:32:13 crc kubenswrapper[4821]: I1125 10:32:13.150566 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:13 crc kubenswrapper[4821]: I1125 10:32:13.150592 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:13 crc kubenswrapper[4821]: I1125 10:32:13.150601 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:13 crc kubenswrapper[4821]: I1125 10:32:13.152306 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"991252e5f56f128827a5054b8e228284bc0266166582cbaedaadc53ca2f5b401"} Nov 25 10:32:13 crc kubenswrapper[4821]: I1125 10:32:13.152333 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"69633332d03d2551ff4abb253d358de398eab18486b92db630ce06c0447fb9cf"} Nov 25 10:32:13 crc kubenswrapper[4821]: I1125 10:32:13.152348 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"9ccb78d9a51ac83ee208c2451685a2512de9deaf742a663bb8c892aa59b76162"} Nov 25 10:32:13 crc kubenswrapper[4821]: I1125 10:32:13.152359 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"a28d43531927e75c3745f100940f7f6bf79fd21026c1309194df8cd7a39361b1"} Nov 25 10:32:13 crc kubenswrapper[4821]: I1125 10:32:13.154216 4821 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751" exitCode=0 Nov 25 10:32:13 crc kubenswrapper[4821]: I1125 10:32:13.154245 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751"} Nov 25 10:32:13 crc kubenswrapper[4821]: I1125 10:32:13.154303 4821 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:32:13 crc kubenswrapper[4821]: I1125 10:32:13.155123 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:13 crc kubenswrapper[4821]: I1125 10:32:13.155144 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:13 crc kubenswrapper[4821]: I1125 10:32:13.155154 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:13 crc kubenswrapper[4821]: I1125 10:32:13.155703 4821 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:32:13 crc kubenswrapper[4821]: I1125 10:32:13.155726 4821 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:32:13 crc kubenswrapper[4821]: I1125 10:32:13.156213 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"e9525225d46e6f9225786ef8dd0ebe108d41efa26f6c94451c069aa5eb30da58"} Nov 25 10:32:13 crc kubenswrapper[4821]: I1125 10:32:13.156434 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:13 crc kubenswrapper[4821]: I1125 10:32:13.156454 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:13 crc kubenswrapper[4821]: I1125 10:32:13.156466 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:13 crc kubenswrapper[4821]: I1125 10:32:13.157538 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:13 crc kubenswrapper[4821]: I1125 10:32:13.157585 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:13 crc kubenswrapper[4821]: I1125 10:32:13.157605 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:13 crc kubenswrapper[4821]: I1125 10:32:13.294385 4821 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:32:13 crc kubenswrapper[4821]: I1125 10:32:13.295783 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:13 crc kubenswrapper[4821]: I1125 10:32:13.295819 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:13 crc kubenswrapper[4821]: I1125 10:32:13.295827 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:13 crc kubenswrapper[4821]: I1125 10:32:13.295850 4821 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 10:32:13 crc kubenswrapper[4821]: E1125 10:32:13.296486 4821 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.252:6443: connect: connection refused" node="crc" Nov 25 10:32:13 crc kubenswrapper[4821]: W1125 10:32:13.738417 4821 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.129.56.252:6443: connect: connection refused Nov 25 10:32:13 crc kubenswrapper[4821]: E1125 10:32:13.738504 4821 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.129.56.252:6443: connect: connection refused" logger="UnhandledError" Nov 25 10:32:14 crc kubenswrapper[4821]: I1125 10:32:14.160957 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"27826e155a9352b1e1866e83fdab683dbcd749cc04b6a4da2db25ecabe3d9a52"} Nov 25 10:32:14 crc kubenswrapper[4821]: I1125 10:32:14.161052 4821 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:32:14 crc kubenswrapper[4821]: I1125 10:32:14.161868 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:14 crc kubenswrapper[4821]: I1125 10:32:14.161898 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:14 crc kubenswrapper[4821]: I1125 10:32:14.161907 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:14 crc kubenswrapper[4821]: I1125 10:32:14.163011 4821 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29" exitCode=0 Nov 25 10:32:14 crc kubenswrapper[4821]: I1125 10:32:14.163077 4821 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:32:14 crc kubenswrapper[4821]: I1125 10:32:14.163095 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29"} Nov 25 10:32:14 crc kubenswrapper[4821]: I1125 10:32:14.163117 4821 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 10:32:14 crc kubenswrapper[4821]: I1125 10:32:14.163189 4821 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:32:14 crc kubenswrapper[4821]: I1125 10:32:14.163191 4821 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:32:14 crc kubenswrapper[4821]: I1125 10:32:14.163631 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:14 crc kubenswrapper[4821]: I1125 10:32:14.163658 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:14 crc kubenswrapper[4821]: I1125 10:32:14.163666 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:14 crc kubenswrapper[4821]: I1125 10:32:14.163848 4821 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:32:14 crc kubenswrapper[4821]: I1125 10:32:14.164217 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:14 crc kubenswrapper[4821]: I1125 10:32:14.164236 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:14 crc kubenswrapper[4821]: I1125 10:32:14.164244 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:14 crc kubenswrapper[4821]: I1125 10:32:14.164296 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:14 crc kubenswrapper[4821]: I1125 10:32:14.164318 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:14 crc kubenswrapper[4821]: I1125 10:32:14.164329 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:14 crc kubenswrapper[4821]: I1125 10:32:14.165261 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:14 crc kubenswrapper[4821]: I1125 10:32:14.165295 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:14 crc kubenswrapper[4821]: I1125 10:32:14.165305 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:14 crc kubenswrapper[4821]: I1125 10:32:14.444685 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:32:15 crc kubenswrapper[4821]: I1125 10:32:15.168079 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"88a06b811e5ec52ce984a2403ba2c98332ba60466d08b3580ad9192fac6caef2"} Nov 25 10:32:15 crc kubenswrapper[4821]: I1125 10:32:15.168129 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"eea762f627aa69fb88a9bd82545fa677e9ccfa97f99a5e92c0aa490e619c0f9a"} Nov 25 10:32:15 crc kubenswrapper[4821]: I1125 10:32:15.168145 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"18465f3a6916c36457a305ba5d568c9077f34f1d000532ed518f237d100008b7"} Nov 25 10:32:15 crc kubenswrapper[4821]: I1125 10:32:15.168150 4821 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:32:15 crc kubenswrapper[4821]: I1125 10:32:15.168189 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:32:15 crc kubenswrapper[4821]: I1125 10:32:15.168853 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:15 crc kubenswrapper[4821]: I1125 10:32:15.168886 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:15 crc kubenswrapper[4821]: I1125 10:32:15.168895 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:15 crc kubenswrapper[4821]: I1125 10:32:15.201586 4821 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 25 10:32:15 crc kubenswrapper[4821]: I1125 10:32:15.201647 4821 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 25 10:32:16 crc kubenswrapper[4821]: I1125 10:32:16.178527 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"b915b565f7fbb3e4e856e7f1035a519e82bfb3023660c121f6e3813de353bb92"} Nov 25 10:32:16 crc kubenswrapper[4821]: I1125 10:32:16.178583 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"fd0385f6ed9597974267f7c45547f6b6725b4ee83a2cc0508ff2d1c1ce8753ad"} Nov 25 10:32:16 crc kubenswrapper[4821]: I1125 10:32:16.178596 4821 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:32:16 crc kubenswrapper[4821]: I1125 10:32:16.178618 4821 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:32:16 crc kubenswrapper[4821]: I1125 10:32:16.179687 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:16 crc kubenswrapper[4821]: I1125 10:32:16.179725 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:16 crc kubenswrapper[4821]: I1125 10:32:16.179736 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:16 crc kubenswrapper[4821]: I1125 10:32:16.179907 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:16 crc kubenswrapper[4821]: I1125 10:32:16.179985 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:16 crc kubenswrapper[4821]: I1125 10:32:16.180006 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:16 crc kubenswrapper[4821]: I1125 10:32:16.496781 4821 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:32:16 crc kubenswrapper[4821]: I1125 10:32:16.497940 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:16 crc kubenswrapper[4821]: I1125 10:32:16.497981 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:16 crc kubenswrapper[4821]: I1125 10:32:16.497992 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:16 crc kubenswrapper[4821]: I1125 10:32:16.498015 4821 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 10:32:16 crc kubenswrapper[4821]: I1125 10:32:16.679915 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 10:32:16 crc kubenswrapper[4821]: I1125 10:32:16.680090 4821 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:32:16 crc kubenswrapper[4821]: I1125 10:32:16.681221 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:16 crc kubenswrapper[4821]: I1125 10:32:16.681262 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:16 crc kubenswrapper[4821]: I1125 10:32:16.681276 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:17 crc kubenswrapper[4821]: I1125 10:32:17.180410 4821 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:32:17 crc kubenswrapper[4821]: I1125 10:32:17.181144 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:17 crc kubenswrapper[4821]: I1125 10:32:17.181205 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:17 crc kubenswrapper[4821]: I1125 10:32:17.181217 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:17 crc kubenswrapper[4821]: I1125 10:32:17.383501 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 10:32:17 crc kubenswrapper[4821]: I1125 10:32:17.383814 4821 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:32:17 crc kubenswrapper[4821]: I1125 10:32:17.385367 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:17 crc kubenswrapper[4821]: I1125 10:32:17.385419 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:17 crc kubenswrapper[4821]: I1125 10:32:17.385441 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:18 crc kubenswrapper[4821]: I1125 10:32:18.123000 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Nov 25 10:32:18 crc kubenswrapper[4821]: I1125 10:32:18.182728 4821 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:32:18 crc kubenswrapper[4821]: I1125 10:32:18.183806 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:18 crc kubenswrapper[4821]: I1125 10:32:18.183850 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:18 crc kubenswrapper[4821]: I1125 10:32:18.183864 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:18 crc kubenswrapper[4821]: I1125 10:32:18.416485 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:32:18 crc kubenswrapper[4821]: I1125 10:32:18.416675 4821 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:32:18 crc kubenswrapper[4821]: I1125 10:32:18.417746 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:18 crc kubenswrapper[4821]: I1125 10:32:18.417783 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:18 crc kubenswrapper[4821]: I1125 10:32:18.417797 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:19 crc kubenswrapper[4821]: I1125 10:32:19.391816 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Nov 25 10:32:19 crc kubenswrapper[4821]: I1125 10:32:19.392029 4821 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:32:19 crc kubenswrapper[4821]: I1125 10:32:19.393474 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:19 crc kubenswrapper[4821]: I1125 10:32:19.393531 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:19 crc kubenswrapper[4821]: I1125 10:32:19.393543 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:19 crc kubenswrapper[4821]: I1125 10:32:19.481992 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 10:32:19 crc kubenswrapper[4821]: I1125 10:32:19.482209 4821 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:32:19 crc kubenswrapper[4821]: I1125 10:32:19.483468 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:19 crc kubenswrapper[4821]: I1125 10:32:19.483518 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:19 crc kubenswrapper[4821]: I1125 10:32:19.483528 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:20 crc kubenswrapper[4821]: E1125 10:32:20.188972 4821 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 25 10:32:22 crc kubenswrapper[4821]: I1125 10:32:22.113687 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 10:32:22 crc kubenswrapper[4821]: I1125 10:32:22.113829 4821 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:32:22 crc kubenswrapper[4821]: I1125 10:32:22.115103 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:22 crc kubenswrapper[4821]: I1125 10:32:22.115138 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:22 crc kubenswrapper[4821]: I1125 10:32:22.115149 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:22 crc kubenswrapper[4821]: I1125 10:32:22.122688 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 10:32:22 crc kubenswrapper[4821]: I1125 10:32:22.190850 4821 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:32:22 crc kubenswrapper[4821]: I1125 10:32:22.192001 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:22 crc kubenswrapper[4821]: I1125 10:32:22.192049 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:22 crc kubenswrapper[4821]: I1125 10:32:22.192065 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:22 crc kubenswrapper[4821]: I1125 10:32:22.194614 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 10:32:23 crc kubenswrapper[4821]: I1125 10:32:23.193346 4821 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:32:23 crc kubenswrapper[4821]: I1125 10:32:23.194459 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:23 crc kubenswrapper[4821]: I1125 10:32:23.194514 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:23 crc kubenswrapper[4821]: I1125 10:32:23.194525 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:24 crc kubenswrapper[4821]: W1125 10:32:24.017322 4821 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout Nov 25 10:32:24 crc kubenswrapper[4821]: I1125 10:32:24.017430 4821 trace.go:236] Trace[968133969]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (25-Nov-2025 10:32:14.015) (total time: 10001ms): Nov 25 10:32:24 crc kubenswrapper[4821]: Trace[968133969]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (10:32:24.017) Nov 25 10:32:24 crc kubenswrapper[4821]: Trace[968133969]: [10.001870787s] [10.001870787s] END Nov 25 10:32:24 crc kubenswrapper[4821]: E1125 10:32:24.017456 4821 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Nov 25 10:32:24 crc kubenswrapper[4821]: I1125 10:32:24.068795 4821 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Nov 25 10:32:24 crc kubenswrapper[4821]: I1125 10:32:24.112435 4821 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 25 10:32:24 crc kubenswrapper[4821]: I1125 10:32:24.112486 4821 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 25 10:32:24 crc kubenswrapper[4821]: I1125 10:32:24.124541 4821 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\": RBAC: [clusterrole.rbac.authorization.k8s.io \"system:public-info-viewer\" not found, clusterrole.rbac.authorization.k8s.io \"system:openshift:public-info-viewer\" not found]","reason":"Forbidden","details":{},"code":403} Nov 25 10:32:24 crc kubenswrapper[4821]: I1125 10:32:24.124599 4821 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 25 10:32:25 crc kubenswrapper[4821]: I1125 10:32:25.200523 4821 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 25 10:32:25 crc kubenswrapper[4821]: I1125 10:32:25.200868 4821 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 25 10:32:28 crc kubenswrapper[4821]: I1125 10:32:28.077545 4821 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Nov 25 10:32:28 crc kubenswrapper[4821]: I1125 10:32:28.156027 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Nov 25 10:32:28 crc kubenswrapper[4821]: I1125 10:32:28.167371 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Nov 25 10:32:28 crc kubenswrapper[4821]: E1125 10:32:28.214252 4821 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"etcd-crc\" already exists" pod="openshift-etcd/etcd-crc" Nov 25 10:32:28 crc kubenswrapper[4821]: I1125 10:32:28.421590 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:32:28 crc kubenswrapper[4821]: I1125 10:32:28.425765 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.063061 4821 apiserver.go:52] "Watching apiserver" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.067188 4821 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.067473 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc"] Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.067809 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.067832 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 10:32:29 crc kubenswrapper[4821]: E1125 10:32:29.067873 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.067815 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.067919 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 10:32:29 crc kubenswrapper[4821]: E1125 10:32:29.067922 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.067855 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.068129 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:32:29 crc kubenswrapper[4821]: E1125 10:32:29.068329 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.069330 4821 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.069803 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.070106 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.070676 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.070822 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.070948 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.070982 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.071369 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.071466 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.071554 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.097260 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62fcb632-f12a-46e2-bb7d-a5311f291ee7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28d43531927e75c3745f100940f7f6bf79fd21026c1309194df8cd7a39361b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69633332d03d2551ff4abb253d358de398eab18486b92db630ce06c0447fb9cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ccb78d9a51ac83ee208c2451685a2512de9deaf742a663bb8c892aa59b76162\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27826e155a9352b1e1866e83fdab683dbcd749cc04b6a4da2db25ecabe3d9a52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://991252e5f56f128827a5054b8e228284bc0266166582cbaedaadc53ca2f5b401\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:32:29 crc kubenswrapper[4821]: E1125 10:32:29.110337 4821 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.110944 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.112129 4821 trace.go:236] Trace[1871870987]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (25-Nov-2025 10:32:14.541) (total time: 14570ms): Nov 25 10:32:29 crc kubenswrapper[4821]: Trace[1871870987]: ---"Objects listed" error: 14570ms (10:32:29.112) Nov 25 10:32:29 crc kubenswrapper[4821]: Trace[1871870987]: [14.570141006s] [14.570141006s] END Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.112171 4821 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.112832 4821 trace.go:236] Trace[1910318667]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (25-Nov-2025 10:32:17.133) (total time: 11978ms): Nov 25 10:32:29 crc kubenswrapper[4821]: Trace[1910318667]: ---"Objects listed" error: 11978ms (10:32:29.112) Nov 25 10:32:29 crc kubenswrapper[4821]: Trace[1910318667]: [11.978927811s] [11.978927811s] END Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.112870 4821 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.113295 4821 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Nov 25 10:32:29 crc kubenswrapper[4821]: E1125 10:32:29.113866 4821 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.114834 4821 trace.go:236] Trace[1683941795]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (25-Nov-2025 10:32:18.027) (total time: 11087ms): Nov 25 10:32:29 crc kubenswrapper[4821]: Trace[1683941795]: ---"Objects listed" error: 11087ms (10:32:29.114) Nov 25 10:32:29 crc kubenswrapper[4821]: Trace[1683941795]: [11.08729442s] [11.08729442s] END Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.115014 4821 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.122662 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.137788 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.158174 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.159055 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d9e58df-1755-4422-8d04-211f98c7c011\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea762f627aa69fb88a9bd82545fa677e9ccfa97f99a5e92c0aa490e619c0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a06b811e5ec52ce984a2403ba2c98332ba60466d08b3580ad9192fac6caef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0385f6ed9597974267f7c45547f6b6725b4ee83a2cc0508ff2d1c1ce8753ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b915b565f7fbb3e4e856e7f1035a519e82bfb3023660c121f6e3813de353bb92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18465f3a6916c36457a305ba5d568c9077f34f1d000532ed518f237d100008b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.170777 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.189025 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d9e58df-1755-4422-8d04-211f98c7c011\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea762f627aa69fb88a9bd82545fa677e9ccfa97f99a5e92c0aa490e619c0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a06b811e5ec52ce984a2403ba2c98332ba60466d08b3580ad9192fac6caef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0385f6ed9597974267f7c45547f6b6725b4ee83a2cc0508ff2d1c1ce8753ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b915b565f7fbb3e4e856e7f1035a519e82bfb3023660c121f6e3813de353bb92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18465f3a6916c36457a305ba5d568c9077f34f1d000532ed518f237d100008b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.204937 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.214879 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.215441 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.215468 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.215487 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.215503 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.215520 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.215541 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.215558 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.215576 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.215594 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.215787 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.215806 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.215864 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.215894 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.215879 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.215914 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.215914 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.215879 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.215932 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.215983 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.215984 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.216007 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.216029 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.216047 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.216064 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.216081 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.216087 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.216097 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.216136 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.216191 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.216214 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.216237 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.216260 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.216282 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.216306 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.216330 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.216351 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.216371 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.216392 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.216411 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.216434 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.216458 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.216481 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.216504 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.216526 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.216547 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.216103 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.216125 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.216248 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.216274 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.216290 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.216298 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.216406 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.216428 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.216548 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.216572 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.216749 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.216748 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.216777 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.216782 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.216785 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.216802 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.216825 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.216839 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.216861 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.216880 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.216896 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.216968 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.216999 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.217020 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.217046 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.217068 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.217089 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.217115 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.217137 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.217177 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.217201 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.217225 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.217295 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.217321 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.217345 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.217369 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.217391 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.217412 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.217432 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.217456 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.217480 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.217502 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.217523 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.217553 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.217607 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.217631 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.217655 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.217677 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.217704 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.217728 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.217750 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.217771 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.217795 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.217819 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.218115 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.218138 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.218154 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.218188 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.218204 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.218220 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.218244 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.218261 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.218277 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.218292 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.218327 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.218342 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.218357 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.218376 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.218391 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.218416 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.218430 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.218448 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.218464 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.218479 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.218495 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.218511 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.218528 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.217020 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.218543 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.217019 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.218560 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.218577 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.218593 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.218609 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.218623 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.218638 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.218656 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.218679 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.218701 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.218717 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.218733 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.218752 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.218776 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.218804 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.218826 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.218850 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.218877 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.218900 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.218922 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.218946 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.218968 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.218989 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.219009 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.219032 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.219054 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.219078 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.220430 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.220475 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.220499 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.220521 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.220543 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.220562 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.220582 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.220602 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.220623 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.220643 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.220664 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.220689 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.220715 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.220739 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.220763 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.220787 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.229728 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.229774 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.229798 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.229897 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.229934 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.229960 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.229977 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.229997 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.230080 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.230128 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.230313 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.230342 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.230370 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.230391 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.230484 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.230507 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.230535 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.230570 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.231153 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.231220 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.231244 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.231264 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.231342 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.231365 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.231386 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.217044 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.217062 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.217088 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.217107 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.217178 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.217283 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.217310 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.217341 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.238387 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.238793 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.217485 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.217534 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.217560 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.217594 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.217688 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.217792 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.217823 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.217837 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.217997 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.218017 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.218042 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.218216 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.218487 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.218526 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.219009 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.219086 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.219345 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.219364 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.219644 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.219706 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.219800 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.219941 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.220075 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.220090 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.220309 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.220348 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.220715 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.233100 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.236565 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.236912 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.237457 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.237665 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.237892 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.238132 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.239729 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.239929 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.240722 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.240741 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.241123 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: E1125 10:32:29.241529 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:32:29.741504691 +0000 UTC m=+20.277824548 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.243497 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.243609 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.243791 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.244051 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.244082 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.244184 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.244203 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.245180 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.245295 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.245573 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.245714 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.245950 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.245955 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.245992 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.246149 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.246322 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.246356 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.246421 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.246442 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.246542 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.246552 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.246629 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.246861 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.247170 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.247395 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.247501 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.247679 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.247266 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.247815 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.246558 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.247847 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.247984 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: E1125 10:32:29.246733 4821 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-apiserver-crc\" already exists" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.247219 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.248256 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.248326 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.248343 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.248349 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.248464 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.248801 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.248856 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.248863 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.248412 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.248846 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.249177 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.249198 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.249239 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.249429 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.249465 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.239664 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.231402 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.249748 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.249769 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.249797 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.249825 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.250033 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.250302 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.250372 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.250569 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.250063 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.250895 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.250920 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.250933 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.250967 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.250993 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.251026 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.251054 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.251096 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.251123 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.251152 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.251222 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.251245 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.251779 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.252013 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.252092 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.252261 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.252534 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.252849 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.253018 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.253436 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.253373 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.253752 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.253922 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.253963 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.253981 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.254001 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.254020 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.254043 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.254074 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.254103 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.254264 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.254308 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.254344 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.254366 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.254391 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.254411 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.254434 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.254456 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.254476 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.254500 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.254969 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.255018 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.255044 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.255101 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.255131 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.255250 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.255276 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.255066 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.255423 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.255504 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.255558 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.255576 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.255589 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.255588 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.255610 4821 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.255624 4821 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.255680 4821 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.255695 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.255747 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.255767 4821 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.255779 4821 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.255793 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.255804 4821 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.255815 4821 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.255826 4821 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.255837 4821 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.255849 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.255862 4821 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.255875 4821 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.255890 4821 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.255904 4821 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.255920 4821 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.255931 4821 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.255944 4821 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.255933 4821 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.255957 4821 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.255970 4821 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.255982 4821 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.255994 4821 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.256005 4821 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.256017 4821 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.256028 4821 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.256040 4821 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.256081 4821 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.256094 4821 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.256107 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.256120 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.256133 4821 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.256145 4821 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.256174 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.256199 4821 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.256150 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.256213 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.256235 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.256249 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.256260 4821 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.256272 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.256284 4821 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.256293 4821 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.256303 4821 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.256313 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.256323 4821 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.256333 4821 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.256342 4821 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.256354 4821 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.256375 4821 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.256384 4821 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.256393 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.256403 4821 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.256414 4821 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.256426 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.256437 4821 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.256447 4821 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.256458 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.256470 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.256480 4821 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.256492 4821 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.256503 4821 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.256512 4821 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.256524 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.256534 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.256546 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.256556 4821 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.256568 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.256581 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.256591 4821 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.256601 4821 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: E1125 10:32:29.256609 4821 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 10:32:29 crc kubenswrapper[4821]: E1125 10:32:29.256670 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 10:32:29.756651451 +0000 UTC m=+20.292971348 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.256612 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.256699 4821 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.256712 4821 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.256724 4821 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.256736 4821 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.256748 4821 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.256760 4821 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.256772 4821 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.256784 4821 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.256796 4821 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.256809 4821 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.256820 4821 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.256833 4821 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.256844 4821 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.256856 4821 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.256859 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.256867 4821 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.256914 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.256928 4821 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.256941 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.256953 4821 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.256965 4821 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.256991 4821 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.257004 4821 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.257017 4821 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.257029 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.257041 4821 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.257052 4821 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.257064 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.257076 4821 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.257090 4821 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.257104 4821 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.257127 4821 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.257141 4821 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.257146 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.257152 4821 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.257201 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.257216 4821 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.257228 4821 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.257240 4821 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.257252 4821 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.257263 4821 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.257275 4821 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.257299 4821 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.257313 4821 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.257326 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.257341 4821 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.257354 4821 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.257366 4821 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.257377 4821 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.257391 4821 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.257409 4821 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.257434 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.257448 4821 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.257461 4821 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.257473 4821 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.257486 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.257499 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.257510 4821 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.257522 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.257533 4821 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.257544 4821 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.257556 4821 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.257569 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.257580 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.257604 4821 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.257616 4821 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.257657 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 10:32:29 crc kubenswrapper[4821]: E1125 10:32:29.257666 4821 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 10:32:29 crc kubenswrapper[4821]: E1125 10:32:29.257732 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 10:32:29.757717823 +0000 UTC m=+20.294037670 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.261188 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.262183 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.262753 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.262996 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.263633 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.263712 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.274237 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.274381 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.274527 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.276822 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.276860 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.276922 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 10:32:29 crc kubenswrapper[4821]: E1125 10:32:29.276971 4821 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 10:32:29 crc kubenswrapper[4821]: E1125 10:32:29.276983 4821 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 10:32:29 crc kubenswrapper[4821]: E1125 10:32:29.276994 4821 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 10:32:29 crc kubenswrapper[4821]: E1125 10:32:29.277043 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 10:32:29.777027764 +0000 UTC m=+20.313347611 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 10:32:29 crc kubenswrapper[4821]: E1125 10:32:29.277098 4821 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 10:32:29 crc kubenswrapper[4821]: E1125 10:32:29.277126 4821 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 10:32:29 crc kubenswrapper[4821]: E1125 10:32:29.277140 4821 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 10:32:29 crc kubenswrapper[4821]: E1125 10:32:29.277209 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 10:32:29.777186197 +0000 UTC m=+20.313506114 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.277355 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62fcb632-f12a-46e2-bb7d-a5311f291ee7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28d43531927e75c3745f100940f7f6bf79fd21026c1309194df8cd7a39361b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69633332d03d2551ff4abb253d358de398eab18486b92db630ce06c0447fb9cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ccb78d9a51ac83ee208c2451685a2512de9deaf742a663bb8c892aa59b76162\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27826e155a9352b1e1866e83fdab683dbcd749cc04b6a4da2db25ecabe3d9a52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://991252e5f56f128827a5054b8e228284bc0266166582cbaedaadc53ca2f5b401\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.277488 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.277635 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.277735 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.277832 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.277897 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.278082 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.278493 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.278642 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.278821 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.278889 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.279032 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.280630 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.281119 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.281151 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.281209 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.281235 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.281271 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.281439 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.281483 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.281723 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.281930 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.283655 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.283737 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.284391 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.286228 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.288319 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.289819 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.290199 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.290317 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.290276 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.290693 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.290855 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.292103 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.292128 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.292262 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.292602 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.293452 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.293677 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.293783 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.294202 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.294286 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.294503 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.294737 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.308856 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.317904 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.323111 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.326499 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.358048 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.358312 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.358446 4821 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.358535 4821 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.358622 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.358699 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.358780 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.358562 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.358851 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.358925 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.358940 4821 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.358243 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.358956 4821 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.358985 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.358997 4821 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.359009 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.359020 4821 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.359029 4821 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.359038 4821 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.359047 4821 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.359056 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.359065 4821 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.359076 4821 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.359088 4821 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.359100 4821 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.359113 4821 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.359151 4821 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.359187 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.359198 4821 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.359210 4821 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.359222 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.359234 4821 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.359245 4821 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.359257 4821 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.359271 4821 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.359283 4821 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.359305 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.359318 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.359328 4821 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.359343 4821 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.359355 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.359367 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.359378 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.359390 4821 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.359402 4821 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.359414 4821 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.359426 4821 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.359438 4821 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.359450 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.359461 4821 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.359472 4821 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.359502 4821 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.359513 4821 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.359524 4821 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.359535 4821 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.359546 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.359557 4821 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.359568 4821 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.359578 4821 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.390425 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.402074 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 10:32:29 crc kubenswrapper[4821]: W1125 10:32:29.404521 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-c91139c769b0a254d219810ece873d1336bc9ba2e83c659bba1e904b3d4dd22b WatchSource:0}: Error finding container c91139c769b0a254d219810ece873d1336bc9ba2e83c659bba1e904b3d4dd22b: Status 404 returned error can't find the container with id c91139c769b0a254d219810ece873d1336bc9ba2e83c659bba1e904b3d4dd22b Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.407359 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 10:32:29 crc kubenswrapper[4821]: W1125 10:32:29.411839 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-bec56556f5188ab8ff5eac30572f8465718db7d0f715e73918c1b6926b87b7f4 WatchSource:0}: Error finding container bec56556f5188ab8ff5eac30572f8465718db7d0f715e73918c1b6926b87b7f4: Status 404 returned error can't find the container with id bec56556f5188ab8ff5eac30572f8465718db7d0f715e73918c1b6926b87b7f4 Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.763237 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:32:29 crc kubenswrapper[4821]: E1125 10:32:29.763433 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:32:30.763409216 +0000 UTC m=+21.299729053 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.763652 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.763686 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:32:29 crc kubenswrapper[4821]: E1125 10:32:29.763750 4821 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 10:32:29 crc kubenswrapper[4821]: E1125 10:32:29.763783 4821 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 10:32:29 crc kubenswrapper[4821]: E1125 10:32:29.763807 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 10:32:30.763796975 +0000 UTC m=+21.300116882 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 10:32:29 crc kubenswrapper[4821]: E1125 10:32:29.763837 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 10:32:30.763820685 +0000 UTC m=+21.300140612 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.864976 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:32:29 crc kubenswrapper[4821]: I1125 10:32:29.865013 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:32:29 crc kubenswrapper[4821]: E1125 10:32:29.865116 4821 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 10:32:29 crc kubenswrapper[4821]: E1125 10:32:29.865133 4821 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 10:32:29 crc kubenswrapper[4821]: E1125 10:32:29.865143 4821 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 10:32:29 crc kubenswrapper[4821]: E1125 10:32:29.865181 4821 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 10:32:29 crc kubenswrapper[4821]: E1125 10:32:29.865212 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 10:32:30.86519876 +0000 UTC m=+21.401518607 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 10:32:29 crc kubenswrapper[4821]: E1125 10:32:29.865216 4821 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 10:32:29 crc kubenswrapper[4821]: E1125 10:32:29.865229 4821 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 10:32:29 crc kubenswrapper[4821]: E1125 10:32:29.865276 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 10:32:30.865262032 +0000 UTC m=+21.401581879 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.117185 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.117657 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.118793 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.119388 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.120292 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.120787 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.121397 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.122336 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.122992 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.124032 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.124552 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.125632 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.126185 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.126326 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:30Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.126673 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.127565 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.128057 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.129075 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.129452 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.130016 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.131035 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.131536 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.133305 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.133737 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.134892 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.135498 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.136129 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.137590 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.138085 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.139025 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.139522 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.140376 4821 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.140476 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.142392 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.142828 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62fcb632-f12a-46e2-bb7d-a5311f291ee7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28d43531927e75c3745f100940f7f6bf79fd21026c1309194df8cd7a39361b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69633332d03d2551ff4abb253d358de398eab18486b92db630ce06c0447fb9cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ccb78d9a51ac83ee208c2451685a2512de9deaf742a663bb8c892aa59b76162\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27826e155a9352b1e1866e83fdab683dbcd749cc04b6a4da2db25ecabe3d9a52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://991252e5f56f128827a5054b8e228284bc0266166582cbaedaadc53ca2f5b401\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:30Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.143460 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.143958 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.145480 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.146442 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.147358 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.147972 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.149018 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.149488 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.150586 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.151373 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.152498 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.153037 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.154409 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.154913 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.156085 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.156590 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.157454 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.157940 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.158945 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.159843 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.160447 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.162309 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:30Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.162716 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-zlk96"] Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.163089 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-zlk96" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.165684 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.167489 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.167728 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.199804 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:30Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.211306 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"bec56556f5188ab8ff5eac30572f8465718db7d0f715e73918c1b6926b87b7f4"} Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.212701 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"06f949936bac99cfbf89d720c9549c074351d9077bd963fa024d87004c6be23e"} Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.212744 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"c91139c769b0a254d219810ece873d1336bc9ba2e83c659bba1e904b3d4dd22b"} Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.214826 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"17b8714be3bd121e06641873e003f08560b1f71e56d6ea936966df7d4896eb03"} Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.214855 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"85731e6b6dbe1777bc7161a010f39f79523ec4b08d71d5d8a975573aa7fe4957"} Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.214865 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"7410b356e2c21ee282877b249ff9751aa72c35e9cc522fdf41642cacc2b96a06"} Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.222809 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:30Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.236621 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:30Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.255789 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d9e58df-1755-4422-8d04-211f98c7c011\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea762f627aa69fb88a9bd82545fa677e9ccfa97f99a5e92c0aa490e619c0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a06b811e5ec52ce984a2403ba2c98332ba60466d08b3580ad9192fac6caef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0385f6ed9597974267f7c45547f6b6725b4ee83a2cc0508ff2d1c1ce8753ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b915b565f7fbb3e4e856e7f1035a519e82bfb3023660c121f6e3813de353bb92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18465f3a6916c36457a305ba5d568c9077f34f1d000532ed518f237d100008b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:30Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.268906 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/3d55aa91-9cb9-4c95-bc27-6fec05684c10-hosts-file\") pod \"node-resolver-zlk96\" (UID: \"3d55aa91-9cb9-4c95-bc27-6fec05684c10\") " pod="openshift-dns/node-resolver-zlk96" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.268955 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5vkd9\" (UniqueName: \"kubernetes.io/projected/3d55aa91-9cb9-4c95-bc27-6fec05684c10-kube-api-access-5vkd9\") pod \"node-resolver-zlk96\" (UID: \"3d55aa91-9cb9-4c95-bc27-6fec05684c10\") " pod="openshift-dns/node-resolver-zlk96" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.276848 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:30Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.297078 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d9e58df-1755-4422-8d04-211f98c7c011\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea762f627aa69fb88a9bd82545fa677e9ccfa97f99a5e92c0aa490e619c0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a06b811e5ec52ce984a2403ba2c98332ba60466d08b3580ad9192fac6caef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0385f6ed9597974267f7c45547f6b6725b4ee83a2cc0508ff2d1c1ce8753ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b915b565f7fbb3e4e856e7f1035a519e82bfb3023660c121f6e3813de353bb92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18465f3a6916c36457a305ba5d568c9077f34f1d000532ed518f237d100008b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:30Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.329105 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06f949936bac99cfbf89d720c9549c074351d9077bd963fa024d87004c6be23e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:30Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.342747 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:30Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.355744 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:30Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.367771 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:30Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.369703 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/3d55aa91-9cb9-4c95-bc27-6fec05684c10-hosts-file\") pod \"node-resolver-zlk96\" (UID: \"3d55aa91-9cb9-4c95-bc27-6fec05684c10\") " pod="openshift-dns/node-resolver-zlk96" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.369775 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5vkd9\" (UniqueName: \"kubernetes.io/projected/3d55aa91-9cb9-4c95-bc27-6fec05684c10-kube-api-access-5vkd9\") pod \"node-resolver-zlk96\" (UID: \"3d55aa91-9cb9-4c95-bc27-6fec05684c10\") " pod="openshift-dns/node-resolver-zlk96" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.369890 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/3d55aa91-9cb9-4c95-bc27-6fec05684c10-hosts-file\") pod \"node-resolver-zlk96\" (UID: \"3d55aa91-9cb9-4c95-bc27-6fec05684c10\") " pod="openshift-dns/node-resolver-zlk96" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.386418 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:30Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.391887 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5vkd9\" (UniqueName: \"kubernetes.io/projected/3d55aa91-9cb9-4c95-bc27-6fec05684c10-kube-api-access-5vkd9\") pod \"node-resolver-zlk96\" (UID: \"3d55aa91-9cb9-4c95-bc27-6fec05684c10\") " pod="openshift-dns/node-resolver-zlk96" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.398007 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17b8714be3bd121e06641873e003f08560b1f71e56d6ea936966df7d4896eb03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85731e6b6dbe1777bc7161a010f39f79523ec4b08d71d5d8a975573aa7fe4957\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:30Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.409639 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zlk96" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d55aa91-9cb9-4c95-bc27-6fec05684c10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vkd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zlk96\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:30Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.422060 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62fcb632-f12a-46e2-bb7d-a5311f291ee7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28d43531927e75c3745f100940f7f6bf79fd21026c1309194df8cd7a39361b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69633332d03d2551ff4abb253d358de398eab18486b92db630ce06c0447fb9cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ccb78d9a51ac83ee208c2451685a2512de9deaf742a663bb8c892aa59b76162\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27826e155a9352b1e1866e83fdab683dbcd749cc04b6a4da2db25ecabe3d9a52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://991252e5f56f128827a5054b8e228284bc0266166582cbaedaadc53ca2f5b401\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:30Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.477481 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-zlk96" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.772502 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.772583 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.772609 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:32:30 crc kubenswrapper[4821]: E1125 10:32:30.772691 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:32:32.772665783 +0000 UTC m=+23.308985630 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:32:30 crc kubenswrapper[4821]: E1125 10:32:30.772714 4821 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 10:32:30 crc kubenswrapper[4821]: E1125 10:32:30.772767 4821 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 10:32:30 crc kubenswrapper[4821]: E1125 10:32:30.772776 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 10:32:32.772752765 +0000 UTC m=+23.309072612 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 10:32:30 crc kubenswrapper[4821]: E1125 10:32:30.772824 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 10:32:32.772811317 +0000 UTC m=+23.309131234 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.873507 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.873547 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:32:30 crc kubenswrapper[4821]: E1125 10:32:30.873656 4821 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 10:32:30 crc kubenswrapper[4821]: E1125 10:32:30.873671 4821 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 10:32:30 crc kubenswrapper[4821]: E1125 10:32:30.873680 4821 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 10:32:30 crc kubenswrapper[4821]: E1125 10:32:30.873719 4821 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 10:32:30 crc kubenswrapper[4821]: E1125 10:32:30.873749 4821 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 10:32:30 crc kubenswrapper[4821]: E1125 10:32:30.873759 4821 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 10:32:30 crc kubenswrapper[4821]: E1125 10:32:30.873726 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 10:32:32.873714243 +0000 UTC m=+23.410034090 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 10:32:30 crc kubenswrapper[4821]: E1125 10:32:30.873821 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 10:32:32.873806595 +0000 UTC m=+23.410126442 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.962775 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-2krbf"] Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.963233 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" Nov 25 10:32:30 crc kubenswrapper[4821]: W1125 10:32:30.965888 4821 reflector.go:561] object-"openshift-machine-config-operator"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-machine-config-operator": no relationship found between node 'crc' and this object Nov 25 10:32:30 crc kubenswrapper[4821]: E1125 10:32:30.965931 4821 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-config-operator\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-machine-config-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 25 10:32:30 crc kubenswrapper[4821]: W1125 10:32:30.966335 4821 reflector.go:561] object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq": failed to list *v1.Secret: secrets "machine-config-daemon-dockercfg-r5tcq" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-machine-config-operator": no relationship found between node 'crc' and this object Nov 25 10:32:30 crc kubenswrapper[4821]: E1125 10:32:30.966373 4821 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-config-operator\"/\"machine-config-daemon-dockercfg-r5tcq\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"machine-config-daemon-dockercfg-r5tcq\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-machine-config-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 25 10:32:30 crc kubenswrapper[4821]: W1125 10:32:30.966421 4821 reflector.go:561] object-"openshift-machine-config-operator"/"proxy-tls": failed to list *v1.Secret: secrets "proxy-tls" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-machine-config-operator": no relationship found between node 'crc' and this object Nov 25 10:32:30 crc kubenswrapper[4821]: W1125 10:32:30.966428 4821 reflector.go:561] object-"openshift-machine-config-operator"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-machine-config-operator": no relationship found between node 'crc' and this object Nov 25 10:32:30 crc kubenswrapper[4821]: W1125 10:32:30.966446 4821 reflector.go:561] object-"openshift-machine-config-operator"/"kube-rbac-proxy": failed to list *v1.ConfigMap: configmaps "kube-rbac-proxy" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-machine-config-operator": no relationship found between node 'crc' and this object Nov 25 10:32:30 crc kubenswrapper[4821]: E1125 10:32:30.966482 4821 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-config-operator\"/\"kube-rbac-proxy\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-rbac-proxy\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-machine-config-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 25 10:32:30 crc kubenswrapper[4821]: E1125 10:32:30.966432 4821 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-config-operator\"/\"proxy-tls\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"proxy-tls\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-machine-config-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 25 10:32:30 crc kubenswrapper[4821]: E1125 10:32:30.966451 4821 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-config-operator\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-machine-config-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.969897 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-qppnw"] Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.970639 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-qppnw" Nov 25 10:32:30 crc kubenswrapper[4821]: W1125 10:32:30.973397 4821 reflector.go:561] object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz": failed to list *v1.Secret: secrets "multus-ancillary-tools-dockercfg-vnmsz" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-multus": no relationship found between node 'crc' and this object Nov 25 10:32:30 crc kubenswrapper[4821]: E1125 10:32:30.973437 4821 reflector.go:158] "Unhandled Error" err="object-\"openshift-multus\"/\"multus-ancillary-tools-dockercfg-vnmsz\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"multus-ancillary-tools-dockercfg-vnmsz\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-multus\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.973486 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.973548 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Nov 25 10:32:30 crc kubenswrapper[4821]: W1125 10:32:30.976806 4821 reflector.go:561] object-"openshift-multus"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-multus": no relationship found between node 'crc' and this object Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.976824 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Nov 25 10:32:30 crc kubenswrapper[4821]: E1125 10:32:30.976853 4821 reflector.go:158] "Unhandled Error" err="object-\"openshift-multus\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-multus\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.982760 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06f949936bac99cfbf89d720c9549c074351d9077bd963fa024d87004c6be23e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:30Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:30 crc kubenswrapper[4821]: I1125 10:32:30.996119 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:30Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.007069 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:31Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.026017 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d9e58df-1755-4422-8d04-211f98c7c011\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea762f627aa69fb88a9bd82545fa677e9ccfa97f99a5e92c0aa490e619c0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a06b811e5ec52ce984a2403ba2c98332ba60466d08b3580ad9192fac6caef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0385f6ed9597974267f7c45547f6b6725b4ee83a2cc0508ff2d1c1ce8753ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b915b565f7fbb3e4e856e7f1035a519e82bfb3023660c121f6e3813de353bb92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18465f3a6916c36457a305ba5d568c9077f34f1d000532ed518f237d100008b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:31Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.041363 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:31Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.052794 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17b8714be3bd121e06641873e003f08560b1f71e56d6ea936966df7d4896eb03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85731e6b6dbe1777bc7161a010f39f79523ec4b08d71d5d8a975573aa7fe4957\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:31Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.061815 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zlk96" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d55aa91-9cb9-4c95-bc27-6fec05684c10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vkd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zlk96\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:31Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.073117 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f948b87-ac86-4de6-ad64-c2ef947f84d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2krbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:31Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.074371 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxd4d\" (UniqueName: \"kubernetes.io/projected/357491b8-cd9c-486e-9898-5ba33678513d-kube-api-access-hxd4d\") pod \"multus-additional-cni-plugins-qppnw\" (UID: \"357491b8-cd9c-486e-9898-5ba33678513d\") " pod="openshift-multus/multus-additional-cni-plugins-qppnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.074416 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/357491b8-cd9c-486e-9898-5ba33678513d-tuning-conf-dir\") pod \"multus-additional-cni-plugins-qppnw\" (UID: \"357491b8-cd9c-486e-9898-5ba33678513d\") " pod="openshift-multus/multus-additional-cni-plugins-qppnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.074446 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/357491b8-cd9c-486e-9898-5ba33678513d-system-cni-dir\") pod \"multus-additional-cni-plugins-qppnw\" (UID: \"357491b8-cd9c-486e-9898-5ba33678513d\") " pod="openshift-multus/multus-additional-cni-plugins-qppnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.074464 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/357491b8-cd9c-486e-9898-5ba33678513d-cnibin\") pod \"multus-additional-cni-plugins-qppnw\" (UID: \"357491b8-cd9c-486e-9898-5ba33678513d\") " pod="openshift-multus/multus-additional-cni-plugins-qppnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.074503 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/357491b8-cd9c-486e-9898-5ba33678513d-cni-binary-copy\") pod \"multus-additional-cni-plugins-qppnw\" (UID: \"357491b8-cd9c-486e-9898-5ba33678513d\") " pod="openshift-multus/multus-additional-cni-plugins-qppnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.074559 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qq48b\" (UniqueName: \"kubernetes.io/projected/5f948b87-ac86-4de6-ad64-c2ef947f84d4-kube-api-access-qq48b\") pod \"machine-config-daemon-2krbf\" (UID: \"5f948b87-ac86-4de6-ad64-c2ef947f84d4\") " pod="openshift-machine-config-operator/machine-config-daemon-2krbf" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.074604 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/5f948b87-ac86-4de6-ad64-c2ef947f84d4-rootfs\") pod \"machine-config-daemon-2krbf\" (UID: \"5f948b87-ac86-4de6-ad64-c2ef947f84d4\") " pod="openshift-machine-config-operator/machine-config-daemon-2krbf" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.074637 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5f948b87-ac86-4de6-ad64-c2ef947f84d4-mcd-auth-proxy-config\") pod \"machine-config-daemon-2krbf\" (UID: \"5f948b87-ac86-4de6-ad64-c2ef947f84d4\") " pod="openshift-machine-config-operator/machine-config-daemon-2krbf" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.074774 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/357491b8-cd9c-486e-9898-5ba33678513d-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-qppnw\" (UID: \"357491b8-cd9c-486e-9898-5ba33678513d\") " pod="openshift-multus/multus-additional-cni-plugins-qppnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.074813 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5f948b87-ac86-4de6-ad64-c2ef947f84d4-proxy-tls\") pod \"machine-config-daemon-2krbf\" (UID: \"5f948b87-ac86-4de6-ad64-c2ef947f84d4\") " pod="openshift-machine-config-operator/machine-config-daemon-2krbf" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.074836 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/357491b8-cd9c-486e-9898-5ba33678513d-os-release\") pod \"multus-additional-cni-plugins-qppnw\" (UID: \"357491b8-cd9c-486e-9898-5ba33678513d\") " pod="openshift-multus/multus-additional-cni-plugins-qppnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.087384 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62fcb632-f12a-46e2-bb7d-a5311f291ee7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28d43531927e75c3745f100940f7f6bf79fd21026c1309194df8cd7a39361b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69633332d03d2551ff4abb253d358de398eab18486b92db630ce06c0447fb9cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ccb78d9a51ac83ee208c2451685a2512de9deaf742a663bb8c892aa59b76162\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27826e155a9352b1e1866e83fdab683dbcd749cc04b6a4da2db25ecabe3d9a52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://991252e5f56f128827a5054b8e228284bc0266166582cbaedaadc53ca2f5b401\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:31Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.113549 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:32:31 crc kubenswrapper[4821]: E1125 10:32:31.113665 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.113946 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:32:31 crc kubenswrapper[4821]: E1125 10:32:31.114086 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.114150 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:32:31 crc kubenswrapper[4821]: E1125 10:32:31.114260 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.119133 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:31Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.148982 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:31Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.166838 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17b8714be3bd121e06641873e003f08560b1f71e56d6ea936966df7d4896eb03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85731e6b6dbe1777bc7161a010f39f79523ec4b08d71d5d8a975573aa7fe4957\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:31Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.176158 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/357491b8-cd9c-486e-9898-5ba33678513d-tuning-conf-dir\") pod \"multus-additional-cni-plugins-qppnw\" (UID: \"357491b8-cd9c-486e-9898-5ba33678513d\") " pod="openshift-multus/multus-additional-cni-plugins-qppnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.176240 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/357491b8-cd9c-486e-9898-5ba33678513d-system-cni-dir\") pod \"multus-additional-cni-plugins-qppnw\" (UID: \"357491b8-cd9c-486e-9898-5ba33678513d\") " pod="openshift-multus/multus-additional-cni-plugins-qppnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.176263 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/357491b8-cd9c-486e-9898-5ba33678513d-cnibin\") pod \"multus-additional-cni-plugins-qppnw\" (UID: \"357491b8-cd9c-486e-9898-5ba33678513d\") " pod="openshift-multus/multus-additional-cni-plugins-qppnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.176280 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/357491b8-cd9c-486e-9898-5ba33678513d-cni-binary-copy\") pod \"multus-additional-cni-plugins-qppnw\" (UID: \"357491b8-cd9c-486e-9898-5ba33678513d\") " pod="openshift-multus/multus-additional-cni-plugins-qppnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.176296 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qq48b\" (UniqueName: \"kubernetes.io/projected/5f948b87-ac86-4de6-ad64-c2ef947f84d4-kube-api-access-qq48b\") pod \"machine-config-daemon-2krbf\" (UID: \"5f948b87-ac86-4de6-ad64-c2ef947f84d4\") " pod="openshift-machine-config-operator/machine-config-daemon-2krbf" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.176312 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/5f948b87-ac86-4de6-ad64-c2ef947f84d4-rootfs\") pod \"machine-config-daemon-2krbf\" (UID: \"5f948b87-ac86-4de6-ad64-c2ef947f84d4\") " pod="openshift-machine-config-operator/machine-config-daemon-2krbf" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.176327 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5f948b87-ac86-4de6-ad64-c2ef947f84d4-mcd-auth-proxy-config\") pod \"machine-config-daemon-2krbf\" (UID: \"5f948b87-ac86-4de6-ad64-c2ef947f84d4\") " pod="openshift-machine-config-operator/machine-config-daemon-2krbf" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.176358 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/357491b8-cd9c-486e-9898-5ba33678513d-system-cni-dir\") pod \"multus-additional-cni-plugins-qppnw\" (UID: \"357491b8-cd9c-486e-9898-5ba33678513d\") " pod="openshift-multus/multus-additional-cni-plugins-qppnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.176387 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/357491b8-cd9c-486e-9898-5ba33678513d-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-qppnw\" (UID: \"357491b8-cd9c-486e-9898-5ba33678513d\") " pod="openshift-multus/multus-additional-cni-plugins-qppnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.176408 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5f948b87-ac86-4de6-ad64-c2ef947f84d4-proxy-tls\") pod \"machine-config-daemon-2krbf\" (UID: \"5f948b87-ac86-4de6-ad64-c2ef947f84d4\") " pod="openshift-machine-config-operator/machine-config-daemon-2krbf" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.176429 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/357491b8-cd9c-486e-9898-5ba33678513d-os-release\") pod \"multus-additional-cni-plugins-qppnw\" (UID: \"357491b8-cd9c-486e-9898-5ba33678513d\") " pod="openshift-multus/multus-additional-cni-plugins-qppnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.176451 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxd4d\" (UniqueName: \"kubernetes.io/projected/357491b8-cd9c-486e-9898-5ba33678513d-kube-api-access-hxd4d\") pod \"multus-additional-cni-plugins-qppnw\" (UID: \"357491b8-cd9c-486e-9898-5ba33678513d\") " pod="openshift-multus/multus-additional-cni-plugins-qppnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.176505 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/5f948b87-ac86-4de6-ad64-c2ef947f84d4-rootfs\") pod \"machine-config-daemon-2krbf\" (UID: \"5f948b87-ac86-4de6-ad64-c2ef947f84d4\") " pod="openshift-machine-config-operator/machine-config-daemon-2krbf" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.176380 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/357491b8-cd9c-486e-9898-5ba33678513d-cnibin\") pod \"multus-additional-cni-plugins-qppnw\" (UID: \"357491b8-cd9c-486e-9898-5ba33678513d\") " pod="openshift-multus/multus-additional-cni-plugins-qppnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.176915 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/357491b8-cd9c-486e-9898-5ba33678513d-tuning-conf-dir\") pod \"multus-additional-cni-plugins-qppnw\" (UID: \"357491b8-cd9c-486e-9898-5ba33678513d\") " pod="openshift-multus/multus-additional-cni-plugins-qppnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.176915 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/357491b8-cd9c-486e-9898-5ba33678513d-os-release\") pod \"multus-additional-cni-plugins-qppnw\" (UID: \"357491b8-cd9c-486e-9898-5ba33678513d\") " pod="openshift-multus/multus-additional-cni-plugins-qppnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.177470 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/357491b8-cd9c-486e-9898-5ba33678513d-cni-binary-copy\") pod \"multus-additional-cni-plugins-qppnw\" (UID: \"357491b8-cd9c-486e-9898-5ba33678513d\") " pod="openshift-multus/multus-additional-cni-plugins-qppnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.177564 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/357491b8-cd9c-486e-9898-5ba33678513d-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-qppnw\" (UID: \"357491b8-cd9c-486e-9898-5ba33678513d\") " pod="openshift-multus/multus-additional-cni-plugins-qppnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.178251 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zlk96" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d55aa91-9cb9-4c95-bc27-6fec05684c10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vkd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zlk96\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:31Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.198322 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d9e58df-1755-4422-8d04-211f98c7c011\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea762f627aa69fb88a9bd82545fa677e9ccfa97f99a5e92c0aa490e619c0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a06b811e5ec52ce984a2403ba2c98332ba60466d08b3580ad9192fac6caef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0385f6ed9597974267f7c45547f6b6725b4ee83a2cc0508ff2d1c1ce8753ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b915b565f7fbb3e4e856e7f1035a519e82bfb3023660c121f6e3813de353bb92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18465f3a6916c36457a305ba5d568c9077f34f1d000532ed518f237d100008b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:31Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.211039 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:31Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.217206 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-zlk96" event={"ID":"3d55aa91-9cb9-4c95-bc27-6fec05684c10","Type":"ContainerStarted","Data":"c4521b5baee45ffc9c7661c781c0b3de458dff325ad525dde283d46af71c0cc0"} Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.217253 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-zlk96" event={"ID":"3d55aa91-9cb9-4c95-bc27-6fec05684c10","Type":"ContainerStarted","Data":"7dfdb4ea9cc706da91629be1a3a5a23a15b778442d3c68037ce9f610844d6b7e"} Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.225192 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62fcb632-f12a-46e2-bb7d-a5311f291ee7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28d43531927e75c3745f100940f7f6bf79fd21026c1309194df8cd7a39361b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69633332d03d2551ff4abb253d358de398eab18486b92db630ce06c0447fb9cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ccb78d9a51ac83ee208c2451685a2512de9deaf742a663bb8c892aa59b76162\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27826e155a9352b1e1866e83fdab683dbcd749cc04b6a4da2db25ecabe3d9a52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://991252e5f56f128827a5054b8e228284bc0266166582cbaedaadc53ca2f5b401\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:31Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.236809 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:31Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.247187 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f948b87-ac86-4de6-ad64-c2ef947f84d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2krbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:31Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.260505 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qppnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"357491b8-cd9c-486e-9898-5ba33678513d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qppnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:31Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.276310 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06f949936bac99cfbf89d720c9549c074351d9077bd963fa024d87004c6be23e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:31Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.290464 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:31Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.301636 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:31Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.321464 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d9e58df-1755-4422-8d04-211f98c7c011\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea762f627aa69fb88a9bd82545fa677e9ccfa97f99a5e92c0aa490e619c0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a06b811e5ec52ce984a2403ba2c98332ba60466d08b3580ad9192fac6caef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0385f6ed9597974267f7c45547f6b6725b4ee83a2cc0508ff2d1c1ce8753ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b915b565f7fbb3e4e856e7f1035a519e82bfb3023660c121f6e3813de353bb92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18465f3a6916c36457a305ba5d568c9077f34f1d000532ed518f237d100008b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:31Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.332870 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:31Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.333516 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-ldzbz"] Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.333886 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-ldzbz" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.335350 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-mzhnw"] Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.335895 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.336084 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.336377 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.337693 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.337719 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.338082 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.338820 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.338856 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.340891 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.341300 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.345402 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f948b87-ac86-4de6-ad64-c2ef947f84d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2krbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:31Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.358628 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qppnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"357491b8-cd9c-486e-9898-5ba33678513d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qppnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:31Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.373235 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62fcb632-f12a-46e2-bb7d-a5311f291ee7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28d43531927e75c3745f100940f7f6bf79fd21026c1309194df8cd7a39361b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69633332d03d2551ff4abb253d358de398eab18486b92db630ce06c0447fb9cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ccb78d9a51ac83ee208c2451685a2512de9deaf742a663bb8c892aa59b76162\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27826e155a9352b1e1866e83fdab683dbcd749cc04b6a4da2db25ecabe3d9a52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://991252e5f56f128827a5054b8e228284bc0266166582cbaedaadc53ca2f5b401\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:31Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.389480 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06f949936bac99cfbf89d720c9549c074351d9077bd963fa024d87004c6be23e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:31Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.401440 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:31Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.412337 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zlk96" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d55aa91-9cb9-4c95-bc27-6fec05684c10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4521b5baee45ffc9c7661c781c0b3de458dff325ad525dde283d46af71c0cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vkd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zlk96\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:31Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.434651 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:31Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.458787 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17b8714be3bd121e06641873e003f08560b1f71e56d6ea936966df7d4896eb03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85731e6b6dbe1777bc7161a010f39f79523ec4b08d71d5d8a975573aa7fe4957\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:31Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.472028 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62fcb632-f12a-46e2-bb7d-a5311f291ee7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28d43531927e75c3745f100940f7f6bf79fd21026c1309194df8cd7a39361b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69633332d03d2551ff4abb253d358de398eab18486b92db630ce06c0447fb9cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ccb78d9a51ac83ee208c2451685a2512de9deaf742a663bb8c892aa59b76162\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27826e155a9352b1e1866e83fdab683dbcd749cc04b6a4da2db25ecabe3d9a52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://991252e5f56f128827a5054b8e228284bc0266166582cbaedaadc53ca2f5b401\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:31Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.478203 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/b84e32ae-4897-4644-a656-6bd82012cc81-multus-daemon-config\") pod \"multus-ldzbz\" (UID: \"b84e32ae-4897-4644-a656-6bd82012cc81\") " pod="openshift-multus/multus-ldzbz" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.478251 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-etc-openvswitch\") pod \"ovnkube-node-mzhnw\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.478277 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/954085fa-d03a-4e88-9dfd-566257e3558d-env-overrides\") pod \"ovnkube-node-mzhnw\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.478312 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-var-lib-openvswitch\") pod \"ovnkube-node-mzhnw\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.478333 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/954085fa-d03a-4e88-9dfd-566257e3558d-ovnkube-config\") pod \"ovnkube-node-mzhnw\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.478353 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/954085fa-d03a-4e88-9dfd-566257e3558d-ovnkube-script-lib\") pod \"ovnkube-node-mzhnw\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.478374 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b84e32ae-4897-4644-a656-6bd82012cc81-host-var-lib-cni-bin\") pod \"multus-ldzbz\" (UID: \"b84e32ae-4897-4644-a656-6bd82012cc81\") " pod="openshift-multus/multus-ldzbz" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.478396 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/b84e32ae-4897-4644-a656-6bd82012cc81-hostroot\") pod \"multus-ldzbz\" (UID: \"b84e32ae-4897-4644-a656-6bd82012cc81\") " pod="openshift-multus/multus-ldzbz" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.478416 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-host-cni-netd\") pod \"ovnkube-node-mzhnw\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.478435 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/b84e32ae-4897-4644-a656-6bd82012cc81-os-release\") pod \"multus-ldzbz\" (UID: \"b84e32ae-4897-4644-a656-6bd82012cc81\") " pod="openshift-multus/multus-ldzbz" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.478457 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/b84e32ae-4897-4644-a656-6bd82012cc81-host-var-lib-kubelet\") pod \"multus-ldzbz\" (UID: \"b84e32ae-4897-4644-a656-6bd82012cc81\") " pod="openshift-multus/multus-ldzbz" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.478481 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/b84e32ae-4897-4644-a656-6bd82012cc81-multus-conf-dir\") pod \"multus-ldzbz\" (UID: \"b84e32ae-4897-4644-a656-6bd82012cc81\") " pod="openshift-multus/multus-ldzbz" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.478504 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-host-run-ovn-kubernetes\") pod \"ovnkube-node-mzhnw\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.478524 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/954085fa-d03a-4e88-9dfd-566257e3558d-ovn-node-metrics-cert\") pod \"ovnkube-node-mzhnw\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.478548 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b84e32ae-4897-4644-a656-6bd82012cc81-host-run-netns\") pod \"multus-ldzbz\" (UID: \"b84e32ae-4897-4644-a656-6bd82012cc81\") " pod="openshift-multus/multus-ldzbz" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.478569 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/b84e32ae-4897-4644-a656-6bd82012cc81-host-run-multus-certs\") pod \"multus-ldzbz\" (UID: \"b84e32ae-4897-4644-a656-6bd82012cc81\") " pod="openshift-multus/multus-ldzbz" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.478593 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b84e32ae-4897-4644-a656-6bd82012cc81-system-cni-dir\") pod \"multus-ldzbz\" (UID: \"b84e32ae-4897-4644-a656-6bd82012cc81\") " pod="openshift-multus/multus-ldzbz" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.478614 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b84e32ae-4897-4644-a656-6bd82012cc81-multus-cni-dir\") pod \"multus-ldzbz\" (UID: \"b84e32ae-4897-4644-a656-6bd82012cc81\") " pod="openshift-multus/multus-ldzbz" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.478636 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/b84e32ae-4897-4644-a656-6bd82012cc81-host-var-lib-cni-multus\") pod \"multus-ldzbz\" (UID: \"b84e32ae-4897-4644-a656-6bd82012cc81\") " pod="openshift-multus/multus-ldzbz" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.478658 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-host-run-netns\") pod \"ovnkube-node-mzhnw\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.478678 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-run-ovn\") pod \"ovnkube-node-mzhnw\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.478707 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/b84e32ae-4897-4644-a656-6bd82012cc81-multus-socket-dir-parent\") pod \"multus-ldzbz\" (UID: \"b84e32ae-4897-4644-a656-6bd82012cc81\") " pod="openshift-multus/multus-ldzbz" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.478742 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/b84e32ae-4897-4644-a656-6bd82012cc81-cni-binary-copy\") pod \"multus-ldzbz\" (UID: \"b84e32ae-4897-4644-a656-6bd82012cc81\") " pod="openshift-multus/multus-ldzbz" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.478763 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hrb87\" (UniqueName: \"kubernetes.io/projected/954085fa-d03a-4e88-9dfd-566257e3558d-kube-api-access-hrb87\") pod \"ovnkube-node-mzhnw\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.478849 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/b84e32ae-4897-4644-a656-6bd82012cc81-cnibin\") pod \"multus-ldzbz\" (UID: \"b84e32ae-4897-4644-a656-6bd82012cc81\") " pod="openshift-multus/multus-ldzbz" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.478896 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b84e32ae-4897-4644-a656-6bd82012cc81-etc-kubernetes\") pod \"multus-ldzbz\" (UID: \"b84e32ae-4897-4644-a656-6bd82012cc81\") " pod="openshift-multus/multus-ldzbz" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.478940 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-host-slash\") pod \"ovnkube-node-mzhnw\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.479000 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-node-log\") pod \"ovnkube-node-mzhnw\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.479056 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-run-systemd\") pod \"ovnkube-node-mzhnw\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.479091 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-run-openvswitch\") pod \"ovnkube-node-mzhnw\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.479116 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-log-socket\") pod \"ovnkube-node-mzhnw\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.479140 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-mzhnw\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.479226 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/b84e32ae-4897-4644-a656-6bd82012cc81-host-run-k8s-cni-cncf-io\") pod \"multus-ldzbz\" (UID: \"b84e32ae-4897-4644-a656-6bd82012cc81\") " pod="openshift-multus/multus-ldzbz" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.479255 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g5pmc\" (UniqueName: \"kubernetes.io/projected/b84e32ae-4897-4644-a656-6bd82012cc81-kube-api-access-g5pmc\") pod \"multus-ldzbz\" (UID: \"b84e32ae-4897-4644-a656-6bd82012cc81\") " pod="openshift-multus/multus-ldzbz" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.479280 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-host-kubelet\") pod \"ovnkube-node-mzhnw\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.479303 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-systemd-units\") pod \"ovnkube-node-mzhnw\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.479334 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-host-cni-bin\") pod \"ovnkube-node-mzhnw\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.487138 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:31Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.500834 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f948b87-ac86-4de6-ad64-c2ef947f84d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2krbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:31Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.515854 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qppnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"357491b8-cd9c-486e-9898-5ba33678513d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qppnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:31Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.531715 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06f949936bac99cfbf89d720c9549c074351d9077bd963fa024d87004c6be23e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:31Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.545202 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:31Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.564596 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:31Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.576726 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17b8714be3bd121e06641873e003f08560b1f71e56d6ea936966df7d4896eb03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85731e6b6dbe1777bc7161a010f39f79523ec4b08d71d5d8a975573aa7fe4957\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:31Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.579813 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/b84e32ae-4897-4644-a656-6bd82012cc81-os-release\") pod \"multus-ldzbz\" (UID: \"b84e32ae-4897-4644-a656-6bd82012cc81\") " pod="openshift-multus/multus-ldzbz" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.579972 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/b84e32ae-4897-4644-a656-6bd82012cc81-os-release\") pod \"multus-ldzbz\" (UID: \"b84e32ae-4897-4644-a656-6bd82012cc81\") " pod="openshift-multus/multus-ldzbz" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.580146 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/b84e32ae-4897-4644-a656-6bd82012cc81-host-var-lib-kubelet\") pod \"multus-ldzbz\" (UID: \"b84e32ae-4897-4644-a656-6bd82012cc81\") " pod="openshift-multus/multus-ldzbz" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.580258 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/b84e32ae-4897-4644-a656-6bd82012cc81-multus-conf-dir\") pod \"multus-ldzbz\" (UID: \"b84e32ae-4897-4644-a656-6bd82012cc81\") " pod="openshift-multus/multus-ldzbz" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.580368 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/954085fa-d03a-4e88-9dfd-566257e3558d-ovn-node-metrics-cert\") pod \"ovnkube-node-mzhnw\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.580462 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-host-run-ovn-kubernetes\") pod \"ovnkube-node-mzhnw\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.580555 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b84e32ae-4897-4644-a656-6bd82012cc81-host-run-netns\") pod \"multus-ldzbz\" (UID: \"b84e32ae-4897-4644-a656-6bd82012cc81\") " pod="openshift-multus/multus-ldzbz" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.580653 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/b84e32ae-4897-4644-a656-6bd82012cc81-host-run-multus-certs\") pod \"multus-ldzbz\" (UID: \"b84e32ae-4897-4644-a656-6bd82012cc81\") " pod="openshift-multus/multus-ldzbz" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.580753 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b84e32ae-4897-4644-a656-6bd82012cc81-system-cni-dir\") pod \"multus-ldzbz\" (UID: \"b84e32ae-4897-4644-a656-6bd82012cc81\") " pod="openshift-multus/multus-ldzbz" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.580253 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/b84e32ae-4897-4644-a656-6bd82012cc81-host-var-lib-kubelet\") pod \"multus-ldzbz\" (UID: \"b84e32ae-4897-4644-a656-6bd82012cc81\") " pod="openshift-multus/multus-ldzbz" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.580586 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b84e32ae-4897-4644-a656-6bd82012cc81-host-run-netns\") pod \"multus-ldzbz\" (UID: \"b84e32ae-4897-4644-a656-6bd82012cc81\") " pod="openshift-multus/multus-ldzbz" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.580293 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/b84e32ae-4897-4644-a656-6bd82012cc81-multus-conf-dir\") pod \"multus-ldzbz\" (UID: \"b84e32ae-4897-4644-a656-6bd82012cc81\") " pod="openshift-multus/multus-ldzbz" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.580899 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b84e32ae-4897-4644-a656-6bd82012cc81-system-cni-dir\") pod \"multus-ldzbz\" (UID: \"b84e32ae-4897-4644-a656-6bd82012cc81\") " pod="openshift-multus/multus-ldzbz" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.580556 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-host-run-ovn-kubernetes\") pod \"ovnkube-node-mzhnw\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.580854 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b84e32ae-4897-4644-a656-6bd82012cc81-multus-cni-dir\") pod \"multus-ldzbz\" (UID: \"b84e32ae-4897-4644-a656-6bd82012cc81\") " pod="openshift-multus/multus-ldzbz" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.581017 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/b84e32ae-4897-4644-a656-6bd82012cc81-host-var-lib-cni-multus\") pod \"multus-ldzbz\" (UID: \"b84e32ae-4897-4644-a656-6bd82012cc81\") " pod="openshift-multus/multus-ldzbz" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.581041 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-host-run-netns\") pod \"ovnkube-node-mzhnw\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.581064 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-run-ovn\") pod \"ovnkube-node-mzhnw\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.581087 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/b84e32ae-4897-4644-a656-6bd82012cc81-host-var-lib-cni-multus\") pod \"multus-ldzbz\" (UID: \"b84e32ae-4897-4644-a656-6bd82012cc81\") " pod="openshift-multus/multus-ldzbz" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.581100 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/b84e32ae-4897-4644-a656-6bd82012cc81-multus-socket-dir-parent\") pod \"multus-ldzbz\" (UID: \"b84e32ae-4897-4644-a656-6bd82012cc81\") " pod="openshift-multus/multus-ldzbz" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.581123 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-host-run-netns\") pod \"ovnkube-node-mzhnw\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.581131 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/b84e32ae-4897-4644-a656-6bd82012cc81-cni-binary-copy\") pod \"multus-ldzbz\" (UID: \"b84e32ae-4897-4644-a656-6bd82012cc81\") " pod="openshift-multus/multus-ldzbz" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.581176 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/b84e32ae-4897-4644-a656-6bd82012cc81-multus-socket-dir-parent\") pod \"multus-ldzbz\" (UID: \"b84e32ae-4897-4644-a656-6bd82012cc81\") " pod="openshift-multus/multus-ldzbz" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.581186 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hrb87\" (UniqueName: \"kubernetes.io/projected/954085fa-d03a-4e88-9dfd-566257e3558d-kube-api-access-hrb87\") pod \"ovnkube-node-mzhnw\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.581203 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-run-ovn\") pod \"ovnkube-node-mzhnw\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.580706 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/b84e32ae-4897-4644-a656-6bd82012cc81-host-run-multus-certs\") pod \"multus-ldzbz\" (UID: \"b84e32ae-4897-4644-a656-6bd82012cc81\") " pod="openshift-multus/multus-ldzbz" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.581226 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b84e32ae-4897-4644-a656-6bd82012cc81-etc-kubernetes\") pod \"multus-ldzbz\" (UID: \"b84e32ae-4897-4644-a656-6bd82012cc81\") " pod="openshift-multus/multus-ldzbz" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.581254 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/b84e32ae-4897-4644-a656-6bd82012cc81-cnibin\") pod \"multus-ldzbz\" (UID: \"b84e32ae-4897-4644-a656-6bd82012cc81\") " pod="openshift-multus/multus-ldzbz" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.581283 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-host-slash\") pod \"ovnkube-node-mzhnw\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.581304 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-node-log\") pod \"ovnkube-node-mzhnw\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.581333 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-run-openvswitch\") pod \"ovnkube-node-mzhnw\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.581356 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-log-socket\") pod \"ovnkube-node-mzhnw\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.581381 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-mzhnw\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.581408 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-run-systemd\") pod \"ovnkube-node-mzhnw\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.581443 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/b84e32ae-4897-4644-a656-6bd82012cc81-host-run-k8s-cni-cncf-io\") pod \"multus-ldzbz\" (UID: \"b84e32ae-4897-4644-a656-6bd82012cc81\") " pod="openshift-multus/multus-ldzbz" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.581468 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g5pmc\" (UniqueName: \"kubernetes.io/projected/b84e32ae-4897-4644-a656-6bd82012cc81-kube-api-access-g5pmc\") pod \"multus-ldzbz\" (UID: \"b84e32ae-4897-4644-a656-6bd82012cc81\") " pod="openshift-multus/multus-ldzbz" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.581490 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-host-kubelet\") pod \"ovnkube-node-mzhnw\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.581528 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-systemd-units\") pod \"ovnkube-node-mzhnw\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.581553 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-host-cni-bin\") pod \"ovnkube-node-mzhnw\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.581575 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/b84e32ae-4897-4644-a656-6bd82012cc81-multus-daemon-config\") pod \"multus-ldzbz\" (UID: \"b84e32ae-4897-4644-a656-6bd82012cc81\") " pod="openshift-multus/multus-ldzbz" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.581598 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-etc-openvswitch\") pod \"ovnkube-node-mzhnw\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.581619 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/954085fa-d03a-4e88-9dfd-566257e3558d-env-overrides\") pod \"ovnkube-node-mzhnw\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.581658 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-var-lib-openvswitch\") pod \"ovnkube-node-mzhnw\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.581682 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/954085fa-d03a-4e88-9dfd-566257e3558d-ovnkube-config\") pod \"ovnkube-node-mzhnw\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.581705 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/954085fa-d03a-4e88-9dfd-566257e3558d-ovnkube-script-lib\") pod \"ovnkube-node-mzhnw\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.581731 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b84e32ae-4897-4644-a656-6bd82012cc81-host-var-lib-cni-bin\") pod \"multus-ldzbz\" (UID: \"b84e32ae-4897-4644-a656-6bd82012cc81\") " pod="openshift-multus/multus-ldzbz" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.581755 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/b84e32ae-4897-4644-a656-6bd82012cc81-hostroot\") pod \"multus-ldzbz\" (UID: \"b84e32ae-4897-4644-a656-6bd82012cc81\") " pod="openshift-multus/multus-ldzbz" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.581776 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-host-cni-netd\") pod \"ovnkube-node-mzhnw\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.581799 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/b84e32ae-4897-4644-a656-6bd82012cc81-cni-binary-copy\") pod \"multus-ldzbz\" (UID: \"b84e32ae-4897-4644-a656-6bd82012cc81\") " pod="openshift-multus/multus-ldzbz" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.581845 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-host-cni-netd\") pod \"ovnkube-node-mzhnw\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.581886 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-host-kubelet\") pod \"ovnkube-node-mzhnw\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.581918 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-systemd-units\") pod \"ovnkube-node-mzhnw\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.581949 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-host-cni-bin\") pod \"ovnkube-node-mzhnw\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.582017 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-var-lib-openvswitch\") pod \"ovnkube-node-mzhnw\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.582152 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b84e32ae-4897-4644-a656-6bd82012cc81-etc-kubernetes\") pod \"multus-ldzbz\" (UID: \"b84e32ae-4897-4644-a656-6bd82012cc81\") " pod="openshift-multus/multus-ldzbz" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.582208 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/b84e32ae-4897-4644-a656-6bd82012cc81-cnibin\") pod \"multus-ldzbz\" (UID: \"b84e32ae-4897-4644-a656-6bd82012cc81\") " pod="openshift-multus/multus-ldzbz" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.582231 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-host-slash\") pod \"ovnkube-node-mzhnw\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.582255 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-node-log\") pod \"ovnkube-node-mzhnw\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.582279 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-run-openvswitch\") pod \"ovnkube-node-mzhnw\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.582303 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-log-socket\") pod \"ovnkube-node-mzhnw\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.582327 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-mzhnw\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.582355 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-run-systemd\") pod \"ovnkube-node-mzhnw\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.582384 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/b84e32ae-4897-4644-a656-6bd82012cc81-host-run-k8s-cni-cncf-io\") pod \"multus-ldzbz\" (UID: \"b84e32ae-4897-4644-a656-6bd82012cc81\") " pod="openshift-multus/multus-ldzbz" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.582584 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/954085fa-d03a-4e88-9dfd-566257e3558d-env-overrides\") pod \"ovnkube-node-mzhnw\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.582625 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-etc-openvswitch\") pod \"ovnkube-node-mzhnw\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.582709 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b84e32ae-4897-4644-a656-6bd82012cc81-host-var-lib-cni-bin\") pod \"multus-ldzbz\" (UID: \"b84e32ae-4897-4644-a656-6bd82012cc81\") " pod="openshift-multus/multus-ldzbz" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.582796 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/954085fa-d03a-4e88-9dfd-566257e3558d-ovnkube-config\") pod \"ovnkube-node-mzhnw\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.582844 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/b84e32ae-4897-4644-a656-6bd82012cc81-hostroot\") pod \"multus-ldzbz\" (UID: \"b84e32ae-4897-4644-a656-6bd82012cc81\") " pod="openshift-multus/multus-ldzbz" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.583013 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/b84e32ae-4897-4644-a656-6bd82012cc81-multus-daemon-config\") pod \"multus-ldzbz\" (UID: \"b84e32ae-4897-4644-a656-6bd82012cc81\") " pod="openshift-multus/multus-ldzbz" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.583295 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/954085fa-d03a-4e88-9dfd-566257e3558d-ovnkube-script-lib\") pod \"ovnkube-node-mzhnw\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.583420 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b84e32ae-4897-4644-a656-6bd82012cc81-multus-cni-dir\") pod \"multus-ldzbz\" (UID: \"b84e32ae-4897-4644-a656-6bd82012cc81\") " pod="openshift-multus/multus-ldzbz" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.589952 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/954085fa-d03a-4e88-9dfd-566257e3558d-ovn-node-metrics-cert\") pod \"ovnkube-node-mzhnw\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.590024 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zlk96" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d55aa91-9cb9-4c95-bc27-6fec05684c10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4521b5baee45ffc9c7661c781c0b3de458dff325ad525dde283d46af71c0cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vkd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zlk96\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:31Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.606125 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hrb87\" (UniqueName: \"kubernetes.io/projected/954085fa-d03a-4e88-9dfd-566257e3558d-kube-api-access-hrb87\") pod \"ovnkube-node-mzhnw\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.610233 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d9e58df-1755-4422-8d04-211f98c7c011\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea762f627aa69fb88a9bd82545fa677e9ccfa97f99a5e92c0aa490e619c0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a06b811e5ec52ce984a2403ba2c98332ba60466d08b3580ad9192fac6caef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0385f6ed9597974267f7c45547f6b6725b4ee83a2cc0508ff2d1c1ce8753ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b915b565f7fbb3e4e856e7f1035a519e82bfb3023660c121f6e3813de353bb92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18465f3a6916c36457a305ba5d568c9077f34f1d000532ed518f237d100008b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:31Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.622141 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:31Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.635410 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ldzbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b84e32ae-4897-4644-a656-6bd82012cc81\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5pmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ldzbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:31Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.653050 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.664685 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"954085fa-d03a-4e88-9dfd-566257e3558d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzhnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:31Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.921271 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.942635 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.950297 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5f948b87-ac86-4de6-ad64-c2ef947f84d4-proxy-tls\") pod \"machine-config-daemon-2krbf\" (UID: \"5f948b87-ac86-4de6-ad64-c2ef947f84d4\") " pod="openshift-machine-config-operator/machine-config-daemon-2krbf" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.950506 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Nov 25 10:32:31 crc kubenswrapper[4821]: I1125 10:32:31.981667 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Nov 25 10:32:32 crc kubenswrapper[4821]: I1125 10:32:32.020618 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Nov 25 10:32:32 crc kubenswrapper[4821]: I1125 10:32:32.027429 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5f948b87-ac86-4de6-ad64-c2ef947f84d4-mcd-auth-proxy-config\") pod \"machine-config-daemon-2krbf\" (UID: \"5f948b87-ac86-4de6-ad64-c2ef947f84d4\") " pod="openshift-machine-config-operator/machine-config-daemon-2krbf" Nov 25 10:32:32 crc kubenswrapper[4821]: E1125 10:32:32.190693 4821 projected.go:288] Couldn't get configMap openshift-multus/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Nov 25 10:32:32 crc kubenswrapper[4821]: E1125 10:32:32.190737 4821 projected.go:194] Error preparing data for projected volume kube-api-access-hxd4d for pod openshift-multus/multus-additional-cni-plugins-qppnw: failed to sync configmap cache: timed out waiting for the condition Nov 25 10:32:32 crc kubenswrapper[4821]: E1125 10:32:32.190797 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/357491b8-cd9c-486e-9898-5ba33678513d-kube-api-access-hxd4d podName:357491b8-cd9c-486e-9898-5ba33678513d nodeName:}" failed. No retries permitted until 2025-11-25 10:32:32.690781137 +0000 UTC m=+23.227100984 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-hxd4d" (UniqueName: "kubernetes.io/projected/357491b8-cd9c-486e-9898-5ba33678513d-kube-api-access-hxd4d") pod "multus-additional-cni-plugins-qppnw" (UID: "357491b8-cd9c-486e-9898-5ba33678513d") : failed to sync configmap cache: timed out waiting for the condition Nov 25 10:32:32 crc kubenswrapper[4821]: I1125 10:32:32.203442 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 10:32:32 crc kubenswrapper[4821]: I1125 10:32:32.206318 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 10:32:32 crc kubenswrapper[4821]: I1125 10:32:32.210201 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Nov 25 10:32:32 crc kubenswrapper[4821]: I1125 10:32:32.221485 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"384d8ba58264ab89796d3b53afb0c4c1be33cb4fba719a8beef4afc2be7d66a2"} Nov 25 10:32:32 crc kubenswrapper[4821]: I1125 10:32:32.223077 4821 generic.go:334] "Generic (PLEG): container finished" podID="954085fa-d03a-4e88-9dfd-566257e3558d" containerID="932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3" exitCode=0 Nov 25 10:32:32 crc kubenswrapper[4821]: I1125 10:32:32.223193 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" event={"ID":"954085fa-d03a-4e88-9dfd-566257e3558d","Type":"ContainerDied","Data":"932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3"} Nov 25 10:32:32 crc kubenswrapper[4821]: I1125 10:32:32.223222 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" event={"ID":"954085fa-d03a-4e88-9dfd-566257e3558d","Type":"ContainerStarted","Data":"0ebf341b9565ca43636f838f682a2a1805f094bdccd2c8362fd1dd4236b179ad"} Nov 25 10:32:32 crc kubenswrapper[4821]: I1125 10:32:32.223393 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d9e58df-1755-4422-8d04-211f98c7c011\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea762f627aa69fb88a9bd82545fa677e9ccfa97f99a5e92c0aa490e619c0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a06b811e5ec52ce984a2403ba2c98332ba60466d08b3580ad9192fac6caef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0385f6ed9597974267f7c45547f6b6725b4ee83a2cc0508ff2d1c1ce8753ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b915b565f7fbb3e4e856e7f1035a519e82bfb3023660c121f6e3813de353bb92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18465f3a6916c36457a305ba5d568c9077f34f1d000532ed518f237d100008b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:32Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:32 crc kubenswrapper[4821]: I1125 10:32:32.235404 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:32Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:32 crc kubenswrapper[4821]: I1125 10:32:32.251127 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ldzbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b84e32ae-4897-4644-a656-6bd82012cc81\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5pmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ldzbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:32Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:32 crc kubenswrapper[4821]: I1125 10:32:32.268511 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"954085fa-d03a-4e88-9dfd-566257e3558d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzhnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:32Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:32 crc kubenswrapper[4821]: I1125 10:32:32.282069 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62fcb632-f12a-46e2-bb7d-a5311f291ee7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28d43531927e75c3745f100940f7f6bf79fd21026c1309194df8cd7a39361b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69633332d03d2551ff4abb253d358de398eab18486b92db630ce06c0447fb9cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ccb78d9a51ac83ee208c2451685a2512de9deaf742a663bb8c892aa59b76162\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27826e155a9352b1e1866e83fdab683dbcd749cc04b6a4da2db25ecabe3d9a52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://991252e5f56f128827a5054b8e228284bc0266166582cbaedaadc53ca2f5b401\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:32Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:32 crc kubenswrapper[4821]: I1125 10:32:32.296001 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:32Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:32 crc kubenswrapper[4821]: I1125 10:32:32.305473 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f948b87-ac86-4de6-ad64-c2ef947f84d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2krbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:32Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:32 crc kubenswrapper[4821]: I1125 10:32:32.317879 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qppnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"357491b8-cd9c-486e-9898-5ba33678513d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qppnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:32Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:32 crc kubenswrapper[4821]: I1125 10:32:32.330215 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06f949936bac99cfbf89d720c9549c074351d9077bd963fa024d87004c6be23e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:32Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:32 crc kubenswrapper[4821]: I1125 10:32:32.340329 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:32Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:32 crc kubenswrapper[4821]: I1125 10:32:32.351804 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:32Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:32 crc kubenswrapper[4821]: I1125 10:32:32.367694 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17b8714be3bd121e06641873e003f08560b1f71e56d6ea936966df7d4896eb03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85731e6b6dbe1777bc7161a010f39f79523ec4b08d71d5d8a975573aa7fe4957\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:32Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:32 crc kubenswrapper[4821]: I1125 10:32:32.382927 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zlk96" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d55aa91-9cb9-4c95-bc27-6fec05684c10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4521b5baee45ffc9c7661c781c0b3de458dff325ad525dde283d46af71c0cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vkd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zlk96\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:32Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:32 crc kubenswrapper[4821]: I1125 10:32:32.398260 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:32Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:32 crc kubenswrapper[4821]: I1125 10:32:32.411934 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17b8714be3bd121e06641873e003f08560b1f71e56d6ea936966df7d4896eb03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85731e6b6dbe1777bc7161a010f39f79523ec4b08d71d5d8a975573aa7fe4957\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:32Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:32 crc kubenswrapper[4821]: I1125 10:32:32.423069 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zlk96" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d55aa91-9cb9-4c95-bc27-6fec05684c10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4521b5baee45ffc9c7661c781c0b3de458dff325ad525dde283d46af71c0cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vkd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zlk96\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:32Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:32 crc kubenswrapper[4821]: I1125 10:32:32.443555 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d9e58df-1755-4422-8d04-211f98c7c011\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea762f627aa69fb88a9bd82545fa677e9ccfa97f99a5e92c0aa490e619c0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a06b811e5ec52ce984a2403ba2c98332ba60466d08b3580ad9192fac6caef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0385f6ed9597974267f7c45547f6b6725b4ee83a2cc0508ff2d1c1ce8753ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b915b565f7fbb3e4e856e7f1035a519e82bfb3023660c121f6e3813de353bb92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18465f3a6916c36457a305ba5d568c9077f34f1d000532ed518f237d100008b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:32Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:32 crc kubenswrapper[4821]: I1125 10:32:32.451365 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Nov 25 10:32:32 crc kubenswrapper[4821]: I1125 10:32:32.458897 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25595ccd-762d-414a-9916-4fe38d478b07\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d4b2d8db4d13392d9789e161fd4b8f6eef983b79128ee584d72e4c06565bc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf780f791771ea43a8c8176d02fdfb5204406e023207299369ae46ceb342baaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0251519fd4262f2ea5c64959db49433afdafe36b7ed2404d660eeb7c365d2746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad5c205e0eeecfd43cbdf0eb1a334e3e8faa5e9a9e8dd202c397d5cbb36e081e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:32Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:32 crc kubenswrapper[4821]: I1125 10:32:32.463862 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qq48b\" (UniqueName: \"kubernetes.io/projected/5f948b87-ac86-4de6-ad64-c2ef947f84d4-kube-api-access-qq48b\") pod \"machine-config-daemon-2krbf\" (UID: \"5f948b87-ac86-4de6-ad64-c2ef947f84d4\") " pod="openshift-machine-config-operator/machine-config-daemon-2krbf" Nov 25 10:32:32 crc kubenswrapper[4821]: I1125 10:32:32.469494 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:32Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:32 crc kubenswrapper[4821]: I1125 10:32:32.473033 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" Nov 25 10:32:32 crc kubenswrapper[4821]: I1125 10:32:32.484379 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ldzbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b84e32ae-4897-4644-a656-6bd82012cc81\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5pmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ldzbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:32Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:32 crc kubenswrapper[4821]: I1125 10:32:32.505879 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"954085fa-d03a-4e88-9dfd-566257e3558d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzhnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:32Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:32 crc kubenswrapper[4821]: I1125 10:32:32.518301 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62fcb632-f12a-46e2-bb7d-a5311f291ee7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28d43531927e75c3745f100940f7f6bf79fd21026c1309194df8cd7a39361b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69633332d03d2551ff4abb253d358de398eab18486b92db630ce06c0447fb9cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ccb78d9a51ac83ee208c2451685a2512de9deaf742a663bb8c892aa59b76162\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27826e155a9352b1e1866e83fdab683dbcd749cc04b6a4da2db25ecabe3d9a52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://991252e5f56f128827a5054b8e228284bc0266166582cbaedaadc53ca2f5b401\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:32Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:32 crc kubenswrapper[4821]: I1125 10:32:32.532446 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:32Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:32 crc kubenswrapper[4821]: I1125 10:32:32.539217 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Nov 25 10:32:32 crc kubenswrapper[4821]: I1125 10:32:32.544546 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f948b87-ac86-4de6-ad64-c2ef947f84d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2krbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:32Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:32 crc kubenswrapper[4821]: I1125 10:32:32.550093 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g5pmc\" (UniqueName: \"kubernetes.io/projected/b84e32ae-4897-4644-a656-6bd82012cc81-kube-api-access-g5pmc\") pod \"multus-ldzbz\" (UID: \"b84e32ae-4897-4644-a656-6bd82012cc81\") " pod="openshift-multus/multus-ldzbz" Nov 25 10:32:32 crc kubenswrapper[4821]: I1125 10:32:32.558374 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qppnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"357491b8-cd9c-486e-9898-5ba33678513d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qppnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:32Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:32 crc kubenswrapper[4821]: I1125 10:32:32.572522 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06f949936bac99cfbf89d720c9549c074351d9077bd963fa024d87004c6be23e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:32Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:32 crc kubenswrapper[4821]: I1125 10:32:32.584822 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384d8ba58264ab89796d3b53afb0c4c1be33cb4fba719a8beef4afc2be7d66a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:32Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:32 crc kubenswrapper[4821]: I1125 10:32:32.695514 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxd4d\" (UniqueName: \"kubernetes.io/projected/357491b8-cd9c-486e-9898-5ba33678513d-kube-api-access-hxd4d\") pod \"multus-additional-cni-plugins-qppnw\" (UID: \"357491b8-cd9c-486e-9898-5ba33678513d\") " pod="openshift-multus/multus-additional-cni-plugins-qppnw" Nov 25 10:32:32 crc kubenswrapper[4821]: I1125 10:32:32.699986 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxd4d\" (UniqueName: \"kubernetes.io/projected/357491b8-cd9c-486e-9898-5ba33678513d-kube-api-access-hxd4d\") pod \"multus-additional-cni-plugins-qppnw\" (UID: \"357491b8-cd9c-486e-9898-5ba33678513d\") " pod="openshift-multus/multus-additional-cni-plugins-qppnw" Nov 25 10:32:32 crc kubenswrapper[4821]: I1125 10:32:32.782413 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-qppnw" Nov 25 10:32:32 crc kubenswrapper[4821]: I1125 10:32:32.795924 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:32:32 crc kubenswrapper[4821]: E1125 10:32:32.796128 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:32:36.796086456 +0000 UTC m=+27.332406333 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:32:32 crc kubenswrapper[4821]: I1125 10:32:32.796211 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:32:32 crc kubenswrapper[4821]: E1125 10:32:32.796486 4821 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 10:32:32 crc kubenswrapper[4821]: I1125 10:32:32.796523 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:32:32 crc kubenswrapper[4821]: E1125 10:32:32.796635 4821 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 10:32:32 crc kubenswrapper[4821]: E1125 10:32:32.796692 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 10:32:36.796582877 +0000 UTC m=+27.332902724 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 10:32:32 crc kubenswrapper[4821]: E1125 10:32:32.796794 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 10:32:36.796784192 +0000 UTC m=+27.333104039 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 10:32:32 crc kubenswrapper[4821]: I1125 10:32:32.820838 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-tgdlt"] Nov 25 10:32:32 crc kubenswrapper[4821]: I1125 10:32:32.821289 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-tgdlt" Nov 25 10:32:32 crc kubenswrapper[4821]: I1125 10:32:32.823088 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Nov 25 10:32:32 crc kubenswrapper[4821]: I1125 10:32:32.823109 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Nov 25 10:32:32 crc kubenswrapper[4821]: I1125 10:32:32.823708 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Nov 25 10:32:32 crc kubenswrapper[4821]: I1125 10:32:32.823898 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Nov 25 10:32:32 crc kubenswrapper[4821]: I1125 10:32:32.837412 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:32Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:32 crc kubenswrapper[4821]: I1125 10:32:32.847320 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-ldzbz" Nov 25 10:32:32 crc kubenswrapper[4821]: I1125 10:32:32.850392 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17b8714be3bd121e06641873e003f08560b1f71e56d6ea936966df7d4896eb03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85731e6b6dbe1777bc7161a010f39f79523ec4b08d71d5d8a975573aa7fe4957\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:32Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:32 crc kubenswrapper[4821]: W1125 10:32:32.860489 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb84e32ae_4897_4644_a656_6bd82012cc81.slice/crio-bce4c37b6883a61d6af98b231fba8c8567ac4ea72dc8e672d4a3c1c006f329a9 WatchSource:0}: Error finding container bce4c37b6883a61d6af98b231fba8c8567ac4ea72dc8e672d4a3c1c006f329a9: Status 404 returned error can't find the container with id bce4c37b6883a61d6af98b231fba8c8567ac4ea72dc8e672d4a3c1c006f329a9 Nov 25 10:32:32 crc kubenswrapper[4821]: I1125 10:32:32.860844 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zlk96" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d55aa91-9cb9-4c95-bc27-6fec05684c10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4521b5baee45ffc9c7661c781c0b3de458dff325ad525dde283d46af71c0cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vkd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zlk96\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:32Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:32 crc kubenswrapper[4821]: I1125 10:32:32.872465 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:32Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:32 crc kubenswrapper[4821]: I1125 10:32:32.886337 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ldzbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b84e32ae-4897-4644-a656-6bd82012cc81\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5pmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ldzbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:32Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:32 crc kubenswrapper[4821]: I1125 10:32:32.897598 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:32:32 crc kubenswrapper[4821]: I1125 10:32:32.897634 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:32:32 crc kubenswrapper[4821]: E1125 10:32:32.897779 4821 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 10:32:32 crc kubenswrapper[4821]: E1125 10:32:32.897792 4821 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 10:32:32 crc kubenswrapper[4821]: E1125 10:32:32.897805 4821 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 10:32:32 crc kubenswrapper[4821]: E1125 10:32:32.897807 4821 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 10:32:32 crc kubenswrapper[4821]: E1125 10:32:32.897817 4821 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 10:32:32 crc kubenswrapper[4821]: E1125 10:32:32.897819 4821 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 10:32:32 crc kubenswrapper[4821]: E1125 10:32:32.897863 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 10:32:36.89784857 +0000 UTC m=+27.434168417 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 10:32:32 crc kubenswrapper[4821]: E1125 10:32:32.897883 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 10:32:36.897875631 +0000 UTC m=+27.434195478 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 10:32:32 crc kubenswrapper[4821]: I1125 10:32:32.902345 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"954085fa-d03a-4e88-9dfd-566257e3558d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzhnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:32Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:32 crc kubenswrapper[4821]: I1125 10:32:32.923986 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d9e58df-1755-4422-8d04-211f98c7c011\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea762f627aa69fb88a9bd82545fa677e9ccfa97f99a5e92c0aa490e619c0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a06b811e5ec52ce984a2403ba2c98332ba60466d08b3580ad9192fac6caef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0385f6ed9597974267f7c45547f6b6725b4ee83a2cc0508ff2d1c1ce8753ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b915b565f7fbb3e4e856e7f1035a519e82bfb3023660c121f6e3813de353bb92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18465f3a6916c36457a305ba5d568c9077f34f1d000532ed518f237d100008b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:32Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:32 crc kubenswrapper[4821]: I1125 10:32:32.937073 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25595ccd-762d-414a-9916-4fe38d478b07\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d4b2d8db4d13392d9789e161fd4b8f6eef983b79128ee584d72e4c06565bc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf780f791771ea43a8c8176d02fdfb5204406e023207299369ae46ceb342baaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0251519fd4262f2ea5c64959db49433afdafe36b7ed2404d660eeb7c365d2746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad5c205e0eeecfd43cbdf0eb1a334e3e8faa5e9a9e8dd202c397d5cbb36e081e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:32Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:32 crc kubenswrapper[4821]: I1125 10:32:32.949116 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f948b87-ac86-4de6-ad64-c2ef947f84d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2krbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:32Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:32 crc kubenswrapper[4821]: I1125 10:32:32.964761 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qppnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"357491b8-cd9c-486e-9898-5ba33678513d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qppnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:32Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:32 crc kubenswrapper[4821]: I1125 10:32:32.992136 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tgdlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2336955b-a703-4db4-8d0f-d08088ed59b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vjfp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tgdlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:32Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:32 crc kubenswrapper[4821]: I1125 10:32:32.998750 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/2336955b-a703-4db4-8d0f-d08088ed59b5-serviceca\") pod \"node-ca-tgdlt\" (UID: \"2336955b-a703-4db4-8d0f-d08088ed59b5\") " pod="openshift-image-registry/node-ca-tgdlt" Nov 25 10:32:32 crc kubenswrapper[4821]: I1125 10:32:32.998784 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2336955b-a703-4db4-8d0f-d08088ed59b5-host\") pod \"node-ca-tgdlt\" (UID: \"2336955b-a703-4db4-8d0f-d08088ed59b5\") " pod="openshift-image-registry/node-ca-tgdlt" Nov 25 10:32:32 crc kubenswrapper[4821]: I1125 10:32:32.998819 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjfp8\" (UniqueName: \"kubernetes.io/projected/2336955b-a703-4db4-8d0f-d08088ed59b5-kube-api-access-vjfp8\") pod \"node-ca-tgdlt\" (UID: \"2336955b-a703-4db4-8d0f-d08088ed59b5\") " pod="openshift-image-registry/node-ca-tgdlt" Nov 25 10:32:33 crc kubenswrapper[4821]: I1125 10:32:33.034726 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62fcb632-f12a-46e2-bb7d-a5311f291ee7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28d43531927e75c3745f100940f7f6bf79fd21026c1309194df8cd7a39361b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69633332d03d2551ff4abb253d358de398eab18486b92db630ce06c0447fb9cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ccb78d9a51ac83ee208c2451685a2512de9deaf742a663bb8c892aa59b76162\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27826e155a9352b1e1866e83fdab683dbcd749cc04b6a4da2db25ecabe3d9a52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://991252e5f56f128827a5054b8e228284bc0266166582cbaedaadc53ca2f5b401\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:33Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:33 crc kubenswrapper[4821]: I1125 10:32:33.073430 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:33Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:33 crc kubenswrapper[4821]: I1125 10:32:33.099317 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjfp8\" (UniqueName: \"kubernetes.io/projected/2336955b-a703-4db4-8d0f-d08088ed59b5-kube-api-access-vjfp8\") pod \"node-ca-tgdlt\" (UID: \"2336955b-a703-4db4-8d0f-d08088ed59b5\") " pod="openshift-image-registry/node-ca-tgdlt" Nov 25 10:32:33 crc kubenswrapper[4821]: I1125 10:32:33.099374 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/2336955b-a703-4db4-8d0f-d08088ed59b5-serviceca\") pod \"node-ca-tgdlt\" (UID: \"2336955b-a703-4db4-8d0f-d08088ed59b5\") " pod="openshift-image-registry/node-ca-tgdlt" Nov 25 10:32:33 crc kubenswrapper[4821]: I1125 10:32:33.099394 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2336955b-a703-4db4-8d0f-d08088ed59b5-host\") pod \"node-ca-tgdlt\" (UID: \"2336955b-a703-4db4-8d0f-d08088ed59b5\") " pod="openshift-image-registry/node-ca-tgdlt" Nov 25 10:32:33 crc kubenswrapper[4821]: I1125 10:32:33.099489 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2336955b-a703-4db4-8d0f-d08088ed59b5-host\") pod \"node-ca-tgdlt\" (UID: \"2336955b-a703-4db4-8d0f-d08088ed59b5\") " pod="openshift-image-registry/node-ca-tgdlt" Nov 25 10:32:33 crc kubenswrapper[4821]: I1125 10:32:33.100530 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/2336955b-a703-4db4-8d0f-d08088ed59b5-serviceca\") pod \"node-ca-tgdlt\" (UID: \"2336955b-a703-4db4-8d0f-d08088ed59b5\") " pod="openshift-image-registry/node-ca-tgdlt" Nov 25 10:32:33 crc kubenswrapper[4821]: I1125 10:32:33.113845 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:32:33 crc kubenswrapper[4821]: I1125 10:32:33.113887 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:32:33 crc kubenswrapper[4821]: I1125 10:32:33.113856 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:32:33 crc kubenswrapper[4821]: E1125 10:32:33.113983 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:32:33 crc kubenswrapper[4821]: E1125 10:32:33.114101 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:32:33 crc kubenswrapper[4821]: E1125 10:32:33.114222 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:32:33 crc kubenswrapper[4821]: I1125 10:32:33.114810 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06f949936bac99cfbf89d720c9549c074351d9077bd963fa024d87004c6be23e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:33Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:33 crc kubenswrapper[4821]: I1125 10:32:33.138887 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjfp8\" (UniqueName: \"kubernetes.io/projected/2336955b-a703-4db4-8d0f-d08088ed59b5-kube-api-access-vjfp8\") pod \"node-ca-tgdlt\" (UID: \"2336955b-a703-4db4-8d0f-d08088ed59b5\") " pod="openshift-image-registry/node-ca-tgdlt" Nov 25 10:32:33 crc kubenswrapper[4821]: I1125 10:32:33.172251 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384d8ba58264ab89796d3b53afb0c4c1be33cb4fba719a8beef4afc2be7d66a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:33Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:33 crc kubenswrapper[4821]: I1125 10:32:33.232379 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" event={"ID":"954085fa-d03a-4e88-9dfd-566257e3558d","Type":"ContainerStarted","Data":"e9d7254196b3702e48083ae98e829b73c539adf69ed5b2bbdc62189d09ae1762"} Nov 25 10:32:33 crc kubenswrapper[4821]: I1125 10:32:33.232428 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" event={"ID":"954085fa-d03a-4e88-9dfd-566257e3558d","Type":"ContainerStarted","Data":"d727430108977574365f0e9c04b44e03307bd7ff89cdbb1e48a6e916f504d6b0"} Nov 25 10:32:33 crc kubenswrapper[4821]: I1125 10:32:33.232441 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" event={"ID":"954085fa-d03a-4e88-9dfd-566257e3558d","Type":"ContainerStarted","Data":"1a14a078ba7d2aed95992b36502008f315f0d3a634b64ed49da70398b2562d44"} Nov 25 10:32:33 crc kubenswrapper[4821]: I1125 10:32:33.232451 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" event={"ID":"954085fa-d03a-4e88-9dfd-566257e3558d","Type":"ContainerStarted","Data":"2f00746930c4588e548f0e2b60d52bf448aff7401b676a7855cb0f4cce00d329"} Nov 25 10:32:33 crc kubenswrapper[4821]: I1125 10:32:33.232463 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" event={"ID":"954085fa-d03a-4e88-9dfd-566257e3558d","Type":"ContainerStarted","Data":"94c7949f077b1ca2b310bedb090aa72fc4bd7e8f2c8bf438096edfca8568f36e"} Nov 25 10:32:33 crc kubenswrapper[4821]: I1125 10:32:33.232475 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" event={"ID":"954085fa-d03a-4e88-9dfd-566257e3558d","Type":"ContainerStarted","Data":"ed2e7cad7bb24b5b56a1295098a6a7c230c8f9b9c29a88ef653ae8b6ac31667d"} Nov 25 10:32:33 crc kubenswrapper[4821]: I1125 10:32:33.233495 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-ldzbz" event={"ID":"b84e32ae-4897-4644-a656-6bd82012cc81","Type":"ContainerStarted","Data":"8ef433744761f26e8e1f6c0db123821a11923ffd774830744285cd4178d89533"} Nov 25 10:32:33 crc kubenswrapper[4821]: I1125 10:32:33.233518 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-ldzbz" event={"ID":"b84e32ae-4897-4644-a656-6bd82012cc81","Type":"ContainerStarted","Data":"bce4c37b6883a61d6af98b231fba8c8567ac4ea72dc8e672d4a3c1c006f329a9"} Nov 25 10:32:33 crc kubenswrapper[4821]: I1125 10:32:33.235506 4821 generic.go:334] "Generic (PLEG): container finished" podID="357491b8-cd9c-486e-9898-5ba33678513d" containerID="dd9c047c5ce1d896f803ff6cafa9bbf3a18a391b44539e1f1312e08c89e51cb8" exitCode=0 Nov 25 10:32:33 crc kubenswrapper[4821]: I1125 10:32:33.235570 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-qppnw" event={"ID":"357491b8-cd9c-486e-9898-5ba33678513d","Type":"ContainerDied","Data":"dd9c047c5ce1d896f803ff6cafa9bbf3a18a391b44539e1f1312e08c89e51cb8"} Nov 25 10:32:33 crc kubenswrapper[4821]: I1125 10:32:33.235591 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-qppnw" event={"ID":"357491b8-cd9c-486e-9898-5ba33678513d","Type":"ContainerStarted","Data":"f269e1ea60cd466695a7f11d8dde0d1b73a0ef1cd33f37e75779a0cd4532d244"} Nov 25 10:32:33 crc kubenswrapper[4821]: I1125 10:32:33.237641 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" event={"ID":"5f948b87-ac86-4de6-ad64-c2ef947f84d4","Type":"ContainerStarted","Data":"4482f73a45bc82bd55770c3fd788becacec94893fed1c305fed1c81c753d43aa"} Nov 25 10:32:33 crc kubenswrapper[4821]: I1125 10:32:33.237735 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" event={"ID":"5f948b87-ac86-4de6-ad64-c2ef947f84d4","Type":"ContainerStarted","Data":"9ba68bd093be94d0d40f3851638db5d26bdfc52d175b1ac78afb91dda7affcfe"} Nov 25 10:32:33 crc kubenswrapper[4821]: I1125 10:32:33.237802 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" event={"ID":"5f948b87-ac86-4de6-ad64-c2ef947f84d4","Type":"ContainerStarted","Data":"a2eda4f119cdf66555182ba6e5ca09d0bff313955a713be8269bd36239793aa8"} Nov 25 10:32:33 crc kubenswrapper[4821]: I1125 10:32:33.245173 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tgdlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2336955b-a703-4db4-8d0f-d08088ed59b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vjfp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tgdlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:33Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:33 crc kubenswrapper[4821]: I1125 10:32:33.259729 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62fcb632-f12a-46e2-bb7d-a5311f291ee7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28d43531927e75c3745f100940f7f6bf79fd21026c1309194df8cd7a39361b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69633332d03d2551ff4abb253d358de398eab18486b92db630ce06c0447fb9cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ccb78d9a51ac83ee208c2451685a2512de9deaf742a663bb8c892aa59b76162\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27826e155a9352b1e1866e83fdab683dbcd749cc04b6a4da2db25ecabe3d9a52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://991252e5f56f128827a5054b8e228284bc0266166582cbaedaadc53ca2f5b401\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:33Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:33 crc kubenswrapper[4821]: I1125 10:32:33.293736 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:33Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:33 crc kubenswrapper[4821]: I1125 10:32:33.332828 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f948b87-ac86-4de6-ad64-c2ef947f84d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2krbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:33Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:33 crc kubenswrapper[4821]: I1125 10:32:33.375751 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qppnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"357491b8-cd9c-486e-9898-5ba33678513d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qppnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:33Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:33 crc kubenswrapper[4821]: I1125 10:32:33.414692 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06f949936bac99cfbf89d720c9549c074351d9077bd963fa024d87004c6be23e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:33Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:33 crc kubenswrapper[4821]: I1125 10:32:33.434107 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-tgdlt" Nov 25 10:32:33 crc kubenswrapper[4821]: W1125 10:32:33.444571 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2336955b_a703_4db4_8d0f_d08088ed59b5.slice/crio-1f40221b7d3776c5023818e0158e889f60504c6a272d353067f0c9ed79719644 WatchSource:0}: Error finding container 1f40221b7d3776c5023818e0158e889f60504c6a272d353067f0c9ed79719644: Status 404 returned error can't find the container with id 1f40221b7d3776c5023818e0158e889f60504c6a272d353067f0c9ed79719644 Nov 25 10:32:33 crc kubenswrapper[4821]: I1125 10:32:33.455621 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384d8ba58264ab89796d3b53afb0c4c1be33cb4fba719a8beef4afc2be7d66a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:33Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:33 crc kubenswrapper[4821]: I1125 10:32:33.492831 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:33Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:33 crc kubenswrapper[4821]: I1125 10:32:33.534652 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17b8714be3bd121e06641873e003f08560b1f71e56d6ea936966df7d4896eb03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85731e6b6dbe1777bc7161a010f39f79523ec4b08d71d5d8a975573aa7fe4957\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:33Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:33 crc kubenswrapper[4821]: I1125 10:32:33.571100 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zlk96" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d55aa91-9cb9-4c95-bc27-6fec05684c10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4521b5baee45ffc9c7661c781c0b3de458dff325ad525dde283d46af71c0cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vkd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zlk96\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:33Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:33 crc kubenswrapper[4821]: I1125 10:32:33.653830 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"954085fa-d03a-4e88-9dfd-566257e3558d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzhnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:33Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:33 crc kubenswrapper[4821]: I1125 10:32:33.687113 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d9e58df-1755-4422-8d04-211f98c7c011\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea762f627aa69fb88a9bd82545fa677e9ccfa97f99a5e92c0aa490e619c0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a06b811e5ec52ce984a2403ba2c98332ba60466d08b3580ad9192fac6caef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0385f6ed9597974267f7c45547f6b6725b4ee83a2cc0508ff2d1c1ce8753ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b915b565f7fbb3e4e856e7f1035a519e82bfb3023660c121f6e3813de353bb92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18465f3a6916c36457a305ba5d568c9077f34f1d000532ed518f237d100008b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:33Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:33 crc kubenswrapper[4821]: I1125 10:32:33.699695 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25595ccd-762d-414a-9916-4fe38d478b07\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d4b2d8db4d13392d9789e161fd4b8f6eef983b79128ee584d72e4c06565bc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf780f791771ea43a8c8176d02fdfb5204406e023207299369ae46ceb342baaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0251519fd4262f2ea5c64959db49433afdafe36b7ed2404d660eeb7c365d2746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad5c205e0eeecfd43cbdf0eb1a334e3e8faa5e9a9e8dd202c397d5cbb36e081e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:33Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:33 crc kubenswrapper[4821]: I1125 10:32:33.734478 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:33Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:33 crc kubenswrapper[4821]: I1125 10:32:33.773128 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ldzbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b84e32ae-4897-4644-a656-6bd82012cc81\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ef433744761f26e8e1f6c0db123821a11923ffd774830744285cd4178d89533\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5pmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ldzbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:33Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:33 crc kubenswrapper[4821]: I1125 10:32:33.818921 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"954085fa-d03a-4e88-9dfd-566257e3558d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzhnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:33Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:33 crc kubenswrapper[4821]: I1125 10:32:33.859545 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d9e58df-1755-4422-8d04-211f98c7c011\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea762f627aa69fb88a9bd82545fa677e9ccfa97f99a5e92c0aa490e619c0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a06b811e5ec52ce984a2403ba2c98332ba60466d08b3580ad9192fac6caef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0385f6ed9597974267f7c45547f6b6725b4ee83a2cc0508ff2d1c1ce8753ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b915b565f7fbb3e4e856e7f1035a519e82bfb3023660c121f6e3813de353bb92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18465f3a6916c36457a305ba5d568c9077f34f1d000532ed518f237d100008b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:33Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:33 crc kubenswrapper[4821]: I1125 10:32:33.893457 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25595ccd-762d-414a-9916-4fe38d478b07\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d4b2d8db4d13392d9789e161fd4b8f6eef983b79128ee584d72e4c06565bc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf780f791771ea43a8c8176d02fdfb5204406e023207299369ae46ceb342baaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0251519fd4262f2ea5c64959db49433afdafe36b7ed2404d660eeb7c365d2746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad5c205e0eeecfd43cbdf0eb1a334e3e8faa5e9a9e8dd202c397d5cbb36e081e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:33Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:33 crc kubenswrapper[4821]: I1125 10:32:33.933840 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:33Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:33 crc kubenswrapper[4821]: I1125 10:32:33.972040 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ldzbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b84e32ae-4897-4644-a656-6bd82012cc81\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ef433744761f26e8e1f6c0db123821a11923ffd774830744285cd4178d89533\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5pmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ldzbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:33Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:34 crc kubenswrapper[4821]: I1125 10:32:34.011211 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tgdlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2336955b-a703-4db4-8d0f-d08088ed59b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vjfp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tgdlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:34Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:34 crc kubenswrapper[4821]: I1125 10:32:34.054406 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62fcb632-f12a-46e2-bb7d-a5311f291ee7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28d43531927e75c3745f100940f7f6bf79fd21026c1309194df8cd7a39361b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69633332d03d2551ff4abb253d358de398eab18486b92db630ce06c0447fb9cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ccb78d9a51ac83ee208c2451685a2512de9deaf742a663bb8c892aa59b76162\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27826e155a9352b1e1866e83fdab683dbcd749cc04b6a4da2db25ecabe3d9a52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://991252e5f56f128827a5054b8e228284bc0266166582cbaedaadc53ca2f5b401\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:34Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:34 crc kubenswrapper[4821]: I1125 10:32:34.092961 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:34Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:34 crc kubenswrapper[4821]: I1125 10:32:34.133403 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f948b87-ac86-4de6-ad64-c2ef947f84d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4482f73a45bc82bd55770c3fd788becacec94893fed1c305fed1c81c753d43aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ba68bd093be94d0d40f3851638db5d26bdfc52d175b1ac78afb91dda7affcfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2krbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:34Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:34 crc kubenswrapper[4821]: I1125 10:32:34.175976 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qppnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"357491b8-cd9c-486e-9898-5ba33678513d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd9c047c5ce1d896f803ff6cafa9bbf3a18a391b44539e1f1312e08c89e51cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd9c047c5ce1d896f803ff6cafa9bbf3a18a391b44539e1f1312e08c89e51cb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qppnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:34Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:34 crc kubenswrapper[4821]: I1125 10:32:34.213892 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06f949936bac99cfbf89d720c9549c074351d9077bd963fa024d87004c6be23e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:34Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:34 crc kubenswrapper[4821]: I1125 10:32:34.243112 4821 generic.go:334] "Generic (PLEG): container finished" podID="357491b8-cd9c-486e-9898-5ba33678513d" containerID="565a6db63fb69abbba81ffdf6085797900858d263cd6affcb0b5f195474c2516" exitCode=0 Nov 25 10:32:34 crc kubenswrapper[4821]: I1125 10:32:34.243472 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-qppnw" event={"ID":"357491b8-cd9c-486e-9898-5ba33678513d","Type":"ContainerDied","Data":"565a6db63fb69abbba81ffdf6085797900858d263cd6affcb0b5f195474c2516"} Nov 25 10:32:34 crc kubenswrapper[4821]: I1125 10:32:34.246918 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-tgdlt" event={"ID":"2336955b-a703-4db4-8d0f-d08088ed59b5","Type":"ContainerStarted","Data":"8cb0362942bfac635ee6292501b1b851772c01d75dd658eef4b17669a716de72"} Nov 25 10:32:34 crc kubenswrapper[4821]: I1125 10:32:34.246954 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-tgdlt" event={"ID":"2336955b-a703-4db4-8d0f-d08088ed59b5","Type":"ContainerStarted","Data":"1f40221b7d3776c5023818e0158e889f60504c6a272d353067f0c9ed79719644"} Nov 25 10:32:34 crc kubenswrapper[4821]: I1125 10:32:34.253485 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384d8ba58264ab89796d3b53afb0c4c1be33cb4fba719a8beef4afc2be7d66a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:34Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:34 crc kubenswrapper[4821]: I1125 10:32:34.292490 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:34Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:34 crc kubenswrapper[4821]: I1125 10:32:34.334681 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17b8714be3bd121e06641873e003f08560b1f71e56d6ea936966df7d4896eb03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85731e6b6dbe1777bc7161a010f39f79523ec4b08d71d5d8a975573aa7fe4957\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:34Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:34 crc kubenswrapper[4821]: I1125 10:32:34.371265 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zlk96" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d55aa91-9cb9-4c95-bc27-6fec05684c10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4521b5baee45ffc9c7661c781c0b3de458dff325ad525dde283d46af71c0cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vkd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zlk96\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:34Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:34 crc kubenswrapper[4821]: I1125 10:32:34.411289 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zlk96" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d55aa91-9cb9-4c95-bc27-6fec05684c10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4521b5baee45ffc9c7661c781c0b3de458dff325ad525dde283d46af71c0cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vkd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zlk96\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:34Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:34 crc kubenswrapper[4821]: I1125 10:32:34.453933 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:34Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:34 crc kubenswrapper[4821]: I1125 10:32:34.492565 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17b8714be3bd121e06641873e003f08560b1f71e56d6ea936966df7d4896eb03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85731e6b6dbe1777bc7161a010f39f79523ec4b08d71d5d8a975573aa7fe4957\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:34Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:34 crc kubenswrapper[4821]: I1125 10:32:34.532602 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25595ccd-762d-414a-9916-4fe38d478b07\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d4b2d8db4d13392d9789e161fd4b8f6eef983b79128ee584d72e4c06565bc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf780f791771ea43a8c8176d02fdfb5204406e023207299369ae46ceb342baaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0251519fd4262f2ea5c64959db49433afdafe36b7ed2404d660eeb7c365d2746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad5c205e0eeecfd43cbdf0eb1a334e3e8faa5e9a9e8dd202c397d5cbb36e081e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:34Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:34 crc kubenswrapper[4821]: I1125 10:32:34.572175 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:34Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:34 crc kubenswrapper[4821]: I1125 10:32:34.612308 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ldzbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b84e32ae-4897-4644-a656-6bd82012cc81\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ef433744761f26e8e1f6c0db123821a11923ffd774830744285cd4178d89533\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5pmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ldzbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:34Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:34 crc kubenswrapper[4821]: I1125 10:32:34.657187 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"954085fa-d03a-4e88-9dfd-566257e3558d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzhnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:34Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:34 crc kubenswrapper[4821]: I1125 10:32:34.697652 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d9e58df-1755-4422-8d04-211f98c7c011\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea762f627aa69fb88a9bd82545fa677e9ccfa97f99a5e92c0aa490e619c0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a06b811e5ec52ce984a2403ba2c98332ba60466d08b3580ad9192fac6caef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0385f6ed9597974267f7c45547f6b6725b4ee83a2cc0508ff2d1c1ce8753ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b915b565f7fbb3e4e856e7f1035a519e82bfb3023660c121f6e3813de353bb92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18465f3a6916c36457a305ba5d568c9077f34f1d000532ed518f237d100008b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:34Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:34 crc kubenswrapper[4821]: I1125 10:32:34.733441 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:34Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:34 crc kubenswrapper[4821]: I1125 10:32:34.773315 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f948b87-ac86-4de6-ad64-c2ef947f84d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4482f73a45bc82bd55770c3fd788becacec94893fed1c305fed1c81c753d43aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ba68bd093be94d0d40f3851638db5d26bdfc52d175b1ac78afb91dda7affcfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2krbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:34Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:34 crc kubenswrapper[4821]: I1125 10:32:34.815044 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qppnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"357491b8-cd9c-486e-9898-5ba33678513d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd9c047c5ce1d896f803ff6cafa9bbf3a18a391b44539e1f1312e08c89e51cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd9c047c5ce1d896f803ff6cafa9bbf3a18a391b44539e1f1312e08c89e51cb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://565a6db63fb69abbba81ffdf6085797900858d263cd6affcb0b5f195474c2516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://565a6db63fb69abbba81ffdf6085797900858d263cd6affcb0b5f195474c2516\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qppnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:34Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:34 crc kubenswrapper[4821]: I1125 10:32:34.851265 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tgdlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2336955b-a703-4db4-8d0f-d08088ed59b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb0362942bfac635ee6292501b1b851772c01d75dd658eef4b17669a716de72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vjfp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tgdlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:34Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:34 crc kubenswrapper[4821]: I1125 10:32:34.894083 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62fcb632-f12a-46e2-bb7d-a5311f291ee7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28d43531927e75c3745f100940f7f6bf79fd21026c1309194df8cd7a39361b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69633332d03d2551ff4abb253d358de398eab18486b92db630ce06c0447fb9cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ccb78d9a51ac83ee208c2451685a2512de9deaf742a663bb8c892aa59b76162\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27826e155a9352b1e1866e83fdab683dbcd749cc04b6a4da2db25ecabe3d9a52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://991252e5f56f128827a5054b8e228284bc0266166582cbaedaadc53ca2f5b401\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:34Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:34 crc kubenswrapper[4821]: I1125 10:32:34.933362 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06f949936bac99cfbf89d720c9549c074351d9077bd963fa024d87004c6be23e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:34Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:34 crc kubenswrapper[4821]: I1125 10:32:34.972253 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384d8ba58264ab89796d3b53afb0c4c1be33cb4fba719a8beef4afc2be7d66a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:34Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:35 crc kubenswrapper[4821]: I1125 10:32:35.114082 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:32:35 crc kubenswrapper[4821]: I1125 10:32:35.114146 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:32:35 crc kubenswrapper[4821]: I1125 10:32:35.114082 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:32:35 crc kubenswrapper[4821]: E1125 10:32:35.114255 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:32:35 crc kubenswrapper[4821]: E1125 10:32:35.114372 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:32:35 crc kubenswrapper[4821]: E1125 10:32:35.114456 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:32:35 crc kubenswrapper[4821]: I1125 10:32:35.251253 4821 generic.go:334] "Generic (PLEG): container finished" podID="357491b8-cd9c-486e-9898-5ba33678513d" containerID="544885302dffff8336a9aa3c4afa80d69088ffaa2402648f4a329e0a798a8c99" exitCode=0 Nov 25 10:32:35 crc kubenswrapper[4821]: I1125 10:32:35.251296 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-qppnw" event={"ID":"357491b8-cd9c-486e-9898-5ba33678513d","Type":"ContainerDied","Data":"544885302dffff8336a9aa3c4afa80d69088ffaa2402648f4a329e0a798a8c99"} Nov 25 10:32:35 crc kubenswrapper[4821]: I1125 10:32:35.255577 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" event={"ID":"954085fa-d03a-4e88-9dfd-566257e3558d","Type":"ContainerStarted","Data":"7dbf499f4ae6161752abaa7a466a027c6f874016be90a534e69f07e3dd3f3d9f"} Nov 25 10:32:35 crc kubenswrapper[4821]: I1125 10:32:35.270823 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d9e58df-1755-4422-8d04-211f98c7c011\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea762f627aa69fb88a9bd82545fa677e9ccfa97f99a5e92c0aa490e619c0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a06b811e5ec52ce984a2403ba2c98332ba60466d08b3580ad9192fac6caef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0385f6ed9597974267f7c45547f6b6725b4ee83a2cc0508ff2d1c1ce8753ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b915b565f7fbb3e4e856e7f1035a519e82bfb3023660c121f6e3813de353bb92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18465f3a6916c36457a305ba5d568c9077f34f1d000532ed518f237d100008b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:35Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:35 crc kubenswrapper[4821]: I1125 10:32:35.284643 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25595ccd-762d-414a-9916-4fe38d478b07\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d4b2d8db4d13392d9789e161fd4b8f6eef983b79128ee584d72e4c06565bc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf780f791771ea43a8c8176d02fdfb5204406e023207299369ae46ceb342baaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0251519fd4262f2ea5c64959db49433afdafe36b7ed2404d660eeb7c365d2746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad5c205e0eeecfd43cbdf0eb1a334e3e8faa5e9a9e8dd202c397d5cbb36e081e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:35Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:35 crc kubenswrapper[4821]: I1125 10:32:35.295053 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:35Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:35 crc kubenswrapper[4821]: I1125 10:32:35.306940 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ldzbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b84e32ae-4897-4644-a656-6bd82012cc81\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ef433744761f26e8e1f6c0db123821a11923ffd774830744285cd4178d89533\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5pmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ldzbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:35Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:35 crc kubenswrapper[4821]: I1125 10:32:35.328842 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"954085fa-d03a-4e88-9dfd-566257e3558d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzhnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:35Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:35 crc kubenswrapper[4821]: I1125 10:32:35.339929 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62fcb632-f12a-46e2-bb7d-a5311f291ee7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28d43531927e75c3745f100940f7f6bf79fd21026c1309194df8cd7a39361b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69633332d03d2551ff4abb253d358de398eab18486b92db630ce06c0447fb9cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ccb78d9a51ac83ee208c2451685a2512de9deaf742a663bb8c892aa59b76162\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27826e155a9352b1e1866e83fdab683dbcd749cc04b6a4da2db25ecabe3d9a52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://991252e5f56f128827a5054b8e228284bc0266166582cbaedaadc53ca2f5b401\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:35Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:35 crc kubenswrapper[4821]: I1125 10:32:35.351676 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:35Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:35 crc kubenswrapper[4821]: I1125 10:32:35.361119 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f948b87-ac86-4de6-ad64-c2ef947f84d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4482f73a45bc82bd55770c3fd788becacec94893fed1c305fed1c81c753d43aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ba68bd093be94d0d40f3851638db5d26bdfc52d175b1ac78afb91dda7affcfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2krbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:35Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:35 crc kubenswrapper[4821]: I1125 10:32:35.377706 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qppnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"357491b8-cd9c-486e-9898-5ba33678513d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd9c047c5ce1d896f803ff6cafa9bbf3a18a391b44539e1f1312e08c89e51cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd9c047c5ce1d896f803ff6cafa9bbf3a18a391b44539e1f1312e08c89e51cb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://565a6db63fb69abbba81ffdf6085797900858d263cd6affcb0b5f195474c2516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://565a6db63fb69abbba81ffdf6085797900858d263cd6affcb0b5f195474c2516\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://544885302dffff8336a9aa3c4afa80d69088ffaa2402648f4a329e0a798a8c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://544885302dffff8336a9aa3c4afa80d69088ffaa2402648f4a329e0a798a8c99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qppnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:35Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:35 crc kubenswrapper[4821]: I1125 10:32:35.389076 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tgdlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2336955b-a703-4db4-8d0f-d08088ed59b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb0362942bfac635ee6292501b1b851772c01d75dd658eef4b17669a716de72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vjfp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tgdlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:35Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:35 crc kubenswrapper[4821]: I1125 10:32:35.415267 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06f949936bac99cfbf89d720c9549c074351d9077bd963fa024d87004c6be23e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:35Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:35 crc kubenswrapper[4821]: I1125 10:32:35.455183 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384d8ba58264ab89796d3b53afb0c4c1be33cb4fba719a8beef4afc2be7d66a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:35Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:35 crc kubenswrapper[4821]: I1125 10:32:35.509411 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:35Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:35 crc kubenswrapper[4821]: I1125 10:32:35.514111 4821 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 10:32:35 crc kubenswrapper[4821]: I1125 10:32:35.517735 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:35 crc kubenswrapper[4821]: I1125 10:32:35.517774 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:35 crc kubenswrapper[4821]: I1125 10:32:35.517787 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:35 crc kubenswrapper[4821]: I1125 10:32:35.517959 4821 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 10:32:35 crc kubenswrapper[4821]: I1125 10:32:35.536281 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17b8714be3bd121e06641873e003f08560b1f71e56d6ea936966df7d4896eb03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85731e6b6dbe1777bc7161a010f39f79523ec4b08d71d5d8a975573aa7fe4957\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:35Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:35 crc kubenswrapper[4821]: I1125 10:32:35.587036 4821 kubelet_node_status.go:115] "Node was previously registered" node="crc" Nov 25 10:32:35 crc kubenswrapper[4821]: I1125 10:32:35.587351 4821 kubelet_node_status.go:79] "Successfully registered node" node="crc" Nov 25 10:32:35 crc kubenswrapper[4821]: I1125 10:32:35.588538 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:35 crc kubenswrapper[4821]: I1125 10:32:35.588568 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:35 crc kubenswrapper[4821]: I1125 10:32:35.588581 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:35 crc kubenswrapper[4821]: I1125 10:32:35.588597 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:35 crc kubenswrapper[4821]: I1125 10:32:35.588610 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:35Z","lastTransitionTime":"2025-11-25T10:32:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:35 crc kubenswrapper[4821]: E1125 10:32:35.605567 4821 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:32:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:32:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:32:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:32:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a5b2b070-39b9-412c-90f1-c7ed2e9e0950\\\",\\\"systemUUID\\\":\\\"ea86e80f-41b7-4190-ad7e-d8c549d535de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:35Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:35 crc kubenswrapper[4821]: I1125 10:32:35.609324 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:35 crc kubenswrapper[4821]: I1125 10:32:35.609380 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:35 crc kubenswrapper[4821]: I1125 10:32:35.609394 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:35 crc kubenswrapper[4821]: I1125 10:32:35.609412 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:35 crc kubenswrapper[4821]: I1125 10:32:35.609426 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:35Z","lastTransitionTime":"2025-11-25T10:32:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:35 crc kubenswrapper[4821]: I1125 10:32:35.615123 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zlk96" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d55aa91-9cb9-4c95-bc27-6fec05684c10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4521b5baee45ffc9c7661c781c0b3de458dff325ad525dde283d46af71c0cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vkd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zlk96\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:35Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:35 crc kubenswrapper[4821]: E1125 10:32:35.622955 4821 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:32:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:32:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:32:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:32:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a5b2b070-39b9-412c-90f1-c7ed2e9e0950\\\",\\\"systemUUID\\\":\\\"ea86e80f-41b7-4190-ad7e-d8c549d535de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:35Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:35 crc kubenswrapper[4821]: I1125 10:32:35.626254 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:35 crc kubenswrapper[4821]: I1125 10:32:35.626276 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:35 crc kubenswrapper[4821]: I1125 10:32:35.626284 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:35 crc kubenswrapper[4821]: I1125 10:32:35.626298 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:35 crc kubenswrapper[4821]: I1125 10:32:35.626307 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:35Z","lastTransitionTime":"2025-11-25T10:32:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:35 crc kubenswrapper[4821]: E1125 10:32:35.638303 4821 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:32:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:32:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:32:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:32:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a5b2b070-39b9-412c-90f1-c7ed2e9e0950\\\",\\\"systemUUID\\\":\\\"ea86e80f-41b7-4190-ad7e-d8c549d535de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:35Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:35 crc kubenswrapper[4821]: I1125 10:32:35.641626 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:35 crc kubenswrapper[4821]: I1125 10:32:35.641674 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:35 crc kubenswrapper[4821]: I1125 10:32:35.641685 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:35 crc kubenswrapper[4821]: I1125 10:32:35.641702 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:35 crc kubenswrapper[4821]: I1125 10:32:35.641718 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:35Z","lastTransitionTime":"2025-11-25T10:32:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:35 crc kubenswrapper[4821]: E1125 10:32:35.652517 4821 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:32:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:32:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:32:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:32:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a5b2b070-39b9-412c-90f1-c7ed2e9e0950\\\",\\\"systemUUID\\\":\\\"ea86e80f-41b7-4190-ad7e-d8c549d535de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:35Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:35 crc kubenswrapper[4821]: I1125 10:32:35.656227 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:35 crc kubenswrapper[4821]: I1125 10:32:35.656264 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:35 crc kubenswrapper[4821]: I1125 10:32:35.656276 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:35 crc kubenswrapper[4821]: I1125 10:32:35.656292 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:35 crc kubenswrapper[4821]: I1125 10:32:35.656302 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:35Z","lastTransitionTime":"2025-11-25T10:32:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:35 crc kubenswrapper[4821]: E1125 10:32:35.667515 4821 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:32:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:32:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:32:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:32:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a5b2b070-39b9-412c-90f1-c7ed2e9e0950\\\",\\\"systemUUID\\\":\\\"ea86e80f-41b7-4190-ad7e-d8c549d535de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:35Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:35 crc kubenswrapper[4821]: E1125 10:32:35.667626 4821 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 10:32:35 crc kubenswrapper[4821]: I1125 10:32:35.669409 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:35 crc kubenswrapper[4821]: I1125 10:32:35.669428 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:35 crc kubenswrapper[4821]: I1125 10:32:35.669436 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:35 crc kubenswrapper[4821]: I1125 10:32:35.669451 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:35 crc kubenswrapper[4821]: I1125 10:32:35.669464 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:35Z","lastTransitionTime":"2025-11-25T10:32:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:35 crc kubenswrapper[4821]: I1125 10:32:35.771559 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:35 crc kubenswrapper[4821]: I1125 10:32:35.771713 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:35 crc kubenswrapper[4821]: I1125 10:32:35.771732 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:35 crc kubenswrapper[4821]: I1125 10:32:35.771749 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:35 crc kubenswrapper[4821]: I1125 10:32:35.771759 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:35Z","lastTransitionTime":"2025-11-25T10:32:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:35 crc kubenswrapper[4821]: I1125 10:32:35.874107 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:35 crc kubenswrapper[4821]: I1125 10:32:35.874133 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:35 crc kubenswrapper[4821]: I1125 10:32:35.874142 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:35 crc kubenswrapper[4821]: I1125 10:32:35.874171 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:35 crc kubenswrapper[4821]: I1125 10:32:35.874181 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:35Z","lastTransitionTime":"2025-11-25T10:32:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:35 crc kubenswrapper[4821]: I1125 10:32:35.977149 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:35 crc kubenswrapper[4821]: I1125 10:32:35.977224 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:35 crc kubenswrapper[4821]: I1125 10:32:35.977391 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:35 crc kubenswrapper[4821]: I1125 10:32:35.977416 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:35 crc kubenswrapper[4821]: I1125 10:32:35.977430 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:35Z","lastTransitionTime":"2025-11-25T10:32:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:36 crc kubenswrapper[4821]: I1125 10:32:36.080270 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:36 crc kubenswrapper[4821]: I1125 10:32:36.080327 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:36 crc kubenswrapper[4821]: I1125 10:32:36.080340 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:36 crc kubenswrapper[4821]: I1125 10:32:36.080358 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:36 crc kubenswrapper[4821]: I1125 10:32:36.080367 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:36Z","lastTransitionTime":"2025-11-25T10:32:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:36 crc kubenswrapper[4821]: I1125 10:32:36.183511 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:36 crc kubenswrapper[4821]: I1125 10:32:36.183551 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:36 crc kubenswrapper[4821]: I1125 10:32:36.183565 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:36 crc kubenswrapper[4821]: I1125 10:32:36.183585 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:36 crc kubenswrapper[4821]: I1125 10:32:36.183597 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:36Z","lastTransitionTime":"2025-11-25T10:32:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:36 crc kubenswrapper[4821]: I1125 10:32:36.262594 4821 generic.go:334] "Generic (PLEG): container finished" podID="357491b8-cd9c-486e-9898-5ba33678513d" containerID="b8bc193fb3f85b980c31af2e826cee90adb633ad2ee82ff1d46e182ab52e8fab" exitCode=0 Nov 25 10:32:36 crc kubenswrapper[4821]: I1125 10:32:36.262641 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-qppnw" event={"ID":"357491b8-cd9c-486e-9898-5ba33678513d","Type":"ContainerDied","Data":"b8bc193fb3f85b980c31af2e826cee90adb633ad2ee82ff1d46e182ab52e8fab"} Nov 25 10:32:36 crc kubenswrapper[4821]: I1125 10:32:36.277737 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06f949936bac99cfbf89d720c9549c074351d9077bd963fa024d87004c6be23e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:36Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:36 crc kubenswrapper[4821]: I1125 10:32:36.285884 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:36 crc kubenswrapper[4821]: I1125 10:32:36.285925 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:36 crc kubenswrapper[4821]: I1125 10:32:36.285936 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:36 crc kubenswrapper[4821]: I1125 10:32:36.285953 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:36 crc kubenswrapper[4821]: I1125 10:32:36.285964 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:36Z","lastTransitionTime":"2025-11-25T10:32:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:36 crc kubenswrapper[4821]: I1125 10:32:36.291080 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384d8ba58264ab89796d3b53afb0c4c1be33cb4fba719a8beef4afc2be7d66a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:36Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:36 crc kubenswrapper[4821]: I1125 10:32:36.301747 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zlk96" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d55aa91-9cb9-4c95-bc27-6fec05684c10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4521b5baee45ffc9c7661c781c0b3de458dff325ad525dde283d46af71c0cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vkd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zlk96\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:36Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:36 crc kubenswrapper[4821]: I1125 10:32:36.314399 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:36Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:36 crc kubenswrapper[4821]: I1125 10:32:36.327198 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17b8714be3bd121e06641873e003f08560b1f71e56d6ea936966df7d4896eb03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85731e6b6dbe1777bc7161a010f39f79523ec4b08d71d5d8a975573aa7fe4957\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:36Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:36 crc kubenswrapper[4821]: I1125 10:32:36.338981 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25595ccd-762d-414a-9916-4fe38d478b07\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d4b2d8db4d13392d9789e161fd4b8f6eef983b79128ee584d72e4c06565bc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf780f791771ea43a8c8176d02fdfb5204406e023207299369ae46ceb342baaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0251519fd4262f2ea5c64959db49433afdafe36b7ed2404d660eeb7c365d2746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad5c205e0eeecfd43cbdf0eb1a334e3e8faa5e9a9e8dd202c397d5cbb36e081e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:36Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:36 crc kubenswrapper[4821]: I1125 10:32:36.355371 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:36Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:36 crc kubenswrapper[4821]: I1125 10:32:36.369212 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ldzbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b84e32ae-4897-4644-a656-6bd82012cc81\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ef433744761f26e8e1f6c0db123821a11923ffd774830744285cd4178d89533\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5pmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ldzbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:36Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:36 crc kubenswrapper[4821]: I1125 10:32:36.388712 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:36 crc kubenswrapper[4821]: I1125 10:32:36.388748 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:36 crc kubenswrapper[4821]: I1125 10:32:36.388758 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:36 crc kubenswrapper[4821]: I1125 10:32:36.388773 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:36 crc kubenswrapper[4821]: I1125 10:32:36.388783 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:36Z","lastTransitionTime":"2025-11-25T10:32:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:36 crc kubenswrapper[4821]: I1125 10:32:36.389487 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"954085fa-d03a-4e88-9dfd-566257e3558d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzhnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:36Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:36 crc kubenswrapper[4821]: I1125 10:32:36.415053 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d9e58df-1755-4422-8d04-211f98c7c011\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea762f627aa69fb88a9bd82545fa677e9ccfa97f99a5e92c0aa490e619c0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a06b811e5ec52ce984a2403ba2c98332ba60466d08b3580ad9192fac6caef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0385f6ed9597974267f7c45547f6b6725b4ee83a2cc0508ff2d1c1ce8753ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b915b565f7fbb3e4e856e7f1035a519e82bfb3023660c121f6e3813de353bb92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18465f3a6916c36457a305ba5d568c9077f34f1d000532ed518f237d100008b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:36Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:36 crc kubenswrapper[4821]: I1125 10:32:36.428548 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:36Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:36 crc kubenswrapper[4821]: I1125 10:32:36.443585 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f948b87-ac86-4de6-ad64-c2ef947f84d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4482f73a45bc82bd55770c3fd788becacec94893fed1c305fed1c81c753d43aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ba68bd093be94d0d40f3851638db5d26bdfc52d175b1ac78afb91dda7affcfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2krbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:36Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:36 crc kubenswrapper[4821]: I1125 10:32:36.458201 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qppnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"357491b8-cd9c-486e-9898-5ba33678513d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd9c047c5ce1d896f803ff6cafa9bbf3a18a391b44539e1f1312e08c89e51cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd9c047c5ce1d896f803ff6cafa9bbf3a18a391b44539e1f1312e08c89e51cb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://565a6db63fb69abbba81ffdf6085797900858d263cd6affcb0b5f195474c2516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://565a6db63fb69abbba81ffdf6085797900858d263cd6affcb0b5f195474c2516\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://544885302dffff8336a9aa3c4afa80d69088ffaa2402648f4a329e0a798a8c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://544885302dffff8336a9aa3c4afa80d69088ffaa2402648f4a329e0a798a8c99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8bc193fb3f85b980c31af2e826cee90adb633ad2ee82ff1d46e182ab52e8fab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8bc193fb3f85b980c31af2e826cee90adb633ad2ee82ff1d46e182ab52e8fab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qppnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:36Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:36 crc kubenswrapper[4821]: I1125 10:32:36.469585 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tgdlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2336955b-a703-4db4-8d0f-d08088ed59b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb0362942bfac635ee6292501b1b851772c01d75dd658eef4b17669a716de72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vjfp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tgdlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:36Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:36 crc kubenswrapper[4821]: I1125 10:32:36.485247 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62fcb632-f12a-46e2-bb7d-a5311f291ee7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28d43531927e75c3745f100940f7f6bf79fd21026c1309194df8cd7a39361b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69633332d03d2551ff4abb253d358de398eab18486b92db630ce06c0447fb9cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ccb78d9a51ac83ee208c2451685a2512de9deaf742a663bb8c892aa59b76162\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27826e155a9352b1e1866e83fdab683dbcd749cc04b6a4da2db25ecabe3d9a52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://991252e5f56f128827a5054b8e228284bc0266166582cbaedaadc53ca2f5b401\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:36Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:36 crc kubenswrapper[4821]: I1125 10:32:36.492119 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:36 crc kubenswrapper[4821]: I1125 10:32:36.492179 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:36 crc kubenswrapper[4821]: I1125 10:32:36.492190 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:36 crc kubenswrapper[4821]: I1125 10:32:36.492222 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:36 crc kubenswrapper[4821]: I1125 10:32:36.492233 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:36Z","lastTransitionTime":"2025-11-25T10:32:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:36 crc kubenswrapper[4821]: I1125 10:32:36.594872 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:36 crc kubenswrapper[4821]: I1125 10:32:36.594905 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:36 crc kubenswrapper[4821]: I1125 10:32:36.594915 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:36 crc kubenswrapper[4821]: I1125 10:32:36.594931 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:36 crc kubenswrapper[4821]: I1125 10:32:36.594942 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:36Z","lastTransitionTime":"2025-11-25T10:32:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:36 crc kubenswrapper[4821]: I1125 10:32:36.696861 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:36 crc kubenswrapper[4821]: I1125 10:32:36.696903 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:36 crc kubenswrapper[4821]: I1125 10:32:36.696911 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:36 crc kubenswrapper[4821]: I1125 10:32:36.696926 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:36 crc kubenswrapper[4821]: I1125 10:32:36.696937 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:36Z","lastTransitionTime":"2025-11-25T10:32:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:36 crc kubenswrapper[4821]: I1125 10:32:36.799132 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:36 crc kubenswrapper[4821]: I1125 10:32:36.799198 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:36 crc kubenswrapper[4821]: I1125 10:32:36.799207 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:36 crc kubenswrapper[4821]: I1125 10:32:36.799221 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:36 crc kubenswrapper[4821]: I1125 10:32:36.799230 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:36Z","lastTransitionTime":"2025-11-25T10:32:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:36 crc kubenswrapper[4821]: I1125 10:32:36.841765 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:32:36 crc kubenswrapper[4821]: I1125 10:32:36.841877 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:32:36 crc kubenswrapper[4821]: E1125 10:32:36.841938 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:32:44.841920465 +0000 UTC m=+35.378240322 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:32:36 crc kubenswrapper[4821]: I1125 10:32:36.842001 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:32:36 crc kubenswrapper[4821]: E1125 10:32:36.842029 4821 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 10:32:36 crc kubenswrapper[4821]: E1125 10:32:36.842105 4821 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 10:32:36 crc kubenswrapper[4821]: E1125 10:32:36.842117 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 10:32:44.842094578 +0000 UTC m=+35.378414465 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 10:32:36 crc kubenswrapper[4821]: E1125 10:32:36.842151 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 10:32:44.842142539 +0000 UTC m=+35.378462466 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 10:32:36 crc kubenswrapper[4821]: I1125 10:32:36.901781 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:36 crc kubenswrapper[4821]: I1125 10:32:36.901831 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:36 crc kubenswrapper[4821]: I1125 10:32:36.901843 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:36 crc kubenswrapper[4821]: I1125 10:32:36.901864 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:36 crc kubenswrapper[4821]: I1125 10:32:36.901875 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:36Z","lastTransitionTime":"2025-11-25T10:32:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:36 crc kubenswrapper[4821]: I1125 10:32:36.943210 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:32:36 crc kubenswrapper[4821]: I1125 10:32:36.943515 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:32:36 crc kubenswrapper[4821]: E1125 10:32:36.943397 4821 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 10:32:36 crc kubenswrapper[4821]: E1125 10:32:36.943555 4821 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 10:32:36 crc kubenswrapper[4821]: E1125 10:32:36.943570 4821 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 10:32:36 crc kubenswrapper[4821]: E1125 10:32:36.943583 4821 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 10:32:36 crc kubenswrapper[4821]: E1125 10:32:36.943593 4821 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 10:32:36 crc kubenswrapper[4821]: E1125 10:32:36.943602 4821 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 10:32:36 crc kubenswrapper[4821]: E1125 10:32:36.943620 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 10:32:44.943603587 +0000 UTC m=+35.479923424 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 10:32:36 crc kubenswrapper[4821]: E1125 10:32:36.943638 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 10:32:44.943630948 +0000 UTC m=+35.479950795 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.003932 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.003975 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.003988 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.004014 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.004029 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:37Z","lastTransitionTime":"2025-11-25T10:32:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.106021 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.106056 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.106065 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.106079 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.106090 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:37Z","lastTransitionTime":"2025-11-25T10:32:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.113873 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:32:37 crc kubenswrapper[4821]: E1125 10:32:37.114004 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.114247 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:32:37 crc kubenswrapper[4821]: E1125 10:32:37.114321 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.114244 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:32:37 crc kubenswrapper[4821]: E1125 10:32:37.114386 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.207971 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.208005 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.208014 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.208028 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.208037 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:37Z","lastTransitionTime":"2025-11-25T10:32:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.268465 4821 generic.go:334] "Generic (PLEG): container finished" podID="357491b8-cd9c-486e-9898-5ba33678513d" containerID="d7713e2a35bac8df1e8987fbcd452459fa3541caad0c52ed57fda187bd7c83d9" exitCode=0 Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.268533 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-qppnw" event={"ID":"357491b8-cd9c-486e-9898-5ba33678513d","Type":"ContainerDied","Data":"d7713e2a35bac8df1e8987fbcd452459fa3541caad0c52ed57fda187bd7c83d9"} Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.273801 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" event={"ID":"954085fa-d03a-4e88-9dfd-566257e3558d","Type":"ContainerStarted","Data":"a4ea2031b69dc207c47ef71a22c04431f4d2dab95e183b3768dd11594dd6bb32"} Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.274026 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.274043 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.281985 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62fcb632-f12a-46e2-bb7d-a5311f291ee7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28d43531927e75c3745f100940f7f6bf79fd21026c1309194df8cd7a39361b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69633332d03d2551ff4abb253d358de398eab18486b92db630ce06c0447fb9cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ccb78d9a51ac83ee208c2451685a2512de9deaf742a663bb8c892aa59b76162\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27826e155a9352b1e1866e83fdab683dbcd749cc04b6a4da2db25ecabe3d9a52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://991252e5f56f128827a5054b8e228284bc0266166582cbaedaadc53ca2f5b401\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:37Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.294729 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:37Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.297223 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.305154 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f948b87-ac86-4de6-ad64-c2ef947f84d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4482f73a45bc82bd55770c3fd788becacec94893fed1c305fed1c81c753d43aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ba68bd093be94d0d40f3851638db5d26bdfc52d175b1ac78afb91dda7affcfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2krbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:37Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.310188 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.310219 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.310227 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.310240 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.310249 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:37Z","lastTransitionTime":"2025-11-25T10:32:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.319622 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qppnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"357491b8-cd9c-486e-9898-5ba33678513d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd9c047c5ce1d896f803ff6cafa9bbf3a18a391b44539e1f1312e08c89e51cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd9c047c5ce1d896f803ff6cafa9bbf3a18a391b44539e1f1312e08c89e51cb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://565a6db63fb69abbba81ffdf6085797900858d263cd6affcb0b5f195474c2516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://565a6db63fb69abbba81ffdf6085797900858d263cd6affcb0b5f195474c2516\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://544885302dffff8336a9aa3c4afa80d69088ffaa2402648f4a329e0a798a8c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://544885302dffff8336a9aa3c4afa80d69088ffaa2402648f4a329e0a798a8c99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8bc193fb3f85b980c31af2e826cee90adb633ad2ee82ff1d46e182ab52e8fab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8bc193fb3f85b980c31af2e826cee90adb633ad2ee82ff1d46e182ab52e8fab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7713e2a35bac8df1e8987fbcd452459fa3541caad0c52ed57fda187bd7c83d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7713e2a35bac8df1e8987fbcd452459fa3541caad0c52ed57fda187bd7c83d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qppnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:37Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.334400 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tgdlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2336955b-a703-4db4-8d0f-d08088ed59b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb0362942bfac635ee6292501b1b851772c01d75dd658eef4b17669a716de72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vjfp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tgdlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:37Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.348620 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06f949936bac99cfbf89d720c9549c074351d9077bd963fa024d87004c6be23e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:37Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.364068 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384d8ba58264ab89796d3b53afb0c4c1be33cb4fba719a8beef4afc2be7d66a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:37Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.376509 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:37Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.388213 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17b8714be3bd121e06641873e003f08560b1f71e56d6ea936966df7d4896eb03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85731e6b6dbe1777bc7161a010f39f79523ec4b08d71d5d8a975573aa7fe4957\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:37Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.397933 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zlk96" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d55aa91-9cb9-4c95-bc27-6fec05684c10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4521b5baee45ffc9c7661c781c0b3de458dff325ad525dde283d46af71c0cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vkd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zlk96\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:37Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.411834 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.411863 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.411871 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.411885 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.411894 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:37Z","lastTransitionTime":"2025-11-25T10:32:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.415126 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d9e58df-1755-4422-8d04-211f98c7c011\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea762f627aa69fb88a9bd82545fa677e9ccfa97f99a5e92c0aa490e619c0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a06b811e5ec52ce984a2403ba2c98332ba60466d08b3580ad9192fac6caef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0385f6ed9597974267f7c45547f6b6725b4ee83a2cc0508ff2d1c1ce8753ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b915b565f7fbb3e4e856e7f1035a519e82bfb3023660c121f6e3813de353bb92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18465f3a6916c36457a305ba5d568c9077f34f1d000532ed518f237d100008b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:37Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.425900 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25595ccd-762d-414a-9916-4fe38d478b07\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d4b2d8db4d13392d9789e161fd4b8f6eef983b79128ee584d72e4c06565bc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf780f791771ea43a8c8176d02fdfb5204406e023207299369ae46ceb342baaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0251519fd4262f2ea5c64959db49433afdafe36b7ed2404d660eeb7c365d2746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad5c205e0eeecfd43cbdf0eb1a334e3e8faa5e9a9e8dd202c397d5cbb36e081e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:37Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.436774 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:37Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.449947 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ldzbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b84e32ae-4897-4644-a656-6bd82012cc81\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ef433744761f26e8e1f6c0db123821a11923ffd774830744285cd4178d89533\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5pmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ldzbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:37Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.467597 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"954085fa-d03a-4e88-9dfd-566257e3558d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzhnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:37Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.479096 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06f949936bac99cfbf89d720c9549c074351d9077bd963fa024d87004c6be23e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:37Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.490688 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384d8ba58264ab89796d3b53afb0c4c1be33cb4fba719a8beef4afc2be7d66a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:37Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.500933 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zlk96" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d55aa91-9cb9-4c95-bc27-6fec05684c10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4521b5baee45ffc9c7661c781c0b3de458dff325ad525dde283d46af71c0cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vkd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zlk96\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:37Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.510613 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:37Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.513941 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.513969 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.513981 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.513997 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.514008 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:37Z","lastTransitionTime":"2025-11-25T10:32:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.523623 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17b8714be3bd121e06641873e003f08560b1f71e56d6ea936966df7d4896eb03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85731e6b6dbe1777bc7161a010f39f79523ec4b08d71d5d8a975573aa7fe4957\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:37Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.535768 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25595ccd-762d-414a-9916-4fe38d478b07\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d4b2d8db4d13392d9789e161fd4b8f6eef983b79128ee584d72e4c06565bc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf780f791771ea43a8c8176d02fdfb5204406e023207299369ae46ceb342baaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0251519fd4262f2ea5c64959db49433afdafe36b7ed2404d660eeb7c365d2746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad5c205e0eeecfd43cbdf0eb1a334e3e8faa5e9a9e8dd202c397d5cbb36e081e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:37Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.546714 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:37Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.557465 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ldzbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b84e32ae-4897-4644-a656-6bd82012cc81\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ef433744761f26e8e1f6c0db123821a11923ffd774830744285cd4178d89533\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5pmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ldzbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:37Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.572632 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"954085fa-d03a-4e88-9dfd-566257e3558d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f00746930c4588e548f0e2b60d52bf448aff7401b676a7855cb0f4cce00d329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a14a078ba7d2aed95992b36502008f315f0d3a634b64ed49da70398b2562d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d7254196b3702e48083ae98e829b73c539adf69ed5b2bbdc62189d09ae1762\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d727430108977574365f0e9c04b44e03307bd7ff89cdbb1e48a6e916f504d6b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94c7949f077b1ca2b310bedb090aa72fc4bd7e8f2c8bf438096edfca8568f36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed2e7cad7bb24b5b56a1295098a6a7c230c8f9b9c29a88ef653ae8b6ac31667d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4ea2031b69dc207c47ef71a22c04431f4d2dab95e183b3768dd11594dd6bb32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dbf499f4ae6161752abaa7a466a027c6f874016be90a534e69f07e3dd3f3d9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzhnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:37Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.590100 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d9e58df-1755-4422-8d04-211f98c7c011\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea762f627aa69fb88a9bd82545fa677e9ccfa97f99a5e92c0aa490e619c0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a06b811e5ec52ce984a2403ba2c98332ba60466d08b3580ad9192fac6caef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0385f6ed9597974267f7c45547f6b6725b4ee83a2cc0508ff2d1c1ce8753ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b915b565f7fbb3e4e856e7f1035a519e82bfb3023660c121f6e3813de353bb92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18465f3a6916c36457a305ba5d568c9077f34f1d000532ed518f237d100008b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:37Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.601949 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:37Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.615810 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.615843 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.615853 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.615869 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.615880 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:37Z","lastTransitionTime":"2025-11-25T10:32:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.616115 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f948b87-ac86-4de6-ad64-c2ef947f84d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4482f73a45bc82bd55770c3fd788becacec94893fed1c305fed1c81c753d43aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ba68bd093be94d0d40f3851638db5d26bdfc52d175b1ac78afb91dda7affcfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2krbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:37Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.630592 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qppnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"357491b8-cd9c-486e-9898-5ba33678513d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd9c047c5ce1d896f803ff6cafa9bbf3a18a391b44539e1f1312e08c89e51cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd9c047c5ce1d896f803ff6cafa9bbf3a18a391b44539e1f1312e08c89e51cb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://565a6db63fb69abbba81ffdf6085797900858d263cd6affcb0b5f195474c2516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://565a6db63fb69abbba81ffdf6085797900858d263cd6affcb0b5f195474c2516\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://544885302dffff8336a9aa3c4afa80d69088ffaa2402648f4a329e0a798a8c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://544885302dffff8336a9aa3c4afa80d69088ffaa2402648f4a329e0a798a8c99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8bc193fb3f85b980c31af2e826cee90adb633ad2ee82ff1d46e182ab52e8fab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8bc193fb3f85b980c31af2e826cee90adb633ad2ee82ff1d46e182ab52e8fab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7713e2a35bac8df1e8987fbcd452459fa3541caad0c52ed57fda187bd7c83d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7713e2a35bac8df1e8987fbcd452459fa3541caad0c52ed57fda187bd7c83d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qppnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:37Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.639985 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tgdlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2336955b-a703-4db4-8d0f-d08088ed59b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb0362942bfac635ee6292501b1b851772c01d75dd658eef4b17669a716de72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vjfp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tgdlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:37Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.658238 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62fcb632-f12a-46e2-bb7d-a5311f291ee7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28d43531927e75c3745f100940f7f6bf79fd21026c1309194df8cd7a39361b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69633332d03d2551ff4abb253d358de398eab18486b92db630ce06c0447fb9cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ccb78d9a51ac83ee208c2451685a2512de9deaf742a663bb8c892aa59b76162\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27826e155a9352b1e1866e83fdab683dbcd749cc04b6a4da2db25ecabe3d9a52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://991252e5f56f128827a5054b8e228284bc0266166582cbaedaadc53ca2f5b401\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:37Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.717971 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.718013 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.718024 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.718045 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.718057 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:37Z","lastTransitionTime":"2025-11-25T10:32:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.820454 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.820491 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.820501 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.820515 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.820524 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:37Z","lastTransitionTime":"2025-11-25T10:32:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.922471 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.922519 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.922530 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.922548 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:37 crc kubenswrapper[4821]: I1125 10:32:37.922561 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:37Z","lastTransitionTime":"2025-11-25T10:32:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.025084 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.025134 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.025146 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.025185 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.025209 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:38Z","lastTransitionTime":"2025-11-25T10:32:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.127314 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.127675 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.127690 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.127709 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.127723 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:38Z","lastTransitionTime":"2025-11-25T10:32:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.230360 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.230442 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.230455 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.230471 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.230484 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:38Z","lastTransitionTime":"2025-11-25T10:32:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.279509 4821 generic.go:334] "Generic (PLEG): container finished" podID="357491b8-cd9c-486e-9898-5ba33678513d" containerID="34adfca89026fc20007a73a8cbd4e7c7b3137a581c3476867af17fabe976f71f" exitCode=0 Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.279618 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-qppnw" event={"ID":"357491b8-cd9c-486e-9898-5ba33678513d","Type":"ContainerDied","Data":"34adfca89026fc20007a73a8cbd4e7c7b3137a581c3476867af17fabe976f71f"} Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.280038 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.295415 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62fcb632-f12a-46e2-bb7d-a5311f291ee7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28d43531927e75c3745f100940f7f6bf79fd21026c1309194df8cd7a39361b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69633332d03d2551ff4abb253d358de398eab18486b92db630ce06c0447fb9cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ccb78d9a51ac83ee208c2451685a2512de9deaf742a663bb8c892aa59b76162\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27826e155a9352b1e1866e83fdab683dbcd749cc04b6a4da2db25ecabe3d9a52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://991252e5f56f128827a5054b8e228284bc0266166582cbaedaadc53ca2f5b401\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:38Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.303349 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.309558 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:38Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.320855 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f948b87-ac86-4de6-ad64-c2ef947f84d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4482f73a45bc82bd55770c3fd788becacec94893fed1c305fed1c81c753d43aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ba68bd093be94d0d40f3851638db5d26bdfc52d175b1ac78afb91dda7affcfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2krbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:38Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.332116 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.332153 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.332176 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.332194 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.332206 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:38Z","lastTransitionTime":"2025-11-25T10:32:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.336955 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qppnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"357491b8-cd9c-486e-9898-5ba33678513d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd9c047c5ce1d896f803ff6cafa9bbf3a18a391b44539e1f1312e08c89e51cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd9c047c5ce1d896f803ff6cafa9bbf3a18a391b44539e1f1312e08c89e51cb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://565a6db63fb69abbba81ffdf6085797900858d263cd6affcb0b5f195474c2516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://565a6db63fb69abbba81ffdf6085797900858d263cd6affcb0b5f195474c2516\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://544885302dffff8336a9aa3c4afa80d69088ffaa2402648f4a329e0a798a8c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://544885302dffff8336a9aa3c4afa80d69088ffaa2402648f4a329e0a798a8c99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8bc193fb3f85b980c31af2e826cee90adb633ad2ee82ff1d46e182ab52e8fab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8bc193fb3f85b980c31af2e826cee90adb633ad2ee82ff1d46e182ab52e8fab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7713e2a35bac8df1e8987fbcd452459fa3541caad0c52ed57fda187bd7c83d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7713e2a35bac8df1e8987fbcd452459fa3541caad0c52ed57fda187bd7c83d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34adfca89026fc20007a73a8cbd4e7c7b3137a581c3476867af17fabe976f71f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34adfca89026fc20007a73a8cbd4e7c7b3137a581c3476867af17fabe976f71f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qppnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:38Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.347353 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tgdlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2336955b-a703-4db4-8d0f-d08088ed59b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb0362942bfac635ee6292501b1b851772c01d75dd658eef4b17669a716de72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vjfp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tgdlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:38Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.360374 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06f949936bac99cfbf89d720c9549c074351d9077bd963fa024d87004c6be23e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:38Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.370995 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384d8ba58264ab89796d3b53afb0c4c1be33cb4fba719a8beef4afc2be7d66a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:38Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.383493 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:38Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.395196 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17b8714be3bd121e06641873e003f08560b1f71e56d6ea936966df7d4896eb03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85731e6b6dbe1777bc7161a010f39f79523ec4b08d71d5d8a975573aa7fe4957\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:38Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.404581 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zlk96" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d55aa91-9cb9-4c95-bc27-6fec05684c10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4521b5baee45ffc9c7661c781c0b3de458dff325ad525dde283d46af71c0cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vkd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zlk96\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:38Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.421484 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d9e58df-1755-4422-8d04-211f98c7c011\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea762f627aa69fb88a9bd82545fa677e9ccfa97f99a5e92c0aa490e619c0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a06b811e5ec52ce984a2403ba2c98332ba60466d08b3580ad9192fac6caef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0385f6ed9597974267f7c45547f6b6725b4ee83a2cc0508ff2d1c1ce8753ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b915b565f7fbb3e4e856e7f1035a519e82bfb3023660c121f6e3813de353bb92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18465f3a6916c36457a305ba5d568c9077f34f1d000532ed518f237d100008b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:38Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.434151 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.434203 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.434246 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.434265 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.434277 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:38Z","lastTransitionTime":"2025-11-25T10:32:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.435464 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25595ccd-762d-414a-9916-4fe38d478b07\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d4b2d8db4d13392d9789e161fd4b8f6eef983b79128ee584d72e4c06565bc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf780f791771ea43a8c8176d02fdfb5204406e023207299369ae46ceb342baaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0251519fd4262f2ea5c64959db49433afdafe36b7ed2404d660eeb7c365d2746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad5c205e0eeecfd43cbdf0eb1a334e3e8faa5e9a9e8dd202c397d5cbb36e081e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:38Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.446537 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:38Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.457775 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ldzbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b84e32ae-4897-4644-a656-6bd82012cc81\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ef433744761f26e8e1f6c0db123821a11923ffd774830744285cd4178d89533\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5pmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ldzbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:38Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.476003 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"954085fa-d03a-4e88-9dfd-566257e3558d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f00746930c4588e548f0e2b60d52bf448aff7401b676a7855cb0f4cce00d329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a14a078ba7d2aed95992b36502008f315f0d3a634b64ed49da70398b2562d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d7254196b3702e48083ae98e829b73c539adf69ed5b2bbdc62189d09ae1762\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d727430108977574365f0e9c04b44e03307bd7ff89cdbb1e48a6e916f504d6b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94c7949f077b1ca2b310bedb090aa72fc4bd7e8f2c8bf438096edfca8568f36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed2e7cad7bb24b5b56a1295098a6a7c230c8f9b9c29a88ef653ae8b6ac31667d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4ea2031b69dc207c47ef71a22c04431f4d2dab95e183b3768dd11594dd6bb32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dbf499f4ae6161752abaa7a466a027c6f874016be90a534e69f07e3dd3f3d9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzhnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:38Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.487774 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06f949936bac99cfbf89d720c9549c074351d9077bd963fa024d87004c6be23e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:38Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.499482 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384d8ba58264ab89796d3b53afb0c4c1be33cb4fba719a8beef4afc2be7d66a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:38Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.510776 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:38Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.520422 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17b8714be3bd121e06641873e003f08560b1f71e56d6ea936966df7d4896eb03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85731e6b6dbe1777bc7161a010f39f79523ec4b08d71d5d8a975573aa7fe4957\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:38Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.529540 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zlk96" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d55aa91-9cb9-4c95-bc27-6fec05684c10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4521b5baee45ffc9c7661c781c0b3de458dff325ad525dde283d46af71c0cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vkd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zlk96\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:38Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.537040 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.537083 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.537095 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.537112 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.537125 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:38Z","lastTransitionTime":"2025-11-25T10:32:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.544950 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"954085fa-d03a-4e88-9dfd-566257e3558d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f00746930c4588e548f0e2b60d52bf448aff7401b676a7855cb0f4cce00d329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a14a078ba7d2aed95992b36502008f315f0d3a634b64ed49da70398b2562d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d7254196b3702e48083ae98e829b73c539adf69ed5b2bbdc62189d09ae1762\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d727430108977574365f0e9c04b44e03307bd7ff89cdbb1e48a6e916f504d6b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94c7949f077b1ca2b310bedb090aa72fc4bd7e8f2c8bf438096edfca8568f36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed2e7cad7bb24b5b56a1295098a6a7c230c8f9b9c29a88ef653ae8b6ac31667d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4ea2031b69dc207c47ef71a22c04431f4d2dab95e183b3768dd11594dd6bb32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dbf499f4ae6161752abaa7a466a027c6f874016be90a534e69f07e3dd3f3d9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzhnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:38Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.563439 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d9e58df-1755-4422-8d04-211f98c7c011\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea762f627aa69fb88a9bd82545fa677e9ccfa97f99a5e92c0aa490e619c0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a06b811e5ec52ce984a2403ba2c98332ba60466d08b3580ad9192fac6caef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0385f6ed9597974267f7c45547f6b6725b4ee83a2cc0508ff2d1c1ce8753ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b915b565f7fbb3e4e856e7f1035a519e82bfb3023660c121f6e3813de353bb92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18465f3a6916c36457a305ba5d568c9077f34f1d000532ed518f237d100008b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:38Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.573921 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25595ccd-762d-414a-9916-4fe38d478b07\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d4b2d8db4d13392d9789e161fd4b8f6eef983b79128ee584d72e4c06565bc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf780f791771ea43a8c8176d02fdfb5204406e023207299369ae46ceb342baaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0251519fd4262f2ea5c64959db49433afdafe36b7ed2404d660eeb7c365d2746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad5c205e0eeecfd43cbdf0eb1a334e3e8faa5e9a9e8dd202c397d5cbb36e081e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:38Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.584944 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:38Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.597074 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ldzbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b84e32ae-4897-4644-a656-6bd82012cc81\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ef433744761f26e8e1f6c0db123821a11923ffd774830744285cd4178d89533\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5pmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ldzbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:38Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.607748 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tgdlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2336955b-a703-4db4-8d0f-d08088ed59b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb0362942bfac635ee6292501b1b851772c01d75dd658eef4b17669a716de72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vjfp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tgdlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:38Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.619082 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62fcb632-f12a-46e2-bb7d-a5311f291ee7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28d43531927e75c3745f100940f7f6bf79fd21026c1309194df8cd7a39361b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69633332d03d2551ff4abb253d358de398eab18486b92db630ce06c0447fb9cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ccb78d9a51ac83ee208c2451685a2512de9deaf742a663bb8c892aa59b76162\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27826e155a9352b1e1866e83fdab683dbcd749cc04b6a4da2db25ecabe3d9a52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://991252e5f56f128827a5054b8e228284bc0266166582cbaedaadc53ca2f5b401\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:38Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.629523 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:38Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.639630 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.639681 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.639692 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.639711 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.639725 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:38Z","lastTransitionTime":"2025-11-25T10:32:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.641270 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f948b87-ac86-4de6-ad64-c2ef947f84d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4482f73a45bc82bd55770c3fd788becacec94893fed1c305fed1c81c753d43aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ba68bd093be94d0d40f3851638db5d26bdfc52d175b1ac78afb91dda7affcfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2krbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:38Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.654328 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qppnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"357491b8-cd9c-486e-9898-5ba33678513d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd9c047c5ce1d896f803ff6cafa9bbf3a18a391b44539e1f1312e08c89e51cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd9c047c5ce1d896f803ff6cafa9bbf3a18a391b44539e1f1312e08c89e51cb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://565a6db63fb69abbba81ffdf6085797900858d263cd6affcb0b5f195474c2516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://565a6db63fb69abbba81ffdf6085797900858d263cd6affcb0b5f195474c2516\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://544885302dffff8336a9aa3c4afa80d69088ffaa2402648f4a329e0a798a8c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://544885302dffff8336a9aa3c4afa80d69088ffaa2402648f4a329e0a798a8c99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8bc193fb3f85b980c31af2e826cee90adb633ad2ee82ff1d46e182ab52e8fab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8bc193fb3f85b980c31af2e826cee90adb633ad2ee82ff1d46e182ab52e8fab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7713e2a35bac8df1e8987fbcd452459fa3541caad0c52ed57fda187bd7c83d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7713e2a35bac8df1e8987fbcd452459fa3541caad0c52ed57fda187bd7c83d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34adfca89026fc20007a73a8cbd4e7c7b3137a581c3476867af17fabe976f71f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34adfca89026fc20007a73a8cbd4e7c7b3137a581c3476867af17fabe976f71f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qppnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:38Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.742273 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.742315 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.742325 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.742340 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.742350 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:38Z","lastTransitionTime":"2025-11-25T10:32:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.845133 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.845226 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.845240 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.845585 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.845622 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:38Z","lastTransitionTime":"2025-11-25T10:32:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.948804 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.948870 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.948881 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.948904 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:38 crc kubenswrapper[4821]: I1125 10:32:38.948920 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:38Z","lastTransitionTime":"2025-11-25T10:32:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:39 crc kubenswrapper[4821]: I1125 10:32:39.050868 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:39 crc kubenswrapper[4821]: I1125 10:32:39.050908 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:39 crc kubenswrapper[4821]: I1125 10:32:39.050919 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:39 crc kubenswrapper[4821]: I1125 10:32:39.050939 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:39 crc kubenswrapper[4821]: I1125 10:32:39.050953 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:39Z","lastTransitionTime":"2025-11-25T10:32:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:39 crc kubenswrapper[4821]: I1125 10:32:39.114096 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:32:39 crc kubenswrapper[4821]: E1125 10:32:39.114230 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:32:39 crc kubenswrapper[4821]: I1125 10:32:39.114639 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:32:39 crc kubenswrapper[4821]: E1125 10:32:39.114695 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:32:39 crc kubenswrapper[4821]: I1125 10:32:39.114735 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:32:39 crc kubenswrapper[4821]: E1125 10:32:39.114775 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:32:39 crc kubenswrapper[4821]: I1125 10:32:39.153104 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:39 crc kubenswrapper[4821]: I1125 10:32:39.153142 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:39 crc kubenswrapper[4821]: I1125 10:32:39.153151 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:39 crc kubenswrapper[4821]: I1125 10:32:39.153181 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:39 crc kubenswrapper[4821]: I1125 10:32:39.153192 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:39Z","lastTransitionTime":"2025-11-25T10:32:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:39 crc kubenswrapper[4821]: I1125 10:32:39.256328 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:39 crc kubenswrapper[4821]: I1125 10:32:39.256449 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:39 crc kubenswrapper[4821]: I1125 10:32:39.256558 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:39 crc kubenswrapper[4821]: I1125 10:32:39.256591 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:39 crc kubenswrapper[4821]: I1125 10:32:39.256604 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:39Z","lastTransitionTime":"2025-11-25T10:32:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:39 crc kubenswrapper[4821]: I1125 10:32:39.286208 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-qppnw" event={"ID":"357491b8-cd9c-486e-9898-5ba33678513d","Type":"ContainerStarted","Data":"dc88d700ca3a449ef1cb32d6124b16d0024d6981ced12e276083f86a1b322ba9"} Nov 25 10:32:39 crc kubenswrapper[4821]: I1125 10:32:39.306462 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d9e58df-1755-4422-8d04-211f98c7c011\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea762f627aa69fb88a9bd82545fa677e9ccfa97f99a5e92c0aa490e619c0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a06b811e5ec52ce984a2403ba2c98332ba60466d08b3580ad9192fac6caef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0385f6ed9597974267f7c45547f6b6725b4ee83a2cc0508ff2d1c1ce8753ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b915b565f7fbb3e4e856e7f1035a519e82bfb3023660c121f6e3813de353bb92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18465f3a6916c36457a305ba5d568c9077f34f1d000532ed518f237d100008b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:39Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:39 crc kubenswrapper[4821]: I1125 10:32:39.318668 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25595ccd-762d-414a-9916-4fe38d478b07\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d4b2d8db4d13392d9789e161fd4b8f6eef983b79128ee584d72e4c06565bc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf780f791771ea43a8c8176d02fdfb5204406e023207299369ae46ceb342baaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0251519fd4262f2ea5c64959db49433afdafe36b7ed2404d660eeb7c365d2746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad5c205e0eeecfd43cbdf0eb1a334e3e8faa5e9a9e8dd202c397d5cbb36e081e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:39Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:39 crc kubenswrapper[4821]: I1125 10:32:39.331334 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:39Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:39 crc kubenswrapper[4821]: I1125 10:32:39.345212 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ldzbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b84e32ae-4897-4644-a656-6bd82012cc81\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ef433744761f26e8e1f6c0db123821a11923ffd774830744285cd4178d89533\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5pmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ldzbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:39Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:39 crc kubenswrapper[4821]: I1125 10:32:39.359223 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:39 crc kubenswrapper[4821]: I1125 10:32:39.359277 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:39 crc kubenswrapper[4821]: I1125 10:32:39.359289 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:39 crc kubenswrapper[4821]: I1125 10:32:39.359303 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:39 crc kubenswrapper[4821]: I1125 10:32:39.359312 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:39Z","lastTransitionTime":"2025-11-25T10:32:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:39 crc kubenswrapper[4821]: I1125 10:32:39.366264 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"954085fa-d03a-4e88-9dfd-566257e3558d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f00746930c4588e548f0e2b60d52bf448aff7401b676a7855cb0f4cce00d329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a14a078ba7d2aed95992b36502008f315f0d3a634b64ed49da70398b2562d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d7254196b3702e48083ae98e829b73c539adf69ed5b2bbdc62189d09ae1762\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d727430108977574365f0e9c04b44e03307bd7ff89cdbb1e48a6e916f504d6b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94c7949f077b1ca2b310bedb090aa72fc4bd7e8f2c8bf438096edfca8568f36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed2e7cad7bb24b5b56a1295098a6a7c230c8f9b9c29a88ef653ae8b6ac31667d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4ea2031b69dc207c47ef71a22c04431f4d2dab95e183b3768dd11594dd6bb32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dbf499f4ae6161752abaa7a466a027c6f874016be90a534e69f07e3dd3f3d9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzhnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:39Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:39 crc kubenswrapper[4821]: I1125 10:32:39.378378 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62fcb632-f12a-46e2-bb7d-a5311f291ee7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28d43531927e75c3745f100940f7f6bf79fd21026c1309194df8cd7a39361b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69633332d03d2551ff4abb253d358de398eab18486b92db630ce06c0447fb9cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ccb78d9a51ac83ee208c2451685a2512de9deaf742a663bb8c892aa59b76162\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27826e155a9352b1e1866e83fdab683dbcd749cc04b6a4da2db25ecabe3d9a52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://991252e5f56f128827a5054b8e228284bc0266166582cbaedaadc53ca2f5b401\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:39Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:39 crc kubenswrapper[4821]: I1125 10:32:39.390557 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:39Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:39 crc kubenswrapper[4821]: I1125 10:32:39.402717 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f948b87-ac86-4de6-ad64-c2ef947f84d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4482f73a45bc82bd55770c3fd788becacec94893fed1c305fed1c81c753d43aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ba68bd093be94d0d40f3851638db5d26bdfc52d175b1ac78afb91dda7affcfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2krbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:39Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:39 crc kubenswrapper[4821]: I1125 10:32:39.415420 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qppnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"357491b8-cd9c-486e-9898-5ba33678513d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc88d700ca3a449ef1cb32d6124b16d0024d6981ced12e276083f86a1b322ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd9c047c5ce1d896f803ff6cafa9bbf3a18a391b44539e1f1312e08c89e51cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd9c047c5ce1d896f803ff6cafa9bbf3a18a391b44539e1f1312e08c89e51cb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://565a6db63fb69abbba81ffdf6085797900858d263cd6affcb0b5f195474c2516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://565a6db63fb69abbba81ffdf6085797900858d263cd6affcb0b5f195474c2516\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://544885302dffff8336a9aa3c4afa80d69088ffaa2402648f4a329e0a798a8c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://544885302dffff8336a9aa3c4afa80d69088ffaa2402648f4a329e0a798a8c99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8bc193fb3f85b980c31af2e826cee90adb633ad2ee82ff1d46e182ab52e8fab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8bc193fb3f85b980c31af2e826cee90adb633ad2ee82ff1d46e182ab52e8fab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7713e2a35bac8df1e8987fbcd452459fa3541caad0c52ed57fda187bd7c83d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7713e2a35bac8df1e8987fbcd452459fa3541caad0c52ed57fda187bd7c83d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34adfca89026fc20007a73a8cbd4e7c7b3137a581c3476867af17fabe976f71f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34adfca89026fc20007a73a8cbd4e7c7b3137a581c3476867af17fabe976f71f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qppnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:39Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:39 crc kubenswrapper[4821]: I1125 10:32:39.424268 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tgdlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2336955b-a703-4db4-8d0f-d08088ed59b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb0362942bfac635ee6292501b1b851772c01d75dd658eef4b17669a716de72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vjfp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tgdlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:39Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:39 crc kubenswrapper[4821]: I1125 10:32:39.435675 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06f949936bac99cfbf89d720c9549c074351d9077bd963fa024d87004c6be23e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:39Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:39 crc kubenswrapper[4821]: I1125 10:32:39.444522 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384d8ba58264ab89796d3b53afb0c4c1be33cb4fba719a8beef4afc2be7d66a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:39Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:39 crc kubenswrapper[4821]: I1125 10:32:39.454546 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:39Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:39 crc kubenswrapper[4821]: I1125 10:32:39.461237 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:39 crc kubenswrapper[4821]: I1125 10:32:39.461271 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:39 crc kubenswrapper[4821]: I1125 10:32:39.461281 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:39 crc kubenswrapper[4821]: I1125 10:32:39.461296 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:39 crc kubenswrapper[4821]: I1125 10:32:39.461306 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:39Z","lastTransitionTime":"2025-11-25T10:32:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:39 crc kubenswrapper[4821]: I1125 10:32:39.465613 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17b8714be3bd121e06641873e003f08560b1f71e56d6ea936966df7d4896eb03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85731e6b6dbe1777bc7161a010f39f79523ec4b08d71d5d8a975573aa7fe4957\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:39Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:39 crc kubenswrapper[4821]: I1125 10:32:39.474758 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zlk96" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d55aa91-9cb9-4c95-bc27-6fec05684c10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4521b5baee45ffc9c7661c781c0b3de458dff325ad525dde283d46af71c0cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vkd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zlk96\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:39Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:39 crc kubenswrapper[4821]: I1125 10:32:39.563338 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:39 crc kubenswrapper[4821]: I1125 10:32:39.563672 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:39 crc kubenswrapper[4821]: I1125 10:32:39.563685 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:39 crc kubenswrapper[4821]: I1125 10:32:39.563701 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:39 crc kubenswrapper[4821]: I1125 10:32:39.563713 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:39Z","lastTransitionTime":"2025-11-25T10:32:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:39 crc kubenswrapper[4821]: I1125 10:32:39.665848 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:39 crc kubenswrapper[4821]: I1125 10:32:39.665886 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:39 crc kubenswrapper[4821]: I1125 10:32:39.665897 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:39 crc kubenswrapper[4821]: I1125 10:32:39.665914 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:39 crc kubenswrapper[4821]: I1125 10:32:39.665924 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:39Z","lastTransitionTime":"2025-11-25T10:32:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:39 crc kubenswrapper[4821]: I1125 10:32:39.767741 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:39 crc kubenswrapper[4821]: I1125 10:32:39.767782 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:39 crc kubenswrapper[4821]: I1125 10:32:39.767794 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:39 crc kubenswrapper[4821]: I1125 10:32:39.767810 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:39 crc kubenswrapper[4821]: I1125 10:32:39.767819 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:39Z","lastTransitionTime":"2025-11-25T10:32:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:39 crc kubenswrapper[4821]: I1125 10:32:39.869829 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:39 crc kubenswrapper[4821]: I1125 10:32:39.869877 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:39 crc kubenswrapper[4821]: I1125 10:32:39.869889 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:39 crc kubenswrapper[4821]: I1125 10:32:39.869907 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:39 crc kubenswrapper[4821]: I1125 10:32:39.869918 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:39Z","lastTransitionTime":"2025-11-25T10:32:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:39 crc kubenswrapper[4821]: I1125 10:32:39.971708 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:39 crc kubenswrapper[4821]: I1125 10:32:39.971740 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:39 crc kubenswrapper[4821]: I1125 10:32:39.971748 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:39 crc kubenswrapper[4821]: I1125 10:32:39.971761 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:39 crc kubenswrapper[4821]: I1125 10:32:39.971771 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:39Z","lastTransitionTime":"2025-11-25T10:32:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.073952 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.073987 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.074025 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.074039 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.074050 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:40Z","lastTransitionTime":"2025-11-25T10:32:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.126234 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06f949936bac99cfbf89d720c9549c074351d9077bd963fa024d87004c6be23e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:40Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.137036 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384d8ba58264ab89796d3b53afb0c4c1be33cb4fba719a8beef4afc2be7d66a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:40Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.145592 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zlk96" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d55aa91-9cb9-4c95-bc27-6fec05684c10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4521b5baee45ffc9c7661c781c0b3de458dff325ad525dde283d46af71c0cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vkd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zlk96\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:40Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.175657 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:40Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.175713 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.175731 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.175741 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.175754 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.175762 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:40Z","lastTransitionTime":"2025-11-25T10:32:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.213855 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17b8714be3bd121e06641873e003f08560b1f71e56d6ea936966df7d4896eb03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85731e6b6dbe1777bc7161a010f39f79523ec4b08d71d5d8a975573aa7fe4957\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:40Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.229184 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25595ccd-762d-414a-9916-4fe38d478b07\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d4b2d8db4d13392d9789e161fd4b8f6eef983b79128ee584d72e4c06565bc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf780f791771ea43a8c8176d02fdfb5204406e023207299369ae46ceb342baaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0251519fd4262f2ea5c64959db49433afdafe36b7ed2404d660eeb7c365d2746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad5c205e0eeecfd43cbdf0eb1a334e3e8faa5e9a9e8dd202c397d5cbb36e081e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:40Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.241523 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:40Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.254669 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ldzbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b84e32ae-4897-4644-a656-6bd82012cc81\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ef433744761f26e8e1f6c0db123821a11923ffd774830744285cd4178d89533\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5pmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ldzbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:40Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.272841 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"954085fa-d03a-4e88-9dfd-566257e3558d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f00746930c4588e548f0e2b60d52bf448aff7401b676a7855cb0f4cce00d329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a14a078ba7d2aed95992b36502008f315f0d3a634b64ed49da70398b2562d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d7254196b3702e48083ae98e829b73c539adf69ed5b2bbdc62189d09ae1762\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d727430108977574365f0e9c04b44e03307bd7ff89cdbb1e48a6e916f504d6b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94c7949f077b1ca2b310bedb090aa72fc4bd7e8f2c8bf438096edfca8568f36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed2e7cad7bb24b5b56a1295098a6a7c230c8f9b9c29a88ef653ae8b6ac31667d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4ea2031b69dc207c47ef71a22c04431f4d2dab95e183b3768dd11594dd6bb32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dbf499f4ae6161752abaa7a466a027c6f874016be90a534e69f07e3dd3f3d9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzhnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:40Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.278122 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.278176 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.278190 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.278208 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.278220 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:40Z","lastTransitionTime":"2025-11-25T10:32:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.289674 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mzhnw_954085fa-d03a-4e88-9dfd-566257e3558d/ovnkube-controller/0.log" Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.292265 4821 generic.go:334] "Generic (PLEG): container finished" podID="954085fa-d03a-4e88-9dfd-566257e3558d" containerID="a4ea2031b69dc207c47ef71a22c04431f4d2dab95e183b3768dd11594dd6bb32" exitCode=1 Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.292346 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" event={"ID":"954085fa-d03a-4e88-9dfd-566257e3558d","Type":"ContainerDied","Data":"a4ea2031b69dc207c47ef71a22c04431f4d2dab95e183b3768dd11594dd6bb32"} Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.293295 4821 scope.go:117] "RemoveContainer" containerID="a4ea2031b69dc207c47ef71a22c04431f4d2dab95e183b3768dd11594dd6bb32" Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.298602 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d9e58df-1755-4422-8d04-211f98c7c011\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea762f627aa69fb88a9bd82545fa677e9ccfa97f99a5e92c0aa490e619c0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a06b811e5ec52ce984a2403ba2c98332ba60466d08b3580ad9192fac6caef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0385f6ed9597974267f7c45547f6b6725b4ee83a2cc0508ff2d1c1ce8753ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b915b565f7fbb3e4e856e7f1035a519e82bfb3023660c121f6e3813de353bb92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18465f3a6916c36457a305ba5d568c9077f34f1d000532ed518f237d100008b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:40Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.316878 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:40Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.327575 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f948b87-ac86-4de6-ad64-c2ef947f84d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4482f73a45bc82bd55770c3fd788becacec94893fed1c305fed1c81c753d43aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ba68bd093be94d0d40f3851638db5d26bdfc52d175b1ac78afb91dda7affcfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2krbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:40Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.340802 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qppnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"357491b8-cd9c-486e-9898-5ba33678513d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc88d700ca3a449ef1cb32d6124b16d0024d6981ced12e276083f86a1b322ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd9c047c5ce1d896f803ff6cafa9bbf3a18a391b44539e1f1312e08c89e51cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd9c047c5ce1d896f803ff6cafa9bbf3a18a391b44539e1f1312e08c89e51cb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://565a6db63fb69abbba81ffdf6085797900858d263cd6affcb0b5f195474c2516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://565a6db63fb69abbba81ffdf6085797900858d263cd6affcb0b5f195474c2516\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://544885302dffff8336a9aa3c4afa80d69088ffaa2402648f4a329e0a798a8c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://544885302dffff8336a9aa3c4afa80d69088ffaa2402648f4a329e0a798a8c99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8bc193fb3f85b980c31af2e826cee90adb633ad2ee82ff1d46e182ab52e8fab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8bc193fb3f85b980c31af2e826cee90adb633ad2ee82ff1d46e182ab52e8fab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7713e2a35bac8df1e8987fbcd452459fa3541caad0c52ed57fda187bd7c83d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7713e2a35bac8df1e8987fbcd452459fa3541caad0c52ed57fda187bd7c83d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34adfca89026fc20007a73a8cbd4e7c7b3137a581c3476867af17fabe976f71f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34adfca89026fc20007a73a8cbd4e7c7b3137a581c3476867af17fabe976f71f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qppnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:40Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.349974 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tgdlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2336955b-a703-4db4-8d0f-d08088ed59b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb0362942bfac635ee6292501b1b851772c01d75dd658eef4b17669a716de72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vjfp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tgdlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:40Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.365744 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62fcb632-f12a-46e2-bb7d-a5311f291ee7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28d43531927e75c3745f100940f7f6bf79fd21026c1309194df8cd7a39361b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69633332d03d2551ff4abb253d358de398eab18486b92db630ce06c0447fb9cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ccb78d9a51ac83ee208c2451685a2512de9deaf742a663bb8c892aa59b76162\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27826e155a9352b1e1866e83fdab683dbcd749cc04b6a4da2db25ecabe3d9a52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://991252e5f56f128827a5054b8e228284bc0266166582cbaedaadc53ca2f5b401\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:40Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.378379 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:40Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.380368 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.380404 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.380415 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.380431 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.380441 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:40Z","lastTransitionTime":"2025-11-25T10:32:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.390883 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17b8714be3bd121e06641873e003f08560b1f71e56d6ea936966df7d4896eb03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85731e6b6dbe1777bc7161a010f39f79523ec4b08d71d5d8a975573aa7fe4957\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:40Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.402447 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zlk96" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d55aa91-9cb9-4c95-bc27-6fec05684c10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4521b5baee45ffc9c7661c781c0b3de458dff325ad525dde283d46af71c0cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vkd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zlk96\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:40Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.424719 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d9e58df-1755-4422-8d04-211f98c7c011\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea762f627aa69fb88a9bd82545fa677e9ccfa97f99a5e92c0aa490e619c0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a06b811e5ec52ce984a2403ba2c98332ba60466d08b3580ad9192fac6caef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0385f6ed9597974267f7c45547f6b6725b4ee83a2cc0508ff2d1c1ce8753ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b915b565f7fbb3e4e856e7f1035a519e82bfb3023660c121f6e3813de353bb92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18465f3a6916c36457a305ba5d568c9077f34f1d000532ed518f237d100008b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:40Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.437810 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25595ccd-762d-414a-9916-4fe38d478b07\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d4b2d8db4d13392d9789e161fd4b8f6eef983b79128ee584d72e4c06565bc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf780f791771ea43a8c8176d02fdfb5204406e023207299369ae46ceb342baaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0251519fd4262f2ea5c64959db49433afdafe36b7ed2404d660eeb7c365d2746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad5c205e0eeecfd43cbdf0eb1a334e3e8faa5e9a9e8dd202c397d5cbb36e081e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:40Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.453647 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:40Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.474429 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ldzbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b84e32ae-4897-4644-a656-6bd82012cc81\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ef433744761f26e8e1f6c0db123821a11923ffd774830744285cd4178d89533\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5pmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ldzbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:40Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.482438 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.482473 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.482482 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.482495 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.482504 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:40Z","lastTransitionTime":"2025-11-25T10:32:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.494414 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"954085fa-d03a-4e88-9dfd-566257e3558d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f00746930c4588e548f0e2b60d52bf448aff7401b676a7855cb0f4cce00d329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a14a078ba7d2aed95992b36502008f315f0d3a634b64ed49da70398b2562d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d7254196b3702e48083ae98e829b73c539adf69ed5b2bbdc62189d09ae1762\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d727430108977574365f0e9c04b44e03307bd7ff89cdbb1e48a6e916f504d6b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94c7949f077b1ca2b310bedb090aa72fc4bd7e8f2c8bf438096edfca8568f36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed2e7cad7bb24b5b56a1295098a6a7c230c8f9b9c29a88ef653ae8b6ac31667d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4ea2031b69dc207c47ef71a22c04431f4d2dab95e183b3768dd11594dd6bb32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4ea2031b69dc207c47ef71a22c04431f4d2dab95e183b3768dd11594dd6bb32\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T10:32:39Z\\\",\\\"message\\\":\\\"vent handler 5\\\\nI1125 10:32:39.744416 6066 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 10:32:39.744433 6066 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 10:32:39.744443 6066 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 10:32:39.744446 6066 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1125 10:32:39.744451 6066 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 10:32:39.744456 6066 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 10:32:39.744495 6066 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1125 10:32:39.744540 6066 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1125 10:32:39.744645 6066 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1125 10:32:39.744727 6066 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1125 10:32:39.744958 6066 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dbf499f4ae6161752abaa7a466a027c6f874016be90a534e69f07e3dd3f3d9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzhnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:40Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.509045 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62fcb632-f12a-46e2-bb7d-a5311f291ee7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28d43531927e75c3745f100940f7f6bf79fd21026c1309194df8cd7a39361b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69633332d03d2551ff4abb253d358de398eab18486b92db630ce06c0447fb9cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ccb78d9a51ac83ee208c2451685a2512de9deaf742a663bb8c892aa59b76162\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27826e155a9352b1e1866e83fdab683dbcd749cc04b6a4da2db25ecabe3d9a52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://991252e5f56f128827a5054b8e228284bc0266166582cbaedaadc53ca2f5b401\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:40Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.521102 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:40Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.533436 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f948b87-ac86-4de6-ad64-c2ef947f84d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4482f73a45bc82bd55770c3fd788becacec94893fed1c305fed1c81c753d43aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ba68bd093be94d0d40f3851638db5d26bdfc52d175b1ac78afb91dda7affcfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2krbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:40Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.549362 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qppnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"357491b8-cd9c-486e-9898-5ba33678513d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc88d700ca3a449ef1cb32d6124b16d0024d6981ced12e276083f86a1b322ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd9c047c5ce1d896f803ff6cafa9bbf3a18a391b44539e1f1312e08c89e51cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd9c047c5ce1d896f803ff6cafa9bbf3a18a391b44539e1f1312e08c89e51cb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://565a6db63fb69abbba81ffdf6085797900858d263cd6affcb0b5f195474c2516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://565a6db63fb69abbba81ffdf6085797900858d263cd6affcb0b5f195474c2516\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://544885302dffff8336a9aa3c4afa80d69088ffaa2402648f4a329e0a798a8c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://544885302dffff8336a9aa3c4afa80d69088ffaa2402648f4a329e0a798a8c99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8bc193fb3f85b980c31af2e826cee90adb633ad2ee82ff1d46e182ab52e8fab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8bc193fb3f85b980c31af2e826cee90adb633ad2ee82ff1d46e182ab52e8fab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7713e2a35bac8df1e8987fbcd452459fa3541caad0c52ed57fda187bd7c83d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7713e2a35bac8df1e8987fbcd452459fa3541caad0c52ed57fda187bd7c83d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34adfca89026fc20007a73a8cbd4e7c7b3137a581c3476867af17fabe976f71f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34adfca89026fc20007a73a8cbd4e7c7b3137a581c3476867af17fabe976f71f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qppnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:40Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.565844 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tgdlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2336955b-a703-4db4-8d0f-d08088ed59b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb0362942bfac635ee6292501b1b851772c01d75dd658eef4b17669a716de72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vjfp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tgdlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:40Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.578702 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06f949936bac99cfbf89d720c9549c074351d9077bd963fa024d87004c6be23e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:40Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.584529 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.584564 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.584591 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.584607 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.584620 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:40Z","lastTransitionTime":"2025-11-25T10:32:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.589398 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384d8ba58264ab89796d3b53afb0c4c1be33cb4fba719a8beef4afc2be7d66a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:40Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.686501 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.686554 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.686565 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.686583 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.686594 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:40Z","lastTransitionTime":"2025-11-25T10:32:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.788592 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.788641 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.788652 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.788668 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.788676 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:40Z","lastTransitionTime":"2025-11-25T10:32:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.890725 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.890770 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.890788 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.890807 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.890820 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:40Z","lastTransitionTime":"2025-11-25T10:32:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.993540 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.993609 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.993619 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.993635 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:40 crc kubenswrapper[4821]: I1125 10:32:40.993644 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:40Z","lastTransitionTime":"2025-11-25T10:32:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:41 crc kubenswrapper[4821]: I1125 10:32:41.096383 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:41 crc kubenswrapper[4821]: I1125 10:32:41.096434 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:41 crc kubenswrapper[4821]: I1125 10:32:41.096451 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:41 crc kubenswrapper[4821]: I1125 10:32:41.096467 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:41 crc kubenswrapper[4821]: I1125 10:32:41.096477 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:41Z","lastTransitionTime":"2025-11-25T10:32:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:41 crc kubenswrapper[4821]: I1125 10:32:41.113690 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:32:41 crc kubenswrapper[4821]: I1125 10:32:41.113747 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:32:41 crc kubenswrapper[4821]: E1125 10:32:41.113792 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:32:41 crc kubenswrapper[4821]: I1125 10:32:41.113854 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:32:41 crc kubenswrapper[4821]: E1125 10:32:41.113895 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:32:41 crc kubenswrapper[4821]: E1125 10:32:41.114061 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:32:41 crc kubenswrapper[4821]: I1125 10:32:41.198729 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:41 crc kubenswrapper[4821]: I1125 10:32:41.198774 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:41 crc kubenswrapper[4821]: I1125 10:32:41.198786 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:41 crc kubenswrapper[4821]: I1125 10:32:41.198801 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:41 crc kubenswrapper[4821]: I1125 10:32:41.198812 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:41Z","lastTransitionTime":"2025-11-25T10:32:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:41 crc kubenswrapper[4821]: I1125 10:32:41.297056 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mzhnw_954085fa-d03a-4e88-9dfd-566257e3558d/ovnkube-controller/1.log" Nov 25 10:32:41 crc kubenswrapper[4821]: I1125 10:32:41.297567 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mzhnw_954085fa-d03a-4e88-9dfd-566257e3558d/ovnkube-controller/0.log" Nov 25 10:32:41 crc kubenswrapper[4821]: I1125 10:32:41.299698 4821 generic.go:334] "Generic (PLEG): container finished" podID="954085fa-d03a-4e88-9dfd-566257e3558d" containerID="5a84f30b6a8ad3333268312e9aadacb17eca12613511195c6f8a7734d5f7af5b" exitCode=1 Nov 25 10:32:41 crc kubenswrapper[4821]: I1125 10:32:41.299749 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" event={"ID":"954085fa-d03a-4e88-9dfd-566257e3558d","Type":"ContainerDied","Data":"5a84f30b6a8ad3333268312e9aadacb17eca12613511195c6f8a7734d5f7af5b"} Nov 25 10:32:41 crc kubenswrapper[4821]: I1125 10:32:41.299808 4821 scope.go:117] "RemoveContainer" containerID="a4ea2031b69dc207c47ef71a22c04431f4d2dab95e183b3768dd11594dd6bb32" Nov 25 10:32:41 crc kubenswrapper[4821]: I1125 10:32:41.300150 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:41 crc kubenswrapper[4821]: I1125 10:32:41.300211 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:41 crc kubenswrapper[4821]: I1125 10:32:41.300230 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:41 crc kubenswrapper[4821]: I1125 10:32:41.300250 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:41 crc kubenswrapper[4821]: I1125 10:32:41.300265 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:41Z","lastTransitionTime":"2025-11-25T10:32:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:41 crc kubenswrapper[4821]: I1125 10:32:41.300391 4821 scope.go:117] "RemoveContainer" containerID="5a84f30b6a8ad3333268312e9aadacb17eca12613511195c6f8a7734d5f7af5b" Nov 25 10:32:41 crc kubenswrapper[4821]: E1125 10:32:41.300555 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-mzhnw_openshift-ovn-kubernetes(954085fa-d03a-4e88-9dfd-566257e3558d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" podUID="954085fa-d03a-4e88-9dfd-566257e3558d" Nov 25 10:32:41 crc kubenswrapper[4821]: I1125 10:32:41.317002 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ldzbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b84e32ae-4897-4644-a656-6bd82012cc81\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ef433744761f26e8e1f6c0db123821a11923ffd774830744285cd4178d89533\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5pmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ldzbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:41Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:41 crc kubenswrapper[4821]: I1125 10:32:41.339955 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"954085fa-d03a-4e88-9dfd-566257e3558d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f00746930c4588e548f0e2b60d52bf448aff7401b676a7855cb0f4cce00d329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a14a078ba7d2aed95992b36502008f315f0d3a634b64ed49da70398b2562d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d7254196b3702e48083ae98e829b73c539adf69ed5b2bbdc62189d09ae1762\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d727430108977574365f0e9c04b44e03307bd7ff89cdbb1e48a6e916f504d6b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94c7949f077b1ca2b310bedb090aa72fc4bd7e8f2c8bf438096edfca8568f36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed2e7cad7bb24b5b56a1295098a6a7c230c8f9b9c29a88ef653ae8b6ac31667d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a84f30b6a8ad3333268312e9aadacb17eca12613511195c6f8a7734d5f7af5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4ea2031b69dc207c47ef71a22c04431f4d2dab95e183b3768dd11594dd6bb32\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T10:32:39Z\\\",\\\"message\\\":\\\"vent handler 5\\\\nI1125 10:32:39.744416 6066 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 10:32:39.744433 6066 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 10:32:39.744443 6066 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 10:32:39.744446 6066 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1125 10:32:39.744451 6066 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 10:32:39.744456 6066 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 10:32:39.744495 6066 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1125 10:32:39.744540 6066 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1125 10:32:39.744645 6066 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1125 10:32:39.744727 6066 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1125 10:32:39.744958 6066 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a84f30b6a8ad3333268312e9aadacb17eca12613511195c6f8a7734d5f7af5b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T10:32:41Z\\\",\\\"message\\\":\\\" 6233 obj_retry.go:365] Adding new object: *v1.Pod openshift-dns/node-resolver-zlk96\\\\nI1125 10:32:41.011897 6233 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-target-xd92c in node crc\\\\nI1125 10:32:41.011901 6233 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-additional-cni-plugins-qppnw after 0 failed attempt(s)\\\\nI1125 10:32:41.011815 6233 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1125 10:32:41.011915 6233 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-additional-cni-plugins-qppnw\\\\nI1125 10:32:41.011902 6233 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-zlk96 in node crc\\\\nI1125 10:32:41.011826 6233 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1125 10:32:41.011923 6233 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-tgdlt after 0 failed attempt(s)\\\\nI1125 10:32:41.011884 6233 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-ingress-operator/metrics]} name:Service_openshift-ingress-operator/metrics_\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dbf499f4ae6161752abaa7a466a027c6f874016be90a534e69f07e3dd3f3d9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzhnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:41Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:41 crc kubenswrapper[4821]: I1125 10:32:41.362091 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d9e58df-1755-4422-8d04-211f98c7c011\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea762f627aa69fb88a9bd82545fa677e9ccfa97f99a5e92c0aa490e619c0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a06b811e5ec52ce984a2403ba2c98332ba60466d08b3580ad9192fac6caef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0385f6ed9597974267f7c45547f6b6725b4ee83a2cc0508ff2d1c1ce8753ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b915b565f7fbb3e4e856e7f1035a519e82bfb3023660c121f6e3813de353bb92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18465f3a6916c36457a305ba5d568c9077f34f1d000532ed518f237d100008b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:41Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:41 crc kubenswrapper[4821]: I1125 10:32:41.374847 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25595ccd-762d-414a-9916-4fe38d478b07\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d4b2d8db4d13392d9789e161fd4b8f6eef983b79128ee584d72e4c06565bc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf780f791771ea43a8c8176d02fdfb5204406e023207299369ae46ceb342baaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0251519fd4262f2ea5c64959db49433afdafe36b7ed2404d660eeb7c365d2746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad5c205e0eeecfd43cbdf0eb1a334e3e8faa5e9a9e8dd202c397d5cbb36e081e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:41Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:41 crc kubenswrapper[4821]: I1125 10:32:41.388016 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:41Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:41 crc kubenswrapper[4821]: I1125 10:32:41.401716 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qppnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"357491b8-cd9c-486e-9898-5ba33678513d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc88d700ca3a449ef1cb32d6124b16d0024d6981ced12e276083f86a1b322ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd9c047c5ce1d896f803ff6cafa9bbf3a18a391b44539e1f1312e08c89e51cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd9c047c5ce1d896f803ff6cafa9bbf3a18a391b44539e1f1312e08c89e51cb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://565a6db63fb69abbba81ffdf6085797900858d263cd6affcb0b5f195474c2516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://565a6db63fb69abbba81ffdf6085797900858d263cd6affcb0b5f195474c2516\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://544885302dffff8336a9aa3c4afa80d69088ffaa2402648f4a329e0a798a8c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://544885302dffff8336a9aa3c4afa80d69088ffaa2402648f4a329e0a798a8c99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8bc193fb3f85b980c31af2e826cee90adb633ad2ee82ff1d46e182ab52e8fab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8bc193fb3f85b980c31af2e826cee90adb633ad2ee82ff1d46e182ab52e8fab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7713e2a35bac8df1e8987fbcd452459fa3541caad0c52ed57fda187bd7c83d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7713e2a35bac8df1e8987fbcd452459fa3541caad0c52ed57fda187bd7c83d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34adfca89026fc20007a73a8cbd4e7c7b3137a581c3476867af17fabe976f71f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34adfca89026fc20007a73a8cbd4e7c7b3137a581c3476867af17fabe976f71f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qppnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:41Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:41 crc kubenswrapper[4821]: I1125 10:32:41.402857 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:41 crc kubenswrapper[4821]: I1125 10:32:41.402886 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:41 crc kubenswrapper[4821]: I1125 10:32:41.402898 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:41 crc kubenswrapper[4821]: I1125 10:32:41.402913 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:41 crc kubenswrapper[4821]: I1125 10:32:41.402923 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:41Z","lastTransitionTime":"2025-11-25T10:32:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:41 crc kubenswrapper[4821]: I1125 10:32:41.411271 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tgdlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2336955b-a703-4db4-8d0f-d08088ed59b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb0362942bfac635ee6292501b1b851772c01d75dd658eef4b17669a716de72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vjfp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tgdlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:41Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:41 crc kubenswrapper[4821]: I1125 10:32:41.426554 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62fcb632-f12a-46e2-bb7d-a5311f291ee7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28d43531927e75c3745f100940f7f6bf79fd21026c1309194df8cd7a39361b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69633332d03d2551ff4abb253d358de398eab18486b92db630ce06c0447fb9cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ccb78d9a51ac83ee208c2451685a2512de9deaf742a663bb8c892aa59b76162\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27826e155a9352b1e1866e83fdab683dbcd749cc04b6a4da2db25ecabe3d9a52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://991252e5f56f128827a5054b8e228284bc0266166582cbaedaadc53ca2f5b401\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:41Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:41 crc kubenswrapper[4821]: I1125 10:32:41.438301 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:41Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:41 crc kubenswrapper[4821]: I1125 10:32:41.449739 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f948b87-ac86-4de6-ad64-c2ef947f84d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4482f73a45bc82bd55770c3fd788becacec94893fed1c305fed1c81c753d43aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ba68bd093be94d0d40f3851638db5d26bdfc52d175b1ac78afb91dda7affcfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2krbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:41Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:41 crc kubenswrapper[4821]: I1125 10:32:41.461711 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06f949936bac99cfbf89d720c9549c074351d9077bd963fa024d87004c6be23e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:41Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:41 crc kubenswrapper[4821]: I1125 10:32:41.472576 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384d8ba58264ab89796d3b53afb0c4c1be33cb4fba719a8beef4afc2be7d66a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:41Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:41 crc kubenswrapper[4821]: I1125 10:32:41.483563 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:41Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:41 crc kubenswrapper[4821]: I1125 10:32:41.495592 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17b8714be3bd121e06641873e003f08560b1f71e56d6ea936966df7d4896eb03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85731e6b6dbe1777bc7161a010f39f79523ec4b08d71d5d8a975573aa7fe4957\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:41Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:41 crc kubenswrapper[4821]: I1125 10:32:41.505618 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:41 crc kubenswrapper[4821]: I1125 10:32:41.505667 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:41 crc kubenswrapper[4821]: I1125 10:32:41.505678 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:41 crc kubenswrapper[4821]: I1125 10:32:41.505699 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:41 crc kubenswrapper[4821]: I1125 10:32:41.505711 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:41Z","lastTransitionTime":"2025-11-25T10:32:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:41 crc kubenswrapper[4821]: I1125 10:32:41.505899 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zlk96" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d55aa91-9cb9-4c95-bc27-6fec05684c10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4521b5baee45ffc9c7661c781c0b3de458dff325ad525dde283d46af71c0cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vkd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zlk96\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:41Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:41 crc kubenswrapper[4821]: I1125 10:32:41.607924 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:41 crc kubenswrapper[4821]: I1125 10:32:41.607963 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:41 crc kubenswrapper[4821]: I1125 10:32:41.607973 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:41 crc kubenswrapper[4821]: I1125 10:32:41.607988 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:41 crc kubenswrapper[4821]: I1125 10:32:41.607998 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:41Z","lastTransitionTime":"2025-11-25T10:32:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:41 crc kubenswrapper[4821]: I1125 10:32:41.710461 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:41 crc kubenswrapper[4821]: I1125 10:32:41.710503 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:41 crc kubenswrapper[4821]: I1125 10:32:41.710512 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:41 crc kubenswrapper[4821]: I1125 10:32:41.710527 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:41 crc kubenswrapper[4821]: I1125 10:32:41.710539 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:41Z","lastTransitionTime":"2025-11-25T10:32:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:41 crc kubenswrapper[4821]: I1125 10:32:41.812148 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:41 crc kubenswrapper[4821]: I1125 10:32:41.812363 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:41 crc kubenswrapper[4821]: I1125 10:32:41.812379 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:41 crc kubenswrapper[4821]: I1125 10:32:41.812396 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:41 crc kubenswrapper[4821]: I1125 10:32:41.812405 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:41Z","lastTransitionTime":"2025-11-25T10:32:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:41 crc kubenswrapper[4821]: I1125 10:32:41.914491 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:41 crc kubenswrapper[4821]: I1125 10:32:41.914535 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:41 crc kubenswrapper[4821]: I1125 10:32:41.914545 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:41 crc kubenswrapper[4821]: I1125 10:32:41.914559 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:41 crc kubenswrapper[4821]: I1125 10:32:41.914572 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:41Z","lastTransitionTime":"2025-11-25T10:32:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:42 crc kubenswrapper[4821]: I1125 10:32:42.016754 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:42 crc kubenswrapper[4821]: I1125 10:32:42.016795 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:42 crc kubenswrapper[4821]: I1125 10:32:42.016806 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:42 crc kubenswrapper[4821]: I1125 10:32:42.016822 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:42 crc kubenswrapper[4821]: I1125 10:32:42.016833 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:42Z","lastTransitionTime":"2025-11-25T10:32:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:42 crc kubenswrapper[4821]: I1125 10:32:42.118846 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:42 crc kubenswrapper[4821]: I1125 10:32:42.118904 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:42 crc kubenswrapper[4821]: I1125 10:32:42.118913 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:42 crc kubenswrapper[4821]: I1125 10:32:42.118925 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:42 crc kubenswrapper[4821]: I1125 10:32:42.118935 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:42Z","lastTransitionTime":"2025-11-25T10:32:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:42 crc kubenswrapper[4821]: I1125 10:32:42.221074 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:42 crc kubenswrapper[4821]: I1125 10:32:42.221149 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:42 crc kubenswrapper[4821]: I1125 10:32:42.221211 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:42 crc kubenswrapper[4821]: I1125 10:32:42.221238 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:42 crc kubenswrapper[4821]: I1125 10:32:42.221254 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:42Z","lastTransitionTime":"2025-11-25T10:32:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:42 crc kubenswrapper[4821]: I1125 10:32:42.304579 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mzhnw_954085fa-d03a-4e88-9dfd-566257e3558d/ovnkube-controller/1.log" Nov 25 10:32:42 crc kubenswrapper[4821]: I1125 10:32:42.307917 4821 scope.go:117] "RemoveContainer" containerID="5a84f30b6a8ad3333268312e9aadacb17eca12613511195c6f8a7734d5f7af5b" Nov 25 10:32:42 crc kubenswrapper[4821]: E1125 10:32:42.308116 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-mzhnw_openshift-ovn-kubernetes(954085fa-d03a-4e88-9dfd-566257e3558d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" podUID="954085fa-d03a-4e88-9dfd-566257e3558d" Nov 25 10:32:42 crc kubenswrapper[4821]: I1125 10:32:42.322755 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ldzbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b84e32ae-4897-4644-a656-6bd82012cc81\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ef433744761f26e8e1f6c0db123821a11923ffd774830744285cd4178d89533\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5pmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ldzbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:42Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:42 crc kubenswrapper[4821]: I1125 10:32:42.323958 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:42 crc kubenswrapper[4821]: I1125 10:32:42.324009 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:42 crc kubenswrapper[4821]: I1125 10:32:42.324022 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:42 crc kubenswrapper[4821]: I1125 10:32:42.324047 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:42 crc kubenswrapper[4821]: I1125 10:32:42.324058 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:42Z","lastTransitionTime":"2025-11-25T10:32:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:42 crc kubenswrapper[4821]: I1125 10:32:42.344403 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"954085fa-d03a-4e88-9dfd-566257e3558d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f00746930c4588e548f0e2b60d52bf448aff7401b676a7855cb0f4cce00d329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a14a078ba7d2aed95992b36502008f315f0d3a634b64ed49da70398b2562d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d7254196b3702e48083ae98e829b73c539adf69ed5b2bbdc62189d09ae1762\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d727430108977574365f0e9c04b44e03307bd7ff89cdbb1e48a6e916f504d6b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94c7949f077b1ca2b310bedb090aa72fc4bd7e8f2c8bf438096edfca8568f36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed2e7cad7bb24b5b56a1295098a6a7c230c8f9b9c29a88ef653ae8b6ac31667d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a84f30b6a8ad3333268312e9aadacb17eca12613511195c6f8a7734d5f7af5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a84f30b6a8ad3333268312e9aadacb17eca12613511195c6f8a7734d5f7af5b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T10:32:41Z\\\",\\\"message\\\":\\\" 6233 obj_retry.go:365] Adding new object: *v1.Pod openshift-dns/node-resolver-zlk96\\\\nI1125 10:32:41.011897 6233 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-target-xd92c in node crc\\\\nI1125 10:32:41.011901 6233 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-additional-cni-plugins-qppnw after 0 failed attempt(s)\\\\nI1125 10:32:41.011815 6233 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1125 10:32:41.011915 6233 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-additional-cni-plugins-qppnw\\\\nI1125 10:32:41.011902 6233 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-zlk96 in node crc\\\\nI1125 10:32:41.011826 6233 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1125 10:32:41.011923 6233 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-tgdlt after 0 failed attempt(s)\\\\nI1125 10:32:41.011884 6233 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-ingress-operator/metrics]} name:Service_openshift-ingress-operator/metrics_\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-mzhnw_openshift-ovn-kubernetes(954085fa-d03a-4e88-9dfd-566257e3558d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dbf499f4ae6161752abaa7a466a027c6f874016be90a534e69f07e3dd3f3d9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzhnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:42Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:42 crc kubenswrapper[4821]: I1125 10:32:42.361804 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d9e58df-1755-4422-8d04-211f98c7c011\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea762f627aa69fb88a9bd82545fa677e9ccfa97f99a5e92c0aa490e619c0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a06b811e5ec52ce984a2403ba2c98332ba60466d08b3580ad9192fac6caef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0385f6ed9597974267f7c45547f6b6725b4ee83a2cc0508ff2d1c1ce8753ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b915b565f7fbb3e4e856e7f1035a519e82bfb3023660c121f6e3813de353bb92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18465f3a6916c36457a305ba5d568c9077f34f1d000532ed518f237d100008b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:42Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:42 crc kubenswrapper[4821]: I1125 10:32:42.372969 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25595ccd-762d-414a-9916-4fe38d478b07\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d4b2d8db4d13392d9789e161fd4b8f6eef983b79128ee584d72e4c06565bc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf780f791771ea43a8c8176d02fdfb5204406e023207299369ae46ceb342baaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0251519fd4262f2ea5c64959db49433afdafe36b7ed2404d660eeb7c365d2746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad5c205e0eeecfd43cbdf0eb1a334e3e8faa5e9a9e8dd202c397d5cbb36e081e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:42Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:42 crc kubenswrapper[4821]: I1125 10:32:42.387631 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:42Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:42 crc kubenswrapper[4821]: I1125 10:32:42.403913 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qppnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"357491b8-cd9c-486e-9898-5ba33678513d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc88d700ca3a449ef1cb32d6124b16d0024d6981ced12e276083f86a1b322ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd9c047c5ce1d896f803ff6cafa9bbf3a18a391b44539e1f1312e08c89e51cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd9c047c5ce1d896f803ff6cafa9bbf3a18a391b44539e1f1312e08c89e51cb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://565a6db63fb69abbba81ffdf6085797900858d263cd6affcb0b5f195474c2516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://565a6db63fb69abbba81ffdf6085797900858d263cd6affcb0b5f195474c2516\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://544885302dffff8336a9aa3c4afa80d69088ffaa2402648f4a329e0a798a8c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://544885302dffff8336a9aa3c4afa80d69088ffaa2402648f4a329e0a798a8c99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8bc193fb3f85b980c31af2e826cee90adb633ad2ee82ff1d46e182ab52e8fab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8bc193fb3f85b980c31af2e826cee90adb633ad2ee82ff1d46e182ab52e8fab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7713e2a35bac8df1e8987fbcd452459fa3541caad0c52ed57fda187bd7c83d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7713e2a35bac8df1e8987fbcd452459fa3541caad0c52ed57fda187bd7c83d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34adfca89026fc20007a73a8cbd4e7c7b3137a581c3476867af17fabe976f71f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34adfca89026fc20007a73a8cbd4e7c7b3137a581c3476867af17fabe976f71f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qppnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:42Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:42 crc kubenswrapper[4821]: I1125 10:32:42.414995 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tgdlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2336955b-a703-4db4-8d0f-d08088ed59b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb0362942bfac635ee6292501b1b851772c01d75dd658eef4b17669a716de72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vjfp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tgdlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:42Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:42 crc kubenswrapper[4821]: I1125 10:32:42.426547 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:42 crc kubenswrapper[4821]: I1125 10:32:42.426587 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:42 crc kubenswrapper[4821]: I1125 10:32:42.426598 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:42 crc kubenswrapper[4821]: I1125 10:32:42.426616 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:42 crc kubenswrapper[4821]: I1125 10:32:42.426626 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:42Z","lastTransitionTime":"2025-11-25T10:32:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:42 crc kubenswrapper[4821]: I1125 10:32:42.428230 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62fcb632-f12a-46e2-bb7d-a5311f291ee7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28d43531927e75c3745f100940f7f6bf79fd21026c1309194df8cd7a39361b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69633332d03d2551ff4abb253d358de398eab18486b92db630ce06c0447fb9cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ccb78d9a51ac83ee208c2451685a2512de9deaf742a663bb8c892aa59b76162\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27826e155a9352b1e1866e83fdab683dbcd749cc04b6a4da2db25ecabe3d9a52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://991252e5f56f128827a5054b8e228284bc0266166582cbaedaadc53ca2f5b401\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:42Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:42 crc kubenswrapper[4821]: I1125 10:32:42.440909 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:42Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:42 crc kubenswrapper[4821]: I1125 10:32:42.452050 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f948b87-ac86-4de6-ad64-c2ef947f84d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4482f73a45bc82bd55770c3fd788becacec94893fed1c305fed1c81c753d43aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ba68bd093be94d0d40f3851638db5d26bdfc52d175b1ac78afb91dda7affcfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2krbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:42Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:42 crc kubenswrapper[4821]: I1125 10:32:42.470242 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06f949936bac99cfbf89d720c9549c074351d9077bd963fa024d87004c6be23e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:42Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:42 crc kubenswrapper[4821]: I1125 10:32:42.481985 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384d8ba58264ab89796d3b53afb0c4c1be33cb4fba719a8beef4afc2be7d66a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:42Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:42 crc kubenswrapper[4821]: I1125 10:32:42.495012 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:42Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:42 crc kubenswrapper[4821]: I1125 10:32:42.507307 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17b8714be3bd121e06641873e003f08560b1f71e56d6ea936966df7d4896eb03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85731e6b6dbe1777bc7161a010f39f79523ec4b08d71d5d8a975573aa7fe4957\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:42Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:42 crc kubenswrapper[4821]: I1125 10:32:42.517996 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zlk96" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d55aa91-9cb9-4c95-bc27-6fec05684c10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4521b5baee45ffc9c7661c781c0b3de458dff325ad525dde283d46af71c0cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vkd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zlk96\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:42Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:42 crc kubenswrapper[4821]: I1125 10:32:42.529026 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:42 crc kubenswrapper[4821]: I1125 10:32:42.529147 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:42 crc kubenswrapper[4821]: I1125 10:32:42.529188 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:42 crc kubenswrapper[4821]: I1125 10:32:42.529208 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:42 crc kubenswrapper[4821]: I1125 10:32:42.529219 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:42Z","lastTransitionTime":"2025-11-25T10:32:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:42 crc kubenswrapper[4821]: I1125 10:32:42.631912 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:42 crc kubenswrapper[4821]: I1125 10:32:42.631974 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:42 crc kubenswrapper[4821]: I1125 10:32:42.631988 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:42 crc kubenswrapper[4821]: I1125 10:32:42.632007 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:42 crc kubenswrapper[4821]: I1125 10:32:42.632019 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:42Z","lastTransitionTime":"2025-11-25T10:32:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:42 crc kubenswrapper[4821]: I1125 10:32:42.734685 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:42 crc kubenswrapper[4821]: I1125 10:32:42.734951 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:42 crc kubenswrapper[4821]: I1125 10:32:42.735023 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:42 crc kubenswrapper[4821]: I1125 10:32:42.735095 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:42 crc kubenswrapper[4821]: I1125 10:32:42.735181 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:42Z","lastTransitionTime":"2025-11-25T10:32:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:42 crc kubenswrapper[4821]: I1125 10:32:42.837460 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:42 crc kubenswrapper[4821]: I1125 10:32:42.837823 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:42 crc kubenswrapper[4821]: I1125 10:32:42.837915 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:42 crc kubenswrapper[4821]: I1125 10:32:42.837999 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:42 crc kubenswrapper[4821]: I1125 10:32:42.838078 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:42Z","lastTransitionTime":"2025-11-25T10:32:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:42 crc kubenswrapper[4821]: I1125 10:32:42.940077 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:42 crc kubenswrapper[4821]: I1125 10:32:42.940113 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:42 crc kubenswrapper[4821]: I1125 10:32:42.940123 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:42 crc kubenswrapper[4821]: I1125 10:32:42.940137 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:42 crc kubenswrapper[4821]: I1125 10:32:42.940147 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:42Z","lastTransitionTime":"2025-11-25T10:32:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.026362 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6hbqd"] Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.027033 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6hbqd" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.029441 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.029463 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.042236 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62fcb632-f12a-46e2-bb7d-a5311f291ee7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28d43531927e75c3745f100940f7f6bf79fd21026c1309194df8cd7a39361b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69633332d03d2551ff4abb253d358de398eab18486b92db630ce06c0447fb9cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ccb78d9a51ac83ee208c2451685a2512de9deaf742a663bb8c892aa59b76162\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27826e155a9352b1e1866e83fdab683dbcd749cc04b6a4da2db25ecabe3d9a52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://991252e5f56f128827a5054b8e228284bc0266166582cbaedaadc53ca2f5b401\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:43Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.042598 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.042639 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.042648 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.042664 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.042673 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:43Z","lastTransitionTime":"2025-11-25T10:32:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.055931 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:43Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.066616 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f948b87-ac86-4de6-ad64-c2ef947f84d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4482f73a45bc82bd55770c3fd788becacec94893fed1c305fed1c81c753d43aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ba68bd093be94d0d40f3851638db5d26bdfc52d175b1ac78afb91dda7affcfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2krbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:43Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.080822 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qppnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"357491b8-cd9c-486e-9898-5ba33678513d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc88d700ca3a449ef1cb32d6124b16d0024d6981ced12e276083f86a1b322ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd9c047c5ce1d896f803ff6cafa9bbf3a18a391b44539e1f1312e08c89e51cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd9c047c5ce1d896f803ff6cafa9bbf3a18a391b44539e1f1312e08c89e51cb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://565a6db63fb69abbba81ffdf6085797900858d263cd6affcb0b5f195474c2516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://565a6db63fb69abbba81ffdf6085797900858d263cd6affcb0b5f195474c2516\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://544885302dffff8336a9aa3c4afa80d69088ffaa2402648f4a329e0a798a8c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://544885302dffff8336a9aa3c4afa80d69088ffaa2402648f4a329e0a798a8c99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8bc193fb3f85b980c31af2e826cee90adb633ad2ee82ff1d46e182ab52e8fab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8bc193fb3f85b980c31af2e826cee90adb633ad2ee82ff1d46e182ab52e8fab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7713e2a35bac8df1e8987fbcd452459fa3541caad0c52ed57fda187bd7c83d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7713e2a35bac8df1e8987fbcd452459fa3541caad0c52ed57fda187bd7c83d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34adfca89026fc20007a73a8cbd4e7c7b3137a581c3476867af17fabe976f71f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34adfca89026fc20007a73a8cbd4e7c7b3137a581c3476867af17fabe976f71f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qppnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:43Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.091114 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tgdlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2336955b-a703-4db4-8d0f-d08088ed59b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb0362942bfac635ee6292501b1b851772c01d75dd658eef4b17669a716de72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vjfp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tgdlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:43Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.102372 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06f949936bac99cfbf89d720c9549c074351d9077bd963fa024d87004c6be23e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:43Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.102638 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5f71b3a2-63cb-443c-9f41-39e57030746c-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-6hbqd\" (UID: \"5f71b3a2-63cb-443c-9f41-39e57030746c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6hbqd" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.102712 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5f71b3a2-63cb-443c-9f41-39e57030746c-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-6hbqd\" (UID: \"5f71b3a2-63cb-443c-9f41-39e57030746c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6hbqd" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.102750 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxdmq\" (UniqueName: \"kubernetes.io/projected/5f71b3a2-63cb-443c-9f41-39e57030746c-kube-api-access-zxdmq\") pod \"ovnkube-control-plane-749d76644c-6hbqd\" (UID: \"5f71b3a2-63cb-443c-9f41-39e57030746c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6hbqd" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.102777 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5f71b3a2-63cb-443c-9f41-39e57030746c-env-overrides\") pod \"ovnkube-control-plane-749d76644c-6hbqd\" (UID: \"5f71b3a2-63cb-443c-9f41-39e57030746c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6hbqd" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.114206 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.114229 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.114212 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384d8ba58264ab89796d3b53afb0c4c1be33cb4fba719a8beef4afc2be7d66a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:43Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.114288 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:32:43 crc kubenswrapper[4821]: E1125 10:32:43.114343 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:32:43 crc kubenswrapper[4821]: E1125 10:32:43.114438 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:32:43 crc kubenswrapper[4821]: E1125 10:32:43.114539 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.126707 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:43Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.139530 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17b8714be3bd121e06641873e003f08560b1f71e56d6ea936966df7d4896eb03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85731e6b6dbe1777bc7161a010f39f79523ec4b08d71d5d8a975573aa7fe4957\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:43Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.144821 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.144856 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.144866 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.144880 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.144890 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:43Z","lastTransitionTime":"2025-11-25T10:32:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.149437 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zlk96" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d55aa91-9cb9-4c95-bc27-6fec05684c10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4521b5baee45ffc9c7661c781c0b3de458dff325ad525dde283d46af71c0cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vkd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zlk96\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:43Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.160073 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6hbqd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f71b3a2-63cb-443c-9f41-39e57030746c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxdmq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxdmq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6hbqd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:43Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.177263 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d9e58df-1755-4422-8d04-211f98c7c011\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea762f627aa69fb88a9bd82545fa677e9ccfa97f99a5e92c0aa490e619c0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a06b811e5ec52ce984a2403ba2c98332ba60466d08b3580ad9192fac6caef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0385f6ed9597974267f7c45547f6b6725b4ee83a2cc0508ff2d1c1ce8753ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b915b565f7fbb3e4e856e7f1035a519e82bfb3023660c121f6e3813de353bb92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18465f3a6916c36457a305ba5d568c9077f34f1d000532ed518f237d100008b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:43Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.188061 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25595ccd-762d-414a-9916-4fe38d478b07\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d4b2d8db4d13392d9789e161fd4b8f6eef983b79128ee584d72e4c06565bc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf780f791771ea43a8c8176d02fdfb5204406e023207299369ae46ceb342baaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0251519fd4262f2ea5c64959db49433afdafe36b7ed2404d660eeb7c365d2746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad5c205e0eeecfd43cbdf0eb1a334e3e8faa5e9a9e8dd202c397d5cbb36e081e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:43Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.199496 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:43Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.203969 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxdmq\" (UniqueName: \"kubernetes.io/projected/5f71b3a2-63cb-443c-9f41-39e57030746c-kube-api-access-zxdmq\") pod \"ovnkube-control-plane-749d76644c-6hbqd\" (UID: \"5f71b3a2-63cb-443c-9f41-39e57030746c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6hbqd" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.204007 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5f71b3a2-63cb-443c-9f41-39e57030746c-env-overrides\") pod \"ovnkube-control-plane-749d76644c-6hbqd\" (UID: \"5f71b3a2-63cb-443c-9f41-39e57030746c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6hbqd" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.204041 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5f71b3a2-63cb-443c-9f41-39e57030746c-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-6hbqd\" (UID: \"5f71b3a2-63cb-443c-9f41-39e57030746c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6hbqd" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.204085 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5f71b3a2-63cb-443c-9f41-39e57030746c-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-6hbqd\" (UID: \"5f71b3a2-63cb-443c-9f41-39e57030746c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6hbqd" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.204669 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5f71b3a2-63cb-443c-9f41-39e57030746c-env-overrides\") pod \"ovnkube-control-plane-749d76644c-6hbqd\" (UID: \"5f71b3a2-63cb-443c-9f41-39e57030746c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6hbqd" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.204915 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5f71b3a2-63cb-443c-9f41-39e57030746c-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-6hbqd\" (UID: \"5f71b3a2-63cb-443c-9f41-39e57030746c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6hbqd" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.210445 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5f71b3a2-63cb-443c-9f41-39e57030746c-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-6hbqd\" (UID: \"5f71b3a2-63cb-443c-9f41-39e57030746c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6hbqd" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.212389 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ldzbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b84e32ae-4897-4644-a656-6bd82012cc81\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ef433744761f26e8e1f6c0db123821a11923ffd774830744285cd4178d89533\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5pmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ldzbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:43Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.219126 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxdmq\" (UniqueName: \"kubernetes.io/projected/5f71b3a2-63cb-443c-9f41-39e57030746c-kube-api-access-zxdmq\") pod \"ovnkube-control-plane-749d76644c-6hbqd\" (UID: \"5f71b3a2-63cb-443c-9f41-39e57030746c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6hbqd" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.230688 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"954085fa-d03a-4e88-9dfd-566257e3558d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f00746930c4588e548f0e2b60d52bf448aff7401b676a7855cb0f4cce00d329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a14a078ba7d2aed95992b36502008f315f0d3a634b64ed49da70398b2562d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d7254196b3702e48083ae98e829b73c539adf69ed5b2bbdc62189d09ae1762\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d727430108977574365f0e9c04b44e03307bd7ff89cdbb1e48a6e916f504d6b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94c7949f077b1ca2b310bedb090aa72fc4bd7e8f2c8bf438096edfca8568f36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed2e7cad7bb24b5b56a1295098a6a7c230c8f9b9c29a88ef653ae8b6ac31667d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a84f30b6a8ad3333268312e9aadacb17eca12613511195c6f8a7734d5f7af5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a84f30b6a8ad3333268312e9aadacb17eca12613511195c6f8a7734d5f7af5b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T10:32:41Z\\\",\\\"message\\\":\\\" 6233 obj_retry.go:365] Adding new object: *v1.Pod openshift-dns/node-resolver-zlk96\\\\nI1125 10:32:41.011897 6233 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-target-xd92c in node crc\\\\nI1125 10:32:41.011901 6233 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-additional-cni-plugins-qppnw after 0 failed attempt(s)\\\\nI1125 10:32:41.011815 6233 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1125 10:32:41.011915 6233 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-additional-cni-plugins-qppnw\\\\nI1125 10:32:41.011902 6233 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-zlk96 in node crc\\\\nI1125 10:32:41.011826 6233 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1125 10:32:41.011923 6233 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-tgdlt after 0 failed attempt(s)\\\\nI1125 10:32:41.011884 6233 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-ingress-operator/metrics]} name:Service_openshift-ingress-operator/metrics_\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-mzhnw_openshift-ovn-kubernetes(954085fa-d03a-4e88-9dfd-566257e3558d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dbf499f4ae6161752abaa7a466a027c6f874016be90a534e69f07e3dd3f3d9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzhnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:43Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.247287 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.247324 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.247332 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.247346 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.247355 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:43Z","lastTransitionTime":"2025-11-25T10:32:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.340040 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6hbqd" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.349774 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.349811 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.349819 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.349835 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.349844 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:43Z","lastTransitionTime":"2025-11-25T10:32:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:43 crc kubenswrapper[4821]: W1125 10:32:43.353199 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5f71b3a2_63cb_443c_9f41_39e57030746c.slice/crio-1e0bd935f02042c12a52d4d2a0d752e655c36af4d41b7400f987756806621a36 WatchSource:0}: Error finding container 1e0bd935f02042c12a52d4d2a0d752e655c36af4d41b7400f987756806621a36: Status 404 returned error can't find the container with id 1e0bd935f02042c12a52d4d2a0d752e655c36af4d41b7400f987756806621a36 Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.452411 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.452439 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.452449 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.452465 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.452473 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:43Z","lastTransitionTime":"2025-11-25T10:32:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.555509 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.555553 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.555564 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.555582 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.555591 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:43Z","lastTransitionTime":"2025-11-25T10:32:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.658226 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.658260 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.658270 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.658285 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.658295 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:43Z","lastTransitionTime":"2025-11-25T10:32:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.760543 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.760590 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.760601 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.760620 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.760637 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:43Z","lastTransitionTime":"2025-11-25T10:32:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.863177 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.863216 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.863226 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.863241 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.863253 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:43Z","lastTransitionTime":"2025-11-25T10:32:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.965197 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.965225 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.965233 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.965246 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:43 crc kubenswrapper[4821]: I1125 10:32:43.965254 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:43Z","lastTransitionTime":"2025-11-25T10:32:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.067455 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.067500 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.067508 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.067522 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.067531 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:44Z","lastTransitionTime":"2025-11-25T10:32:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.169646 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.169687 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.169698 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.169716 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.169728 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:44Z","lastTransitionTime":"2025-11-25T10:32:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.272498 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.272564 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.272582 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.272609 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.272626 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:44Z","lastTransitionTime":"2025-11-25T10:32:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.313991 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6hbqd" event={"ID":"5f71b3a2-63cb-443c-9f41-39e57030746c","Type":"ContainerStarted","Data":"595f9ea75461f18e763d84799e368becc07b72acdc5337d20d6995bf53c3d27a"} Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.314047 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6hbqd" event={"ID":"5f71b3a2-63cb-443c-9f41-39e57030746c","Type":"ContainerStarted","Data":"65473b0e63c2e9c7c09a2bb8f31396daf1d87c2c83d20ed3508ce2fe7358dda3"} Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.314061 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6hbqd" event={"ID":"5f71b3a2-63cb-443c-9f41-39e57030746c","Type":"ContainerStarted","Data":"1e0bd935f02042c12a52d4d2a0d752e655c36af4d41b7400f987756806621a36"} Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.327450 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:44Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.341541 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f948b87-ac86-4de6-ad64-c2ef947f84d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4482f73a45bc82bd55770c3fd788becacec94893fed1c305fed1c81c753d43aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ba68bd093be94d0d40f3851638db5d26bdfc52d175b1ac78afb91dda7affcfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2krbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:44Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.355505 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qppnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"357491b8-cd9c-486e-9898-5ba33678513d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc88d700ca3a449ef1cb32d6124b16d0024d6981ced12e276083f86a1b322ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd9c047c5ce1d896f803ff6cafa9bbf3a18a391b44539e1f1312e08c89e51cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd9c047c5ce1d896f803ff6cafa9bbf3a18a391b44539e1f1312e08c89e51cb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://565a6db63fb69abbba81ffdf6085797900858d263cd6affcb0b5f195474c2516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://565a6db63fb69abbba81ffdf6085797900858d263cd6affcb0b5f195474c2516\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://544885302dffff8336a9aa3c4afa80d69088ffaa2402648f4a329e0a798a8c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://544885302dffff8336a9aa3c4afa80d69088ffaa2402648f4a329e0a798a8c99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8bc193fb3f85b980c31af2e826cee90adb633ad2ee82ff1d46e182ab52e8fab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8bc193fb3f85b980c31af2e826cee90adb633ad2ee82ff1d46e182ab52e8fab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7713e2a35bac8df1e8987fbcd452459fa3541caad0c52ed57fda187bd7c83d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7713e2a35bac8df1e8987fbcd452459fa3541caad0c52ed57fda187bd7c83d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34adfca89026fc20007a73a8cbd4e7c7b3137a581c3476867af17fabe976f71f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34adfca89026fc20007a73a8cbd4e7c7b3137a581c3476867af17fabe976f71f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qppnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:44Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.366734 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tgdlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2336955b-a703-4db4-8d0f-d08088ed59b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb0362942bfac635ee6292501b1b851772c01d75dd658eef4b17669a716de72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vjfp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tgdlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:44Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.375733 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.375811 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.375835 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.375861 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.375891 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:44Z","lastTransitionTime":"2025-11-25T10:32:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.383690 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62fcb632-f12a-46e2-bb7d-a5311f291ee7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28d43531927e75c3745f100940f7f6bf79fd21026c1309194df8cd7a39361b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69633332d03d2551ff4abb253d358de398eab18486b92db630ce06c0447fb9cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ccb78d9a51ac83ee208c2451685a2512de9deaf742a663bb8c892aa59b76162\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27826e155a9352b1e1866e83fdab683dbcd749cc04b6a4da2db25ecabe3d9a52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://991252e5f56f128827a5054b8e228284bc0266166582cbaedaadc53ca2f5b401\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:44Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.405154 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06f949936bac99cfbf89d720c9549c074351d9077bd963fa024d87004c6be23e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:44Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.418009 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384d8ba58264ab89796d3b53afb0c4c1be33cb4fba719a8beef4afc2be7d66a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:44Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.427616 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zlk96" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d55aa91-9cb9-4c95-bc27-6fec05684c10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4521b5baee45ffc9c7661c781c0b3de458dff325ad525dde283d46af71c0cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vkd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zlk96\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:44Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.437577 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6hbqd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f71b3a2-63cb-443c-9f41-39e57030746c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65473b0e63c2e9c7c09a2bb8f31396daf1d87c2c83d20ed3508ce2fe7358dda3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxdmq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://595f9ea75461f18e763d84799e368becc07b72acdc5337d20d6995bf53c3d27a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxdmq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6hbqd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:44Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.447016 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:44Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.457805 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17b8714be3bd121e06641873e003f08560b1f71e56d6ea936966df7d4896eb03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85731e6b6dbe1777bc7161a010f39f79523ec4b08d71d5d8a975573aa7fe4957\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:44Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.472910 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25595ccd-762d-414a-9916-4fe38d478b07\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d4b2d8db4d13392d9789e161fd4b8f6eef983b79128ee584d72e4c06565bc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf780f791771ea43a8c8176d02fdfb5204406e023207299369ae46ceb342baaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0251519fd4262f2ea5c64959db49433afdafe36b7ed2404d660eeb7c365d2746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad5c205e0eeecfd43cbdf0eb1a334e3e8faa5e9a9e8dd202c397d5cbb36e081e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:44Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.475682 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-d4msn"] Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.476108 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4msn" Nov 25 10:32:44 crc kubenswrapper[4821]: E1125 10:32:44.476178 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4msn" podUID="1f12cef3-7e40-4f61-836e-23cd0e578e71" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.477513 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.477551 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.477560 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.477575 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.477586 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:44Z","lastTransitionTime":"2025-11-25T10:32:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.486445 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:44Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.499583 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ldzbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b84e32ae-4897-4644-a656-6bd82012cc81\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ef433744761f26e8e1f6c0db123821a11923ffd774830744285cd4178d89533\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5pmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ldzbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:44Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.515482 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"954085fa-d03a-4e88-9dfd-566257e3558d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f00746930c4588e548f0e2b60d52bf448aff7401b676a7855cb0f4cce00d329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a14a078ba7d2aed95992b36502008f315f0d3a634b64ed49da70398b2562d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d7254196b3702e48083ae98e829b73c539adf69ed5b2bbdc62189d09ae1762\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d727430108977574365f0e9c04b44e03307bd7ff89cdbb1e48a6e916f504d6b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94c7949f077b1ca2b310bedb090aa72fc4bd7e8f2c8bf438096edfca8568f36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed2e7cad7bb24b5b56a1295098a6a7c230c8f9b9c29a88ef653ae8b6ac31667d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a84f30b6a8ad3333268312e9aadacb17eca12613511195c6f8a7734d5f7af5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a84f30b6a8ad3333268312e9aadacb17eca12613511195c6f8a7734d5f7af5b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T10:32:41Z\\\",\\\"message\\\":\\\" 6233 obj_retry.go:365] Adding new object: *v1.Pod openshift-dns/node-resolver-zlk96\\\\nI1125 10:32:41.011897 6233 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-target-xd92c in node crc\\\\nI1125 10:32:41.011901 6233 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-additional-cni-plugins-qppnw after 0 failed attempt(s)\\\\nI1125 10:32:41.011815 6233 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1125 10:32:41.011915 6233 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-additional-cni-plugins-qppnw\\\\nI1125 10:32:41.011902 6233 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-zlk96 in node crc\\\\nI1125 10:32:41.011826 6233 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1125 10:32:41.011923 6233 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-tgdlt after 0 failed attempt(s)\\\\nI1125 10:32:41.011884 6233 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-ingress-operator/metrics]} name:Service_openshift-ingress-operator/metrics_\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-mzhnw_openshift-ovn-kubernetes(954085fa-d03a-4e88-9dfd-566257e3558d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dbf499f4ae6161752abaa7a466a027c6f874016be90a534e69f07e3dd3f3d9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzhnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:44Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.517600 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s7lkv\" (UniqueName: \"kubernetes.io/projected/1f12cef3-7e40-4f61-836e-23cd0e578e71-kube-api-access-s7lkv\") pod \"network-metrics-daemon-d4msn\" (UID: \"1f12cef3-7e40-4f61-836e-23cd0e578e71\") " pod="openshift-multus/network-metrics-daemon-d4msn" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.517632 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1f12cef3-7e40-4f61-836e-23cd0e578e71-metrics-certs\") pod \"network-metrics-daemon-d4msn\" (UID: \"1f12cef3-7e40-4f61-836e-23cd0e578e71\") " pod="openshift-multus/network-metrics-daemon-d4msn" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.531860 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d9e58df-1755-4422-8d04-211f98c7c011\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea762f627aa69fb88a9bd82545fa677e9ccfa97f99a5e92c0aa490e619c0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a06b811e5ec52ce984a2403ba2c98332ba60466d08b3580ad9192fac6caef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0385f6ed9597974267f7c45547f6b6725b4ee83a2cc0508ff2d1c1ce8753ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b915b565f7fbb3e4e856e7f1035a519e82bfb3023660c121f6e3813de353bb92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18465f3a6916c36457a305ba5d568c9077f34f1d000532ed518f237d100008b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:44Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.543083 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62fcb632-f12a-46e2-bb7d-a5311f291ee7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28d43531927e75c3745f100940f7f6bf79fd21026c1309194df8cd7a39361b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69633332d03d2551ff4abb253d358de398eab18486b92db630ce06c0447fb9cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ccb78d9a51ac83ee208c2451685a2512de9deaf742a663bb8c892aa59b76162\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27826e155a9352b1e1866e83fdab683dbcd749cc04b6a4da2db25ecabe3d9a52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://991252e5f56f128827a5054b8e228284bc0266166582cbaedaadc53ca2f5b401\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:44Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.552730 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:44Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.562632 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f948b87-ac86-4de6-ad64-c2ef947f84d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4482f73a45bc82bd55770c3fd788becacec94893fed1c305fed1c81c753d43aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ba68bd093be94d0d40f3851638db5d26bdfc52d175b1ac78afb91dda7affcfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2krbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:44Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.573969 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qppnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"357491b8-cd9c-486e-9898-5ba33678513d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc88d700ca3a449ef1cb32d6124b16d0024d6981ced12e276083f86a1b322ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd9c047c5ce1d896f803ff6cafa9bbf3a18a391b44539e1f1312e08c89e51cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd9c047c5ce1d896f803ff6cafa9bbf3a18a391b44539e1f1312e08c89e51cb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://565a6db63fb69abbba81ffdf6085797900858d263cd6affcb0b5f195474c2516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://565a6db63fb69abbba81ffdf6085797900858d263cd6affcb0b5f195474c2516\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://544885302dffff8336a9aa3c4afa80d69088ffaa2402648f4a329e0a798a8c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://544885302dffff8336a9aa3c4afa80d69088ffaa2402648f4a329e0a798a8c99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8bc193fb3f85b980c31af2e826cee90adb633ad2ee82ff1d46e182ab52e8fab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8bc193fb3f85b980c31af2e826cee90adb633ad2ee82ff1d46e182ab52e8fab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7713e2a35bac8df1e8987fbcd452459fa3541caad0c52ed57fda187bd7c83d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7713e2a35bac8df1e8987fbcd452459fa3541caad0c52ed57fda187bd7c83d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34adfca89026fc20007a73a8cbd4e7c7b3137a581c3476867af17fabe976f71f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34adfca89026fc20007a73a8cbd4e7c7b3137a581c3476867af17fabe976f71f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qppnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:44Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.579495 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.579541 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.579549 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.579562 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.579571 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:44Z","lastTransitionTime":"2025-11-25T10:32:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.583632 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tgdlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2336955b-a703-4db4-8d0f-d08088ed59b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb0362942bfac635ee6292501b1b851772c01d75dd658eef4b17669a716de72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vjfp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tgdlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:44Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.594607 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06f949936bac99cfbf89d720c9549c074351d9077bd963fa024d87004c6be23e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:44Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.604337 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384d8ba58264ab89796d3b53afb0c4c1be33cb4fba719a8beef4afc2be7d66a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:44Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.614142 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:44Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.618739 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s7lkv\" (UniqueName: \"kubernetes.io/projected/1f12cef3-7e40-4f61-836e-23cd0e578e71-kube-api-access-s7lkv\") pod \"network-metrics-daemon-d4msn\" (UID: \"1f12cef3-7e40-4f61-836e-23cd0e578e71\") " pod="openshift-multus/network-metrics-daemon-d4msn" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.618784 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1f12cef3-7e40-4f61-836e-23cd0e578e71-metrics-certs\") pod \"network-metrics-daemon-d4msn\" (UID: \"1f12cef3-7e40-4f61-836e-23cd0e578e71\") " pod="openshift-multus/network-metrics-daemon-d4msn" Nov 25 10:32:44 crc kubenswrapper[4821]: E1125 10:32:44.618899 4821 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 10:32:44 crc kubenswrapper[4821]: E1125 10:32:44.618956 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1f12cef3-7e40-4f61-836e-23cd0e578e71-metrics-certs podName:1f12cef3-7e40-4f61-836e-23cd0e578e71 nodeName:}" failed. No retries permitted until 2025-11-25 10:32:45.118939459 +0000 UTC m=+35.655259306 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/1f12cef3-7e40-4f61-836e-23cd0e578e71-metrics-certs") pod "network-metrics-daemon-d4msn" (UID: "1f12cef3-7e40-4f61-836e-23cd0e578e71") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.627007 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17b8714be3bd121e06641873e003f08560b1f71e56d6ea936966df7d4896eb03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85731e6b6dbe1777bc7161a010f39f79523ec4b08d71d5d8a975573aa7fe4957\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:44Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.634752 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s7lkv\" (UniqueName: \"kubernetes.io/projected/1f12cef3-7e40-4f61-836e-23cd0e578e71-kube-api-access-s7lkv\") pod \"network-metrics-daemon-d4msn\" (UID: \"1f12cef3-7e40-4f61-836e-23cd0e578e71\") " pod="openshift-multus/network-metrics-daemon-d4msn" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.638554 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zlk96" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d55aa91-9cb9-4c95-bc27-6fec05684c10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4521b5baee45ffc9c7661c781c0b3de458dff325ad525dde283d46af71c0cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vkd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zlk96\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:44Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.653620 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6hbqd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f71b3a2-63cb-443c-9f41-39e57030746c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65473b0e63c2e9c7c09a2bb8f31396daf1d87c2c83d20ed3508ce2fe7358dda3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxdmq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://595f9ea75461f18e763d84799e368becc07b72acdc5337d20d6995bf53c3d27a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxdmq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6hbqd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:44Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.663978 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-d4msn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f12cef3-7e40-4f61-836e-23cd0e578e71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s7lkv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s7lkv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:44Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-d4msn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:44Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.682063 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.682112 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.682121 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.682134 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.682143 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:44Z","lastTransitionTime":"2025-11-25T10:32:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.685036 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d9e58df-1755-4422-8d04-211f98c7c011\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea762f627aa69fb88a9bd82545fa677e9ccfa97f99a5e92c0aa490e619c0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a06b811e5ec52ce984a2403ba2c98332ba60466d08b3580ad9192fac6caef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0385f6ed9597974267f7c45547f6b6725b4ee83a2cc0508ff2d1c1ce8753ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b915b565f7fbb3e4e856e7f1035a519e82bfb3023660c121f6e3813de353bb92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18465f3a6916c36457a305ba5d568c9077f34f1d000532ed518f237d100008b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:44Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.698993 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25595ccd-762d-414a-9916-4fe38d478b07\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d4b2d8db4d13392d9789e161fd4b8f6eef983b79128ee584d72e4c06565bc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf780f791771ea43a8c8176d02fdfb5204406e023207299369ae46ceb342baaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0251519fd4262f2ea5c64959db49433afdafe36b7ed2404d660eeb7c365d2746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad5c205e0eeecfd43cbdf0eb1a334e3e8faa5e9a9e8dd202c397d5cbb36e081e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:44Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.711310 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:44Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.724514 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ldzbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b84e32ae-4897-4644-a656-6bd82012cc81\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ef433744761f26e8e1f6c0db123821a11923ffd774830744285cd4178d89533\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5pmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ldzbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:44Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.741263 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"954085fa-d03a-4e88-9dfd-566257e3558d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f00746930c4588e548f0e2b60d52bf448aff7401b676a7855cb0f4cce00d329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a14a078ba7d2aed95992b36502008f315f0d3a634b64ed49da70398b2562d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d7254196b3702e48083ae98e829b73c539adf69ed5b2bbdc62189d09ae1762\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d727430108977574365f0e9c04b44e03307bd7ff89cdbb1e48a6e916f504d6b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94c7949f077b1ca2b310bedb090aa72fc4bd7e8f2c8bf438096edfca8568f36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed2e7cad7bb24b5b56a1295098a6a7c230c8f9b9c29a88ef653ae8b6ac31667d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a84f30b6a8ad3333268312e9aadacb17eca12613511195c6f8a7734d5f7af5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a84f30b6a8ad3333268312e9aadacb17eca12613511195c6f8a7734d5f7af5b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T10:32:41Z\\\",\\\"message\\\":\\\" 6233 obj_retry.go:365] Adding new object: *v1.Pod openshift-dns/node-resolver-zlk96\\\\nI1125 10:32:41.011897 6233 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-target-xd92c in node crc\\\\nI1125 10:32:41.011901 6233 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-additional-cni-plugins-qppnw after 0 failed attempt(s)\\\\nI1125 10:32:41.011815 6233 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1125 10:32:41.011915 6233 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-additional-cni-plugins-qppnw\\\\nI1125 10:32:41.011902 6233 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-zlk96 in node crc\\\\nI1125 10:32:41.011826 6233 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1125 10:32:41.011923 6233 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-tgdlt after 0 failed attempt(s)\\\\nI1125 10:32:41.011884 6233 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-ingress-operator/metrics]} name:Service_openshift-ingress-operator/metrics_\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-mzhnw_openshift-ovn-kubernetes(954085fa-d03a-4e88-9dfd-566257e3558d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dbf499f4ae6161752abaa7a466a027c6f874016be90a534e69f07e3dd3f3d9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzhnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:44Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.785096 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.785139 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.785152 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.785189 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.785201 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:44Z","lastTransitionTime":"2025-11-25T10:32:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.887477 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.887521 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.887530 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.887546 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.887556 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:44Z","lastTransitionTime":"2025-11-25T10:32:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.921903 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:32:44 crc kubenswrapper[4821]: E1125 10:32:44.922008 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:33:00.92198582 +0000 UTC m=+51.458305677 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.922037 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.922073 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:32:44 crc kubenswrapper[4821]: E1125 10:32:44.922203 4821 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 10:32:44 crc kubenswrapper[4821]: E1125 10:32:44.922226 4821 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 10:32:44 crc kubenswrapper[4821]: E1125 10:32:44.922252 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 10:33:00.922242595 +0000 UTC m=+51.458562442 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 10:32:44 crc kubenswrapper[4821]: E1125 10:32:44.922268 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 10:33:00.922260616 +0000 UTC m=+51.458580463 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.989943 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.989973 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.989981 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.989993 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:44 crc kubenswrapper[4821]: I1125 10:32:44.990001 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:44Z","lastTransitionTime":"2025-11-25T10:32:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.022613 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.022647 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:32:45 crc kubenswrapper[4821]: E1125 10:32:45.022713 4821 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 10:32:45 crc kubenswrapper[4821]: E1125 10:32:45.022730 4821 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 10:32:45 crc kubenswrapper[4821]: E1125 10:32:45.022737 4821 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 10:32:45 crc kubenswrapper[4821]: E1125 10:32:45.022741 4821 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 10:32:45 crc kubenswrapper[4821]: E1125 10:32:45.022748 4821 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 10:32:45 crc kubenswrapper[4821]: E1125 10:32:45.022756 4821 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 10:32:45 crc kubenswrapper[4821]: E1125 10:32:45.022785 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 10:33:01.022771572 +0000 UTC m=+51.559091419 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 10:32:45 crc kubenswrapper[4821]: E1125 10:32:45.022799 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 10:33:01.022792822 +0000 UTC m=+51.559112669 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.092505 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.092544 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.092552 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.092572 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.092582 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:45Z","lastTransitionTime":"2025-11-25T10:32:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.114084 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.114156 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.114194 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:32:45 crc kubenswrapper[4821]: E1125 10:32:45.114214 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:32:45 crc kubenswrapper[4821]: E1125 10:32:45.114330 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:32:45 crc kubenswrapper[4821]: E1125 10:32:45.114405 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.123652 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1f12cef3-7e40-4f61-836e-23cd0e578e71-metrics-certs\") pod \"network-metrics-daemon-d4msn\" (UID: \"1f12cef3-7e40-4f61-836e-23cd0e578e71\") " pod="openshift-multus/network-metrics-daemon-d4msn" Nov 25 10:32:45 crc kubenswrapper[4821]: E1125 10:32:45.123784 4821 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 10:32:45 crc kubenswrapper[4821]: E1125 10:32:45.123859 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1f12cef3-7e40-4f61-836e-23cd0e578e71-metrics-certs podName:1f12cef3-7e40-4f61-836e-23cd0e578e71 nodeName:}" failed. No retries permitted until 2025-11-25 10:32:46.12384029 +0000 UTC m=+36.660160137 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/1f12cef3-7e40-4f61-836e-23cd0e578e71-metrics-certs") pod "network-metrics-daemon-d4msn" (UID: "1f12cef3-7e40-4f61-836e-23cd0e578e71") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.194425 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.194463 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.194477 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.194492 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.194501 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:45Z","lastTransitionTime":"2025-11-25T10:32:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.296602 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.296635 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.296643 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.296659 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.296669 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:45Z","lastTransitionTime":"2025-11-25T10:32:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.398837 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.398894 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.398903 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.398920 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.398933 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:45Z","lastTransitionTime":"2025-11-25T10:32:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.501010 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.501063 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.501075 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.501091 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.501103 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:45Z","lastTransitionTime":"2025-11-25T10:32:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.603190 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.603274 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.603283 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.603299 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.603309 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:45Z","lastTransitionTime":"2025-11-25T10:32:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.704978 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.705020 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.705036 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.705051 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.705063 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:45Z","lastTransitionTime":"2025-11-25T10:32:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.808000 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.808058 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.808071 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.808091 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.808103 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:45Z","lastTransitionTime":"2025-11-25T10:32:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.826549 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.826590 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.826602 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.826618 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.826630 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:45Z","lastTransitionTime":"2025-11-25T10:32:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:45 crc kubenswrapper[4821]: E1125 10:32:45.838590 4821 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:32:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:32:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:32:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:32:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a5b2b070-39b9-412c-90f1-c7ed2e9e0950\\\",\\\"systemUUID\\\":\\\"ea86e80f-41b7-4190-ad7e-d8c549d535de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:45Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.841592 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.841624 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.841632 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.841645 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.841656 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:45Z","lastTransitionTime":"2025-11-25T10:32:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:45 crc kubenswrapper[4821]: E1125 10:32:45.852552 4821 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:32:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:32:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:32:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:32:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a5b2b070-39b9-412c-90f1-c7ed2e9e0950\\\",\\\"systemUUID\\\":\\\"ea86e80f-41b7-4190-ad7e-d8c549d535de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:45Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.855283 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.855312 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.855321 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.855334 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.855342 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:45Z","lastTransitionTime":"2025-11-25T10:32:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:45 crc kubenswrapper[4821]: E1125 10:32:45.866070 4821 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:32:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:32:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:32:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:32:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a5b2b070-39b9-412c-90f1-c7ed2e9e0950\\\",\\\"systemUUID\\\":\\\"ea86e80f-41b7-4190-ad7e-d8c549d535de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:45Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.869249 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.869282 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.869293 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.869309 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.869321 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:45Z","lastTransitionTime":"2025-11-25T10:32:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:45 crc kubenswrapper[4821]: E1125 10:32:45.882072 4821 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:32:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:32:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:32:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:32:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a5b2b070-39b9-412c-90f1-c7ed2e9e0950\\\",\\\"systemUUID\\\":\\\"ea86e80f-41b7-4190-ad7e-d8c549d535de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:45Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.885864 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.885900 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.885911 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.885933 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.885944 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:45Z","lastTransitionTime":"2025-11-25T10:32:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:45 crc kubenswrapper[4821]: E1125 10:32:45.900817 4821 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:32:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:32:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:32:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:32:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a5b2b070-39b9-412c-90f1-c7ed2e9e0950\\\",\\\"systemUUID\\\":\\\"ea86e80f-41b7-4190-ad7e-d8c549d535de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:45Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:45 crc kubenswrapper[4821]: E1125 10:32:45.900944 4821 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.910829 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.910881 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.910890 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.910904 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:45 crc kubenswrapper[4821]: I1125 10:32:45.910913 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:45Z","lastTransitionTime":"2025-11-25T10:32:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:46 crc kubenswrapper[4821]: I1125 10:32:46.017799 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:46 crc kubenswrapper[4821]: I1125 10:32:46.017996 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:46 crc kubenswrapper[4821]: I1125 10:32:46.018027 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:46 crc kubenswrapper[4821]: I1125 10:32:46.018109 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:46 crc kubenswrapper[4821]: I1125 10:32:46.018134 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:46Z","lastTransitionTime":"2025-11-25T10:32:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:46 crc kubenswrapper[4821]: I1125 10:32:46.113695 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4msn" Nov 25 10:32:46 crc kubenswrapper[4821]: E1125 10:32:46.113908 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4msn" podUID="1f12cef3-7e40-4f61-836e-23cd0e578e71" Nov 25 10:32:46 crc kubenswrapper[4821]: I1125 10:32:46.120767 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:46 crc kubenswrapper[4821]: I1125 10:32:46.120864 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:46 crc kubenswrapper[4821]: I1125 10:32:46.120893 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:46 crc kubenswrapper[4821]: I1125 10:32:46.120937 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:46 crc kubenswrapper[4821]: I1125 10:32:46.120969 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:46Z","lastTransitionTime":"2025-11-25T10:32:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:46 crc kubenswrapper[4821]: I1125 10:32:46.133771 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1f12cef3-7e40-4f61-836e-23cd0e578e71-metrics-certs\") pod \"network-metrics-daemon-d4msn\" (UID: \"1f12cef3-7e40-4f61-836e-23cd0e578e71\") " pod="openshift-multus/network-metrics-daemon-d4msn" Nov 25 10:32:46 crc kubenswrapper[4821]: E1125 10:32:46.134150 4821 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 10:32:46 crc kubenswrapper[4821]: E1125 10:32:46.134251 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1f12cef3-7e40-4f61-836e-23cd0e578e71-metrics-certs podName:1f12cef3-7e40-4f61-836e-23cd0e578e71 nodeName:}" failed. No retries permitted until 2025-11-25 10:32:48.134229207 +0000 UTC m=+38.670549074 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/1f12cef3-7e40-4f61-836e-23cd0e578e71-metrics-certs") pod "network-metrics-daemon-d4msn" (UID: "1f12cef3-7e40-4f61-836e-23cd0e578e71") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 10:32:46 crc kubenswrapper[4821]: I1125 10:32:46.223914 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:46 crc kubenswrapper[4821]: I1125 10:32:46.223944 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:46 crc kubenswrapper[4821]: I1125 10:32:46.223952 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:46 crc kubenswrapper[4821]: I1125 10:32:46.223964 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:46 crc kubenswrapper[4821]: I1125 10:32:46.223974 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:46Z","lastTransitionTime":"2025-11-25T10:32:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:46 crc kubenswrapper[4821]: I1125 10:32:46.326654 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:46 crc kubenswrapper[4821]: I1125 10:32:46.326696 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:46 crc kubenswrapper[4821]: I1125 10:32:46.326708 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:46 crc kubenswrapper[4821]: I1125 10:32:46.326724 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:46 crc kubenswrapper[4821]: I1125 10:32:46.326734 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:46Z","lastTransitionTime":"2025-11-25T10:32:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:46 crc kubenswrapper[4821]: I1125 10:32:46.428978 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:46 crc kubenswrapper[4821]: I1125 10:32:46.429017 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:46 crc kubenswrapper[4821]: I1125 10:32:46.429025 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:46 crc kubenswrapper[4821]: I1125 10:32:46.429039 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:46 crc kubenswrapper[4821]: I1125 10:32:46.429047 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:46Z","lastTransitionTime":"2025-11-25T10:32:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:46 crc kubenswrapper[4821]: I1125 10:32:46.531054 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:46 crc kubenswrapper[4821]: I1125 10:32:46.531092 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:46 crc kubenswrapper[4821]: I1125 10:32:46.531100 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:46 crc kubenswrapper[4821]: I1125 10:32:46.531114 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:46 crc kubenswrapper[4821]: I1125 10:32:46.531123 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:46Z","lastTransitionTime":"2025-11-25T10:32:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:46 crc kubenswrapper[4821]: I1125 10:32:46.633831 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:46 crc kubenswrapper[4821]: I1125 10:32:46.633878 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:46 crc kubenswrapper[4821]: I1125 10:32:46.633891 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:46 crc kubenswrapper[4821]: I1125 10:32:46.633906 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:46 crc kubenswrapper[4821]: I1125 10:32:46.633917 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:46Z","lastTransitionTime":"2025-11-25T10:32:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:46 crc kubenswrapper[4821]: I1125 10:32:46.736740 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:46 crc kubenswrapper[4821]: I1125 10:32:46.736785 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:46 crc kubenswrapper[4821]: I1125 10:32:46.736797 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:46 crc kubenswrapper[4821]: I1125 10:32:46.736814 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:46 crc kubenswrapper[4821]: I1125 10:32:46.736827 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:46Z","lastTransitionTime":"2025-11-25T10:32:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:46 crc kubenswrapper[4821]: I1125 10:32:46.839482 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:46 crc kubenswrapper[4821]: I1125 10:32:46.839533 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:46 crc kubenswrapper[4821]: I1125 10:32:46.839545 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:46 crc kubenswrapper[4821]: I1125 10:32:46.839561 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:46 crc kubenswrapper[4821]: I1125 10:32:46.839570 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:46Z","lastTransitionTime":"2025-11-25T10:32:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:46 crc kubenswrapper[4821]: I1125 10:32:46.941280 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:46 crc kubenswrapper[4821]: I1125 10:32:46.941334 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:46 crc kubenswrapper[4821]: I1125 10:32:46.941346 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:46 crc kubenswrapper[4821]: I1125 10:32:46.941367 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:46 crc kubenswrapper[4821]: I1125 10:32:46.941379 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:46Z","lastTransitionTime":"2025-11-25T10:32:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:47 crc kubenswrapper[4821]: I1125 10:32:47.043651 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:47 crc kubenswrapper[4821]: I1125 10:32:47.043682 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:47 crc kubenswrapper[4821]: I1125 10:32:47.043691 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:47 crc kubenswrapper[4821]: I1125 10:32:47.043704 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:47 crc kubenswrapper[4821]: I1125 10:32:47.043712 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:47Z","lastTransitionTime":"2025-11-25T10:32:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:47 crc kubenswrapper[4821]: I1125 10:32:47.113988 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:32:47 crc kubenswrapper[4821]: I1125 10:32:47.114032 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:32:47 crc kubenswrapper[4821]: I1125 10:32:47.114074 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:32:47 crc kubenswrapper[4821]: E1125 10:32:47.114125 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:32:47 crc kubenswrapper[4821]: E1125 10:32:47.114248 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:32:47 crc kubenswrapper[4821]: E1125 10:32:47.114348 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:32:47 crc kubenswrapper[4821]: I1125 10:32:47.146588 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:47 crc kubenswrapper[4821]: I1125 10:32:47.146877 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:47 crc kubenswrapper[4821]: I1125 10:32:47.146908 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:47 crc kubenswrapper[4821]: I1125 10:32:47.146926 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:47 crc kubenswrapper[4821]: I1125 10:32:47.146936 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:47Z","lastTransitionTime":"2025-11-25T10:32:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:47 crc kubenswrapper[4821]: I1125 10:32:47.249609 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:47 crc kubenswrapper[4821]: I1125 10:32:47.249654 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:47 crc kubenswrapper[4821]: I1125 10:32:47.249666 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:47 crc kubenswrapper[4821]: I1125 10:32:47.249683 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:47 crc kubenswrapper[4821]: I1125 10:32:47.249714 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:47Z","lastTransitionTime":"2025-11-25T10:32:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:47 crc kubenswrapper[4821]: I1125 10:32:47.352467 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:47 crc kubenswrapper[4821]: I1125 10:32:47.352509 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:47 crc kubenswrapper[4821]: I1125 10:32:47.352521 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:47 crc kubenswrapper[4821]: I1125 10:32:47.352538 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:47 crc kubenswrapper[4821]: I1125 10:32:47.352549 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:47Z","lastTransitionTime":"2025-11-25T10:32:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:47 crc kubenswrapper[4821]: I1125 10:32:47.455354 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:47 crc kubenswrapper[4821]: I1125 10:32:47.455398 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:47 crc kubenswrapper[4821]: I1125 10:32:47.455414 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:47 crc kubenswrapper[4821]: I1125 10:32:47.455431 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:47 crc kubenswrapper[4821]: I1125 10:32:47.455442 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:47Z","lastTransitionTime":"2025-11-25T10:32:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:47 crc kubenswrapper[4821]: I1125 10:32:47.557650 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:47 crc kubenswrapper[4821]: I1125 10:32:47.557699 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:47 crc kubenswrapper[4821]: I1125 10:32:47.557710 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:47 crc kubenswrapper[4821]: I1125 10:32:47.557723 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:47 crc kubenswrapper[4821]: I1125 10:32:47.557733 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:47Z","lastTransitionTime":"2025-11-25T10:32:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:47 crc kubenswrapper[4821]: I1125 10:32:47.660404 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:47 crc kubenswrapper[4821]: I1125 10:32:47.660454 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:47 crc kubenswrapper[4821]: I1125 10:32:47.660463 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:47 crc kubenswrapper[4821]: I1125 10:32:47.660480 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:47 crc kubenswrapper[4821]: I1125 10:32:47.660488 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:47Z","lastTransitionTime":"2025-11-25T10:32:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:47 crc kubenswrapper[4821]: I1125 10:32:47.762925 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:47 crc kubenswrapper[4821]: I1125 10:32:47.763343 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:47 crc kubenswrapper[4821]: I1125 10:32:47.763452 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:47 crc kubenswrapper[4821]: I1125 10:32:47.763545 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:47 crc kubenswrapper[4821]: I1125 10:32:47.763618 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:47Z","lastTransitionTime":"2025-11-25T10:32:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:47 crc kubenswrapper[4821]: I1125 10:32:47.866631 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:47 crc kubenswrapper[4821]: I1125 10:32:47.866692 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:47 crc kubenswrapper[4821]: I1125 10:32:47.866704 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:47 crc kubenswrapper[4821]: I1125 10:32:47.866729 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:47 crc kubenswrapper[4821]: I1125 10:32:47.866745 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:47Z","lastTransitionTime":"2025-11-25T10:32:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:47 crc kubenswrapper[4821]: I1125 10:32:47.969766 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:47 crc kubenswrapper[4821]: I1125 10:32:47.970076 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:47 crc kubenswrapper[4821]: I1125 10:32:47.970179 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:47 crc kubenswrapper[4821]: I1125 10:32:47.970287 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:47 crc kubenswrapper[4821]: I1125 10:32:47.970357 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:47Z","lastTransitionTime":"2025-11-25T10:32:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:48 crc kubenswrapper[4821]: I1125 10:32:48.074131 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:48 crc kubenswrapper[4821]: I1125 10:32:48.074257 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:48 crc kubenswrapper[4821]: I1125 10:32:48.074275 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:48 crc kubenswrapper[4821]: I1125 10:32:48.074302 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:48 crc kubenswrapper[4821]: I1125 10:32:48.074318 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:48Z","lastTransitionTime":"2025-11-25T10:32:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:48 crc kubenswrapper[4821]: I1125 10:32:48.114015 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4msn" Nov 25 10:32:48 crc kubenswrapper[4821]: E1125 10:32:48.114337 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4msn" podUID="1f12cef3-7e40-4f61-836e-23cd0e578e71" Nov 25 10:32:48 crc kubenswrapper[4821]: I1125 10:32:48.153210 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1f12cef3-7e40-4f61-836e-23cd0e578e71-metrics-certs\") pod \"network-metrics-daemon-d4msn\" (UID: \"1f12cef3-7e40-4f61-836e-23cd0e578e71\") " pod="openshift-multus/network-metrics-daemon-d4msn" Nov 25 10:32:48 crc kubenswrapper[4821]: E1125 10:32:48.154129 4821 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 10:32:48 crc kubenswrapper[4821]: E1125 10:32:48.154364 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1f12cef3-7e40-4f61-836e-23cd0e578e71-metrics-certs podName:1f12cef3-7e40-4f61-836e-23cd0e578e71 nodeName:}" failed. No retries permitted until 2025-11-25 10:32:52.154339615 +0000 UTC m=+42.690659462 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/1f12cef3-7e40-4f61-836e-23cd0e578e71-metrics-certs") pod "network-metrics-daemon-d4msn" (UID: "1f12cef3-7e40-4f61-836e-23cd0e578e71") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 10:32:48 crc kubenswrapper[4821]: I1125 10:32:48.177103 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:48 crc kubenswrapper[4821]: I1125 10:32:48.177152 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:48 crc kubenswrapper[4821]: I1125 10:32:48.177179 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:48 crc kubenswrapper[4821]: I1125 10:32:48.177198 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:48 crc kubenswrapper[4821]: I1125 10:32:48.177210 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:48Z","lastTransitionTime":"2025-11-25T10:32:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:48 crc kubenswrapper[4821]: I1125 10:32:48.279434 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:48 crc kubenswrapper[4821]: I1125 10:32:48.279477 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:48 crc kubenswrapper[4821]: I1125 10:32:48.279488 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:48 crc kubenswrapper[4821]: I1125 10:32:48.279505 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:48 crc kubenswrapper[4821]: I1125 10:32:48.279515 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:48Z","lastTransitionTime":"2025-11-25T10:32:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:48 crc kubenswrapper[4821]: I1125 10:32:48.381298 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:48 crc kubenswrapper[4821]: I1125 10:32:48.381344 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:48 crc kubenswrapper[4821]: I1125 10:32:48.381356 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:48 crc kubenswrapper[4821]: I1125 10:32:48.381372 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:48 crc kubenswrapper[4821]: I1125 10:32:48.381383 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:48Z","lastTransitionTime":"2025-11-25T10:32:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:48 crc kubenswrapper[4821]: I1125 10:32:48.483352 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:48 crc kubenswrapper[4821]: I1125 10:32:48.483398 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:48 crc kubenswrapper[4821]: I1125 10:32:48.483406 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:48 crc kubenswrapper[4821]: I1125 10:32:48.483420 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:48 crc kubenswrapper[4821]: I1125 10:32:48.483429 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:48Z","lastTransitionTime":"2025-11-25T10:32:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:48 crc kubenswrapper[4821]: I1125 10:32:48.586298 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:48 crc kubenswrapper[4821]: I1125 10:32:48.586355 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:48 crc kubenswrapper[4821]: I1125 10:32:48.586367 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:48 crc kubenswrapper[4821]: I1125 10:32:48.586384 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:48 crc kubenswrapper[4821]: I1125 10:32:48.586394 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:48Z","lastTransitionTime":"2025-11-25T10:32:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:48 crc kubenswrapper[4821]: I1125 10:32:48.688340 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:48 crc kubenswrapper[4821]: I1125 10:32:48.688372 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:48 crc kubenswrapper[4821]: I1125 10:32:48.688379 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:48 crc kubenswrapper[4821]: I1125 10:32:48.688393 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:48 crc kubenswrapper[4821]: I1125 10:32:48.688402 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:48Z","lastTransitionTime":"2025-11-25T10:32:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:48 crc kubenswrapper[4821]: I1125 10:32:48.790743 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:48 crc kubenswrapper[4821]: I1125 10:32:48.790792 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:48 crc kubenswrapper[4821]: I1125 10:32:48.790806 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:48 crc kubenswrapper[4821]: I1125 10:32:48.790851 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:48 crc kubenswrapper[4821]: I1125 10:32:48.790864 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:48Z","lastTransitionTime":"2025-11-25T10:32:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:48 crc kubenswrapper[4821]: I1125 10:32:48.893372 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:48 crc kubenswrapper[4821]: I1125 10:32:48.893478 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:48 crc kubenswrapper[4821]: I1125 10:32:48.893489 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:48 crc kubenswrapper[4821]: I1125 10:32:48.893501 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:48 crc kubenswrapper[4821]: I1125 10:32:48.893510 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:48Z","lastTransitionTime":"2025-11-25T10:32:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:48 crc kubenswrapper[4821]: I1125 10:32:48.996578 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:48 crc kubenswrapper[4821]: I1125 10:32:48.996624 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:48 crc kubenswrapper[4821]: I1125 10:32:48.996632 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:48 crc kubenswrapper[4821]: I1125 10:32:48.996647 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:48 crc kubenswrapper[4821]: I1125 10:32:48.996658 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:48Z","lastTransitionTime":"2025-11-25T10:32:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:49 crc kubenswrapper[4821]: I1125 10:32:49.099550 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:49 crc kubenswrapper[4821]: I1125 10:32:49.099584 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:49 crc kubenswrapper[4821]: I1125 10:32:49.099596 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:49 crc kubenswrapper[4821]: I1125 10:32:49.099612 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:49 crc kubenswrapper[4821]: I1125 10:32:49.099622 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:49Z","lastTransitionTime":"2025-11-25T10:32:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:49 crc kubenswrapper[4821]: I1125 10:32:49.114032 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:32:49 crc kubenswrapper[4821]: I1125 10:32:49.114099 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:32:49 crc kubenswrapper[4821]: I1125 10:32:49.114121 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:32:49 crc kubenswrapper[4821]: E1125 10:32:49.114312 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:32:49 crc kubenswrapper[4821]: E1125 10:32:49.114370 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:32:49 crc kubenswrapper[4821]: E1125 10:32:49.114466 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:32:49 crc kubenswrapper[4821]: I1125 10:32:49.202275 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:49 crc kubenswrapper[4821]: I1125 10:32:49.202380 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:49 crc kubenswrapper[4821]: I1125 10:32:49.202397 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:49 crc kubenswrapper[4821]: I1125 10:32:49.202421 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:49 crc kubenswrapper[4821]: I1125 10:32:49.202432 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:49Z","lastTransitionTime":"2025-11-25T10:32:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:49 crc kubenswrapper[4821]: I1125 10:32:49.305940 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:49 crc kubenswrapper[4821]: I1125 10:32:49.305977 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:49 crc kubenswrapper[4821]: I1125 10:32:49.305987 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:49 crc kubenswrapper[4821]: I1125 10:32:49.306003 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:49 crc kubenswrapper[4821]: I1125 10:32:49.306014 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:49Z","lastTransitionTime":"2025-11-25T10:32:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:49 crc kubenswrapper[4821]: I1125 10:32:49.408828 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:49 crc kubenswrapper[4821]: I1125 10:32:49.408871 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:49 crc kubenswrapper[4821]: I1125 10:32:49.408883 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:49 crc kubenswrapper[4821]: I1125 10:32:49.408898 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:49 crc kubenswrapper[4821]: I1125 10:32:49.408914 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:49Z","lastTransitionTime":"2025-11-25T10:32:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:49 crc kubenswrapper[4821]: I1125 10:32:49.511091 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:49 crc kubenswrapper[4821]: I1125 10:32:49.511133 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:49 crc kubenswrapper[4821]: I1125 10:32:49.511143 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:49 crc kubenswrapper[4821]: I1125 10:32:49.511192 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:49 crc kubenswrapper[4821]: I1125 10:32:49.511205 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:49Z","lastTransitionTime":"2025-11-25T10:32:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:49 crc kubenswrapper[4821]: I1125 10:32:49.613003 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:49 crc kubenswrapper[4821]: I1125 10:32:49.613042 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:49 crc kubenswrapper[4821]: I1125 10:32:49.613051 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:49 crc kubenswrapper[4821]: I1125 10:32:49.613066 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:49 crc kubenswrapper[4821]: I1125 10:32:49.613075 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:49Z","lastTransitionTime":"2025-11-25T10:32:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:49 crc kubenswrapper[4821]: I1125 10:32:49.714936 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:49 crc kubenswrapper[4821]: I1125 10:32:49.714971 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:49 crc kubenswrapper[4821]: I1125 10:32:49.714978 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:49 crc kubenswrapper[4821]: I1125 10:32:49.714992 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:49 crc kubenswrapper[4821]: I1125 10:32:49.715000 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:49Z","lastTransitionTime":"2025-11-25T10:32:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:49 crc kubenswrapper[4821]: I1125 10:32:49.816848 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:49 crc kubenswrapper[4821]: I1125 10:32:49.816897 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:49 crc kubenswrapper[4821]: I1125 10:32:49.816912 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:49 crc kubenswrapper[4821]: I1125 10:32:49.816931 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:49 crc kubenswrapper[4821]: I1125 10:32:49.816943 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:49Z","lastTransitionTime":"2025-11-25T10:32:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:49 crc kubenswrapper[4821]: I1125 10:32:49.919257 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:49 crc kubenswrapper[4821]: I1125 10:32:49.919298 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:49 crc kubenswrapper[4821]: I1125 10:32:49.919306 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:49 crc kubenswrapper[4821]: I1125 10:32:49.919320 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:49 crc kubenswrapper[4821]: I1125 10:32:49.919329 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:49Z","lastTransitionTime":"2025-11-25T10:32:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:50 crc kubenswrapper[4821]: I1125 10:32:50.022034 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:50 crc kubenswrapper[4821]: I1125 10:32:50.022097 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:50 crc kubenswrapper[4821]: I1125 10:32:50.022110 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:50 crc kubenswrapper[4821]: I1125 10:32:50.022126 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:50 crc kubenswrapper[4821]: I1125 10:32:50.022136 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:50Z","lastTransitionTime":"2025-11-25T10:32:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:50 crc kubenswrapper[4821]: I1125 10:32:50.114113 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4msn" Nov 25 10:32:50 crc kubenswrapper[4821]: E1125 10:32:50.114792 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4msn" podUID="1f12cef3-7e40-4f61-836e-23cd0e578e71" Nov 25 10:32:50 crc kubenswrapper[4821]: I1125 10:32:50.125611 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:50 crc kubenswrapper[4821]: I1125 10:32:50.125655 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:50 crc kubenswrapper[4821]: I1125 10:32:50.125670 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:50 crc kubenswrapper[4821]: I1125 10:32:50.125690 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:50 crc kubenswrapper[4821]: I1125 10:32:50.125707 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:50Z","lastTransitionTime":"2025-11-25T10:32:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:50 crc kubenswrapper[4821]: I1125 10:32:50.128881 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:50Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:50 crc kubenswrapper[4821]: I1125 10:32:50.139763 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f948b87-ac86-4de6-ad64-c2ef947f84d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4482f73a45bc82bd55770c3fd788becacec94893fed1c305fed1c81c753d43aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ba68bd093be94d0d40f3851638db5d26bdfc52d175b1ac78afb91dda7affcfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2krbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:50Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:50 crc kubenswrapper[4821]: I1125 10:32:50.156010 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qppnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"357491b8-cd9c-486e-9898-5ba33678513d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc88d700ca3a449ef1cb32d6124b16d0024d6981ced12e276083f86a1b322ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd9c047c5ce1d896f803ff6cafa9bbf3a18a391b44539e1f1312e08c89e51cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd9c047c5ce1d896f803ff6cafa9bbf3a18a391b44539e1f1312e08c89e51cb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://565a6db63fb69abbba81ffdf6085797900858d263cd6affcb0b5f195474c2516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://565a6db63fb69abbba81ffdf6085797900858d263cd6affcb0b5f195474c2516\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://544885302dffff8336a9aa3c4afa80d69088ffaa2402648f4a329e0a798a8c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://544885302dffff8336a9aa3c4afa80d69088ffaa2402648f4a329e0a798a8c99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8bc193fb3f85b980c31af2e826cee90adb633ad2ee82ff1d46e182ab52e8fab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8bc193fb3f85b980c31af2e826cee90adb633ad2ee82ff1d46e182ab52e8fab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7713e2a35bac8df1e8987fbcd452459fa3541caad0c52ed57fda187bd7c83d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7713e2a35bac8df1e8987fbcd452459fa3541caad0c52ed57fda187bd7c83d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34adfca89026fc20007a73a8cbd4e7c7b3137a581c3476867af17fabe976f71f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34adfca89026fc20007a73a8cbd4e7c7b3137a581c3476867af17fabe976f71f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qppnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:50Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:50 crc kubenswrapper[4821]: I1125 10:32:50.166612 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tgdlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2336955b-a703-4db4-8d0f-d08088ed59b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb0362942bfac635ee6292501b1b851772c01d75dd658eef4b17669a716de72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vjfp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tgdlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:50Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:50 crc kubenswrapper[4821]: I1125 10:32:50.179671 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62fcb632-f12a-46e2-bb7d-a5311f291ee7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28d43531927e75c3745f100940f7f6bf79fd21026c1309194df8cd7a39361b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69633332d03d2551ff4abb253d358de398eab18486b92db630ce06c0447fb9cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ccb78d9a51ac83ee208c2451685a2512de9deaf742a663bb8c892aa59b76162\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27826e155a9352b1e1866e83fdab683dbcd749cc04b6a4da2db25ecabe3d9a52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://991252e5f56f128827a5054b8e228284bc0266166582cbaedaadc53ca2f5b401\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:50Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:50 crc kubenswrapper[4821]: I1125 10:32:50.194093 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384d8ba58264ab89796d3b53afb0c4c1be33cb4fba719a8beef4afc2be7d66a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:50Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:50 crc kubenswrapper[4821]: I1125 10:32:50.207960 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06f949936bac99cfbf89d720c9549c074351d9077bd963fa024d87004c6be23e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:50Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:50 crc kubenswrapper[4821]: I1125 10:32:50.220096 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17b8714be3bd121e06641873e003f08560b1f71e56d6ea936966df7d4896eb03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85731e6b6dbe1777bc7161a010f39f79523ec4b08d71d5d8a975573aa7fe4957\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:50Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:50 crc kubenswrapper[4821]: I1125 10:32:50.228546 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:50 crc kubenswrapper[4821]: I1125 10:32:50.228584 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:50 crc kubenswrapper[4821]: I1125 10:32:50.228593 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:50 crc kubenswrapper[4821]: I1125 10:32:50.228608 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:50 crc kubenswrapper[4821]: I1125 10:32:50.228622 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:50Z","lastTransitionTime":"2025-11-25T10:32:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:50 crc kubenswrapper[4821]: I1125 10:32:50.231982 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zlk96" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d55aa91-9cb9-4c95-bc27-6fec05684c10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4521b5baee45ffc9c7661c781c0b3de458dff325ad525dde283d46af71c0cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vkd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zlk96\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:50Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:50 crc kubenswrapper[4821]: I1125 10:32:50.243313 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6hbqd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f71b3a2-63cb-443c-9f41-39e57030746c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65473b0e63c2e9c7c09a2bb8f31396daf1d87c2c83d20ed3508ce2fe7358dda3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxdmq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://595f9ea75461f18e763d84799e368becc07b72acdc5337d20d6995bf53c3d27a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxdmq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6hbqd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:50Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:50 crc kubenswrapper[4821]: I1125 10:32:50.256017 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:50Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:50 crc kubenswrapper[4821]: I1125 10:32:50.290819 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d9e58df-1755-4422-8d04-211f98c7c011\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea762f627aa69fb88a9bd82545fa677e9ccfa97f99a5e92c0aa490e619c0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a06b811e5ec52ce984a2403ba2c98332ba60466d08b3580ad9192fac6caef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0385f6ed9597974267f7c45547f6b6725b4ee83a2cc0508ff2d1c1ce8753ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b915b565f7fbb3e4e856e7f1035a519e82bfb3023660c121f6e3813de353bb92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18465f3a6916c36457a305ba5d568c9077f34f1d000532ed518f237d100008b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:50Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:50 crc kubenswrapper[4821]: I1125 10:32:50.317514 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25595ccd-762d-414a-9916-4fe38d478b07\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d4b2d8db4d13392d9789e161fd4b8f6eef983b79128ee584d72e4c06565bc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf780f791771ea43a8c8176d02fdfb5204406e023207299369ae46ceb342baaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0251519fd4262f2ea5c64959db49433afdafe36b7ed2404d660eeb7c365d2746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad5c205e0eeecfd43cbdf0eb1a334e3e8faa5e9a9e8dd202c397d5cbb36e081e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:50Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:50 crc kubenswrapper[4821]: I1125 10:32:50.328673 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:50Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:50 crc kubenswrapper[4821]: I1125 10:32:50.330256 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:50 crc kubenswrapper[4821]: I1125 10:32:50.330300 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:50 crc kubenswrapper[4821]: I1125 10:32:50.330313 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:50 crc kubenswrapper[4821]: I1125 10:32:50.330330 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:50 crc kubenswrapper[4821]: I1125 10:32:50.330341 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:50Z","lastTransitionTime":"2025-11-25T10:32:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:50 crc kubenswrapper[4821]: I1125 10:32:50.341497 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ldzbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b84e32ae-4897-4644-a656-6bd82012cc81\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ef433744761f26e8e1f6c0db123821a11923ffd774830744285cd4178d89533\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5pmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ldzbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:50Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:50 crc kubenswrapper[4821]: I1125 10:32:50.358230 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"954085fa-d03a-4e88-9dfd-566257e3558d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f00746930c4588e548f0e2b60d52bf448aff7401b676a7855cb0f4cce00d329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a14a078ba7d2aed95992b36502008f315f0d3a634b64ed49da70398b2562d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d7254196b3702e48083ae98e829b73c539adf69ed5b2bbdc62189d09ae1762\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d727430108977574365f0e9c04b44e03307bd7ff89cdbb1e48a6e916f504d6b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94c7949f077b1ca2b310bedb090aa72fc4bd7e8f2c8bf438096edfca8568f36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed2e7cad7bb24b5b56a1295098a6a7c230c8f9b9c29a88ef653ae8b6ac31667d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a84f30b6a8ad3333268312e9aadacb17eca12613511195c6f8a7734d5f7af5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a84f30b6a8ad3333268312e9aadacb17eca12613511195c6f8a7734d5f7af5b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T10:32:41Z\\\",\\\"message\\\":\\\" 6233 obj_retry.go:365] Adding new object: *v1.Pod openshift-dns/node-resolver-zlk96\\\\nI1125 10:32:41.011897 6233 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-target-xd92c in node crc\\\\nI1125 10:32:41.011901 6233 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-additional-cni-plugins-qppnw after 0 failed attempt(s)\\\\nI1125 10:32:41.011815 6233 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1125 10:32:41.011915 6233 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-additional-cni-plugins-qppnw\\\\nI1125 10:32:41.011902 6233 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-zlk96 in node crc\\\\nI1125 10:32:41.011826 6233 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1125 10:32:41.011923 6233 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-tgdlt after 0 failed attempt(s)\\\\nI1125 10:32:41.011884 6233 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-ingress-operator/metrics]} name:Service_openshift-ingress-operator/metrics_\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-mzhnw_openshift-ovn-kubernetes(954085fa-d03a-4e88-9dfd-566257e3558d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dbf499f4ae6161752abaa7a466a027c6f874016be90a534e69f07e3dd3f3d9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzhnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:50Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:50 crc kubenswrapper[4821]: I1125 10:32:50.367808 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-d4msn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f12cef3-7e40-4f61-836e-23cd0e578e71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s7lkv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s7lkv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:44Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-d4msn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:50Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:50 crc kubenswrapper[4821]: I1125 10:32:50.432775 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:50 crc kubenswrapper[4821]: I1125 10:32:50.432811 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:50 crc kubenswrapper[4821]: I1125 10:32:50.432822 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:50 crc kubenswrapper[4821]: I1125 10:32:50.432838 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:50 crc kubenswrapper[4821]: I1125 10:32:50.432851 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:50Z","lastTransitionTime":"2025-11-25T10:32:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:50 crc kubenswrapper[4821]: I1125 10:32:50.535032 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:50 crc kubenswrapper[4821]: I1125 10:32:50.535098 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:50 crc kubenswrapper[4821]: I1125 10:32:50.535107 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:50 crc kubenswrapper[4821]: I1125 10:32:50.535124 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:50 crc kubenswrapper[4821]: I1125 10:32:50.535134 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:50Z","lastTransitionTime":"2025-11-25T10:32:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:50 crc kubenswrapper[4821]: I1125 10:32:50.637929 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:50 crc kubenswrapper[4821]: I1125 10:32:50.637967 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:50 crc kubenswrapper[4821]: I1125 10:32:50.637977 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:50 crc kubenswrapper[4821]: I1125 10:32:50.637992 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:50 crc kubenswrapper[4821]: I1125 10:32:50.638001 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:50Z","lastTransitionTime":"2025-11-25T10:32:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:50 crc kubenswrapper[4821]: I1125 10:32:50.740248 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:50 crc kubenswrapper[4821]: I1125 10:32:50.740543 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:50 crc kubenswrapper[4821]: I1125 10:32:50.740641 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:50 crc kubenswrapper[4821]: I1125 10:32:50.740733 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:50 crc kubenswrapper[4821]: I1125 10:32:50.740812 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:50Z","lastTransitionTime":"2025-11-25T10:32:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:50 crc kubenswrapper[4821]: I1125 10:32:50.843393 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:50 crc kubenswrapper[4821]: I1125 10:32:50.843431 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:50 crc kubenswrapper[4821]: I1125 10:32:50.843442 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:50 crc kubenswrapper[4821]: I1125 10:32:50.843458 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:50 crc kubenswrapper[4821]: I1125 10:32:50.843469 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:50Z","lastTransitionTime":"2025-11-25T10:32:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:50 crc kubenswrapper[4821]: I1125 10:32:50.947890 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:50 crc kubenswrapper[4821]: I1125 10:32:50.947936 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:50 crc kubenswrapper[4821]: I1125 10:32:50.947947 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:50 crc kubenswrapper[4821]: I1125 10:32:50.947960 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:50 crc kubenswrapper[4821]: I1125 10:32:50.947972 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:50Z","lastTransitionTime":"2025-11-25T10:32:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:51 crc kubenswrapper[4821]: I1125 10:32:51.050196 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:51 crc kubenswrapper[4821]: I1125 10:32:51.050438 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:51 crc kubenswrapper[4821]: I1125 10:32:51.050520 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:51 crc kubenswrapper[4821]: I1125 10:32:51.050612 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:51 crc kubenswrapper[4821]: I1125 10:32:51.050702 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:51Z","lastTransitionTime":"2025-11-25T10:32:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:51 crc kubenswrapper[4821]: I1125 10:32:51.113631 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:32:51 crc kubenswrapper[4821]: I1125 10:32:51.113666 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:32:51 crc kubenswrapper[4821]: E1125 10:32:51.113787 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:32:51 crc kubenswrapper[4821]: E1125 10:32:51.113965 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:32:51 crc kubenswrapper[4821]: I1125 10:32:51.114115 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:32:51 crc kubenswrapper[4821]: E1125 10:32:51.114294 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:32:51 crc kubenswrapper[4821]: I1125 10:32:51.153210 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:51 crc kubenswrapper[4821]: I1125 10:32:51.153240 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:51 crc kubenswrapper[4821]: I1125 10:32:51.153248 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:51 crc kubenswrapper[4821]: I1125 10:32:51.153267 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:51 crc kubenswrapper[4821]: I1125 10:32:51.153286 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:51Z","lastTransitionTime":"2025-11-25T10:32:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:51 crc kubenswrapper[4821]: I1125 10:32:51.255359 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:51 crc kubenswrapper[4821]: I1125 10:32:51.255403 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:51 crc kubenswrapper[4821]: I1125 10:32:51.255413 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:51 crc kubenswrapper[4821]: I1125 10:32:51.255429 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:51 crc kubenswrapper[4821]: I1125 10:32:51.255440 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:51Z","lastTransitionTime":"2025-11-25T10:32:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:51 crc kubenswrapper[4821]: I1125 10:32:51.357950 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:51 crc kubenswrapper[4821]: I1125 10:32:51.357992 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:51 crc kubenswrapper[4821]: I1125 10:32:51.358005 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:51 crc kubenswrapper[4821]: I1125 10:32:51.358023 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:51 crc kubenswrapper[4821]: I1125 10:32:51.358034 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:51Z","lastTransitionTime":"2025-11-25T10:32:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:51 crc kubenswrapper[4821]: I1125 10:32:51.460555 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:51 crc kubenswrapper[4821]: I1125 10:32:51.460632 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:51 crc kubenswrapper[4821]: I1125 10:32:51.460647 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:51 crc kubenswrapper[4821]: I1125 10:32:51.460665 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:51 crc kubenswrapper[4821]: I1125 10:32:51.460675 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:51Z","lastTransitionTime":"2025-11-25T10:32:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:51 crc kubenswrapper[4821]: I1125 10:32:51.563291 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:51 crc kubenswrapper[4821]: I1125 10:32:51.563336 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:51 crc kubenswrapper[4821]: I1125 10:32:51.563348 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:51 crc kubenswrapper[4821]: I1125 10:32:51.563363 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:51 crc kubenswrapper[4821]: I1125 10:32:51.563381 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:51Z","lastTransitionTime":"2025-11-25T10:32:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:51 crc kubenswrapper[4821]: I1125 10:32:51.666991 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:51 crc kubenswrapper[4821]: I1125 10:32:51.667045 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:51 crc kubenswrapper[4821]: I1125 10:32:51.667058 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:51 crc kubenswrapper[4821]: I1125 10:32:51.667075 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:51 crc kubenswrapper[4821]: I1125 10:32:51.667084 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:51Z","lastTransitionTime":"2025-11-25T10:32:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:51 crc kubenswrapper[4821]: I1125 10:32:51.769497 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:51 crc kubenswrapper[4821]: I1125 10:32:51.769768 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:51 crc kubenswrapper[4821]: I1125 10:32:51.769878 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:51 crc kubenswrapper[4821]: I1125 10:32:51.769967 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:51 crc kubenswrapper[4821]: I1125 10:32:51.770056 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:51Z","lastTransitionTime":"2025-11-25T10:32:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:51 crc kubenswrapper[4821]: I1125 10:32:51.872122 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:51 crc kubenswrapper[4821]: I1125 10:32:51.872194 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:51 crc kubenswrapper[4821]: I1125 10:32:51.872208 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:51 crc kubenswrapper[4821]: I1125 10:32:51.872223 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:51 crc kubenswrapper[4821]: I1125 10:32:51.872235 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:51Z","lastTransitionTime":"2025-11-25T10:32:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:51 crc kubenswrapper[4821]: I1125 10:32:51.975245 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:51 crc kubenswrapper[4821]: I1125 10:32:51.975281 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:51 crc kubenswrapper[4821]: I1125 10:32:51.975291 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:51 crc kubenswrapper[4821]: I1125 10:32:51.975305 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:51 crc kubenswrapper[4821]: I1125 10:32:51.975314 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:51Z","lastTransitionTime":"2025-11-25T10:32:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:52 crc kubenswrapper[4821]: I1125 10:32:52.077592 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:52 crc kubenswrapper[4821]: I1125 10:32:52.077660 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:52 crc kubenswrapper[4821]: I1125 10:32:52.077670 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:52 crc kubenswrapper[4821]: I1125 10:32:52.077709 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:52 crc kubenswrapper[4821]: I1125 10:32:52.077725 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:52Z","lastTransitionTime":"2025-11-25T10:32:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:52 crc kubenswrapper[4821]: I1125 10:32:52.113648 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4msn" Nov 25 10:32:52 crc kubenswrapper[4821]: E1125 10:32:52.114013 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4msn" podUID="1f12cef3-7e40-4f61-836e-23cd0e578e71" Nov 25 10:32:52 crc kubenswrapper[4821]: I1125 10:32:52.179818 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:52 crc kubenswrapper[4821]: I1125 10:32:52.179865 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:52 crc kubenswrapper[4821]: I1125 10:32:52.179878 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:52 crc kubenswrapper[4821]: I1125 10:32:52.179893 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:52 crc kubenswrapper[4821]: I1125 10:32:52.179905 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:52Z","lastTransitionTime":"2025-11-25T10:32:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:52 crc kubenswrapper[4821]: I1125 10:32:52.194342 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1f12cef3-7e40-4f61-836e-23cd0e578e71-metrics-certs\") pod \"network-metrics-daemon-d4msn\" (UID: \"1f12cef3-7e40-4f61-836e-23cd0e578e71\") " pod="openshift-multus/network-metrics-daemon-d4msn" Nov 25 10:32:52 crc kubenswrapper[4821]: E1125 10:32:52.194501 4821 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 10:32:52 crc kubenswrapper[4821]: E1125 10:32:52.194555 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1f12cef3-7e40-4f61-836e-23cd0e578e71-metrics-certs podName:1f12cef3-7e40-4f61-836e-23cd0e578e71 nodeName:}" failed. No retries permitted until 2025-11-25 10:33:00.19454112 +0000 UTC m=+50.730860967 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/1f12cef3-7e40-4f61-836e-23cd0e578e71-metrics-certs") pod "network-metrics-daemon-d4msn" (UID: "1f12cef3-7e40-4f61-836e-23cd0e578e71") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 10:32:52 crc kubenswrapper[4821]: I1125 10:32:52.282856 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:52 crc kubenswrapper[4821]: I1125 10:32:52.282905 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:52 crc kubenswrapper[4821]: I1125 10:32:52.282916 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:52 crc kubenswrapper[4821]: I1125 10:32:52.282935 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:52 crc kubenswrapper[4821]: I1125 10:32:52.282947 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:52Z","lastTransitionTime":"2025-11-25T10:32:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:52 crc kubenswrapper[4821]: I1125 10:32:52.386479 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:52 crc kubenswrapper[4821]: I1125 10:32:52.386552 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:52 crc kubenswrapper[4821]: I1125 10:32:52.386567 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:52 crc kubenswrapper[4821]: I1125 10:32:52.386589 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:52 crc kubenswrapper[4821]: I1125 10:32:52.386627 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:52Z","lastTransitionTime":"2025-11-25T10:32:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:52 crc kubenswrapper[4821]: I1125 10:32:52.489529 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:52 crc kubenswrapper[4821]: I1125 10:32:52.489573 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:52 crc kubenswrapper[4821]: I1125 10:32:52.489585 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:52 crc kubenswrapper[4821]: I1125 10:32:52.489600 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:52 crc kubenswrapper[4821]: I1125 10:32:52.489610 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:52Z","lastTransitionTime":"2025-11-25T10:32:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:52 crc kubenswrapper[4821]: I1125 10:32:52.591596 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:52 crc kubenswrapper[4821]: I1125 10:32:52.591630 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:52 crc kubenswrapper[4821]: I1125 10:32:52.591640 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:52 crc kubenswrapper[4821]: I1125 10:32:52.591655 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:52 crc kubenswrapper[4821]: I1125 10:32:52.591666 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:52Z","lastTransitionTime":"2025-11-25T10:32:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:52 crc kubenswrapper[4821]: I1125 10:32:52.694628 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:52 crc kubenswrapper[4821]: I1125 10:32:52.694668 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:52 crc kubenswrapper[4821]: I1125 10:32:52.694676 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:52 crc kubenswrapper[4821]: I1125 10:32:52.694691 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:52 crc kubenswrapper[4821]: I1125 10:32:52.694702 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:52Z","lastTransitionTime":"2025-11-25T10:32:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:52 crc kubenswrapper[4821]: I1125 10:32:52.797056 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:52 crc kubenswrapper[4821]: I1125 10:32:52.797094 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:52 crc kubenswrapper[4821]: I1125 10:32:52.797101 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:52 crc kubenswrapper[4821]: I1125 10:32:52.797116 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:52 crc kubenswrapper[4821]: I1125 10:32:52.797125 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:52Z","lastTransitionTime":"2025-11-25T10:32:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:52 crc kubenswrapper[4821]: I1125 10:32:52.899855 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:52 crc kubenswrapper[4821]: I1125 10:32:52.899933 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:52 crc kubenswrapper[4821]: I1125 10:32:52.899952 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:52 crc kubenswrapper[4821]: I1125 10:32:52.899972 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:52 crc kubenswrapper[4821]: I1125 10:32:52.899986 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:52Z","lastTransitionTime":"2025-11-25T10:32:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:53 crc kubenswrapper[4821]: I1125 10:32:53.002375 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:53 crc kubenswrapper[4821]: I1125 10:32:53.002415 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:53 crc kubenswrapper[4821]: I1125 10:32:53.002423 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:53 crc kubenswrapper[4821]: I1125 10:32:53.002437 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:53 crc kubenswrapper[4821]: I1125 10:32:53.002446 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:53Z","lastTransitionTime":"2025-11-25T10:32:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:53 crc kubenswrapper[4821]: I1125 10:32:53.105002 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:53 crc kubenswrapper[4821]: I1125 10:32:53.105048 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:53 crc kubenswrapper[4821]: I1125 10:32:53.105057 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:53 crc kubenswrapper[4821]: I1125 10:32:53.105079 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:53 crc kubenswrapper[4821]: I1125 10:32:53.105097 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:53Z","lastTransitionTime":"2025-11-25T10:32:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:53 crc kubenswrapper[4821]: I1125 10:32:53.113239 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:32:53 crc kubenswrapper[4821]: I1125 10:32:53.113337 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:32:53 crc kubenswrapper[4821]: E1125 10:32:53.113368 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:32:53 crc kubenswrapper[4821]: I1125 10:32:53.113404 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:32:53 crc kubenswrapper[4821]: E1125 10:32:53.113538 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:32:53 crc kubenswrapper[4821]: E1125 10:32:53.113684 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:32:53 crc kubenswrapper[4821]: I1125 10:32:53.207668 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:53 crc kubenswrapper[4821]: I1125 10:32:53.207700 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:53 crc kubenswrapper[4821]: I1125 10:32:53.207709 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:53 crc kubenswrapper[4821]: I1125 10:32:53.207750 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:53 crc kubenswrapper[4821]: I1125 10:32:53.207760 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:53Z","lastTransitionTime":"2025-11-25T10:32:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:53 crc kubenswrapper[4821]: I1125 10:32:53.309882 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:53 crc kubenswrapper[4821]: I1125 10:32:53.309932 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:53 crc kubenswrapper[4821]: I1125 10:32:53.309943 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:53 crc kubenswrapper[4821]: I1125 10:32:53.309959 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:53 crc kubenswrapper[4821]: I1125 10:32:53.309968 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:53Z","lastTransitionTime":"2025-11-25T10:32:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:53 crc kubenswrapper[4821]: I1125 10:32:53.413362 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:53 crc kubenswrapper[4821]: I1125 10:32:53.413401 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:53 crc kubenswrapper[4821]: I1125 10:32:53.413419 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:53 crc kubenswrapper[4821]: I1125 10:32:53.413436 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:53 crc kubenswrapper[4821]: I1125 10:32:53.413447 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:53Z","lastTransitionTime":"2025-11-25T10:32:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:53 crc kubenswrapper[4821]: I1125 10:32:53.514950 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:53 crc kubenswrapper[4821]: I1125 10:32:53.514998 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:53 crc kubenswrapper[4821]: I1125 10:32:53.515011 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:53 crc kubenswrapper[4821]: I1125 10:32:53.515027 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:53 crc kubenswrapper[4821]: I1125 10:32:53.515038 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:53Z","lastTransitionTime":"2025-11-25T10:32:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:53 crc kubenswrapper[4821]: I1125 10:32:53.617448 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:53 crc kubenswrapper[4821]: I1125 10:32:53.617490 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:53 crc kubenswrapper[4821]: I1125 10:32:53.617500 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:53 crc kubenswrapper[4821]: I1125 10:32:53.617514 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:53 crc kubenswrapper[4821]: I1125 10:32:53.617523 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:53Z","lastTransitionTime":"2025-11-25T10:32:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:53 crc kubenswrapper[4821]: I1125 10:32:53.719906 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:53 crc kubenswrapper[4821]: I1125 10:32:53.719938 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:53 crc kubenswrapper[4821]: I1125 10:32:53.719949 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:53 crc kubenswrapper[4821]: I1125 10:32:53.719964 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:53 crc kubenswrapper[4821]: I1125 10:32:53.719975 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:53Z","lastTransitionTime":"2025-11-25T10:32:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:53 crc kubenswrapper[4821]: I1125 10:32:53.822544 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:53 crc kubenswrapper[4821]: I1125 10:32:53.822586 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:53 crc kubenswrapper[4821]: I1125 10:32:53.822602 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:53 crc kubenswrapper[4821]: I1125 10:32:53.822618 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:53 crc kubenswrapper[4821]: I1125 10:32:53.822629 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:53Z","lastTransitionTime":"2025-11-25T10:32:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:53 crc kubenswrapper[4821]: I1125 10:32:53.925254 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:53 crc kubenswrapper[4821]: I1125 10:32:53.925308 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:53 crc kubenswrapper[4821]: I1125 10:32:53.925319 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:53 crc kubenswrapper[4821]: I1125 10:32:53.925337 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:53 crc kubenswrapper[4821]: I1125 10:32:53.925348 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:53Z","lastTransitionTime":"2025-11-25T10:32:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:54 crc kubenswrapper[4821]: I1125 10:32:54.027666 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:54 crc kubenswrapper[4821]: I1125 10:32:54.027731 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:54 crc kubenswrapper[4821]: I1125 10:32:54.027748 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:54 crc kubenswrapper[4821]: I1125 10:32:54.027769 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:54 crc kubenswrapper[4821]: I1125 10:32:54.027784 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:54Z","lastTransitionTime":"2025-11-25T10:32:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:54 crc kubenswrapper[4821]: I1125 10:32:54.114122 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4msn" Nov 25 10:32:54 crc kubenswrapper[4821]: E1125 10:32:54.114444 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4msn" podUID="1f12cef3-7e40-4f61-836e-23cd0e578e71" Nov 25 10:32:54 crc kubenswrapper[4821]: I1125 10:32:54.130040 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:54 crc kubenswrapper[4821]: I1125 10:32:54.130085 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:54 crc kubenswrapper[4821]: I1125 10:32:54.130095 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:54 crc kubenswrapper[4821]: I1125 10:32:54.130114 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:54 crc kubenswrapper[4821]: I1125 10:32:54.130125 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:54Z","lastTransitionTime":"2025-11-25T10:32:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:54 crc kubenswrapper[4821]: I1125 10:32:54.232701 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:54 crc kubenswrapper[4821]: I1125 10:32:54.232738 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:54 crc kubenswrapper[4821]: I1125 10:32:54.232748 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:54 crc kubenswrapper[4821]: I1125 10:32:54.232763 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:54 crc kubenswrapper[4821]: I1125 10:32:54.232774 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:54Z","lastTransitionTime":"2025-11-25T10:32:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:54 crc kubenswrapper[4821]: I1125 10:32:54.335847 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:54 crc kubenswrapper[4821]: I1125 10:32:54.335895 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:54 crc kubenswrapper[4821]: I1125 10:32:54.335907 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:54 crc kubenswrapper[4821]: I1125 10:32:54.335925 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:54 crc kubenswrapper[4821]: I1125 10:32:54.335940 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:54Z","lastTransitionTime":"2025-11-25T10:32:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:54 crc kubenswrapper[4821]: I1125 10:32:54.438433 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:54 crc kubenswrapper[4821]: I1125 10:32:54.438478 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:54 crc kubenswrapper[4821]: I1125 10:32:54.438486 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:54 crc kubenswrapper[4821]: I1125 10:32:54.438502 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:54 crc kubenswrapper[4821]: I1125 10:32:54.438511 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:54Z","lastTransitionTime":"2025-11-25T10:32:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:54 crc kubenswrapper[4821]: I1125 10:32:54.540767 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:54 crc kubenswrapper[4821]: I1125 10:32:54.540796 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:54 crc kubenswrapper[4821]: I1125 10:32:54.540804 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:54 crc kubenswrapper[4821]: I1125 10:32:54.540819 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:54 crc kubenswrapper[4821]: I1125 10:32:54.540827 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:54Z","lastTransitionTime":"2025-11-25T10:32:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:54 crc kubenswrapper[4821]: I1125 10:32:54.643474 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:54 crc kubenswrapper[4821]: I1125 10:32:54.643518 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:54 crc kubenswrapper[4821]: I1125 10:32:54.643526 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:54 crc kubenswrapper[4821]: I1125 10:32:54.643547 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:54 crc kubenswrapper[4821]: I1125 10:32:54.643559 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:54Z","lastTransitionTime":"2025-11-25T10:32:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:54 crc kubenswrapper[4821]: I1125 10:32:54.747100 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:54 crc kubenswrapper[4821]: I1125 10:32:54.747185 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:54 crc kubenswrapper[4821]: I1125 10:32:54.747202 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:54 crc kubenswrapper[4821]: I1125 10:32:54.747219 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:54 crc kubenswrapper[4821]: I1125 10:32:54.747231 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:54Z","lastTransitionTime":"2025-11-25T10:32:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:54 crc kubenswrapper[4821]: I1125 10:32:54.849543 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:54 crc kubenswrapper[4821]: I1125 10:32:54.849804 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:54 crc kubenswrapper[4821]: I1125 10:32:54.849895 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:54 crc kubenswrapper[4821]: I1125 10:32:54.849983 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:54 crc kubenswrapper[4821]: I1125 10:32:54.850066 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:54Z","lastTransitionTime":"2025-11-25T10:32:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:54 crc kubenswrapper[4821]: I1125 10:32:54.952438 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:54 crc kubenswrapper[4821]: I1125 10:32:54.952753 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:54 crc kubenswrapper[4821]: I1125 10:32:54.952847 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:54 crc kubenswrapper[4821]: I1125 10:32:54.952935 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:54 crc kubenswrapper[4821]: I1125 10:32:54.953031 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:54Z","lastTransitionTime":"2025-11-25T10:32:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:55 crc kubenswrapper[4821]: I1125 10:32:55.055522 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:55 crc kubenswrapper[4821]: I1125 10:32:55.055562 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:55 crc kubenswrapper[4821]: I1125 10:32:55.055572 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:55 crc kubenswrapper[4821]: I1125 10:32:55.055588 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:55 crc kubenswrapper[4821]: I1125 10:32:55.055599 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:55Z","lastTransitionTime":"2025-11-25T10:32:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:55 crc kubenswrapper[4821]: I1125 10:32:55.114060 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:32:55 crc kubenswrapper[4821]: I1125 10:32:55.114124 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:32:55 crc kubenswrapper[4821]: I1125 10:32:55.114480 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:32:55 crc kubenswrapper[4821]: E1125 10:32:55.114609 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:32:55 crc kubenswrapper[4821]: E1125 10:32:55.114783 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:32:55 crc kubenswrapper[4821]: E1125 10:32:55.114906 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:32:55 crc kubenswrapper[4821]: I1125 10:32:55.114953 4821 scope.go:117] "RemoveContainer" containerID="5a84f30b6a8ad3333268312e9aadacb17eca12613511195c6f8a7734d5f7af5b" Nov 25 10:32:55 crc kubenswrapper[4821]: I1125 10:32:55.158276 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:55 crc kubenswrapper[4821]: I1125 10:32:55.158328 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:55 crc kubenswrapper[4821]: I1125 10:32:55.158339 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:55 crc kubenswrapper[4821]: I1125 10:32:55.158354 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:55 crc kubenswrapper[4821]: I1125 10:32:55.158366 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:55Z","lastTransitionTime":"2025-11-25T10:32:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:55 crc kubenswrapper[4821]: I1125 10:32:55.272455 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:55 crc kubenswrapper[4821]: I1125 10:32:55.273552 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:55 crc kubenswrapper[4821]: I1125 10:32:55.273578 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:55 crc kubenswrapper[4821]: I1125 10:32:55.273598 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:55 crc kubenswrapper[4821]: I1125 10:32:55.273616 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:55Z","lastTransitionTime":"2025-11-25T10:32:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:55 crc kubenswrapper[4821]: I1125 10:32:55.376637 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:55 crc kubenswrapper[4821]: I1125 10:32:55.376704 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:55 crc kubenswrapper[4821]: I1125 10:32:55.376717 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:55 crc kubenswrapper[4821]: I1125 10:32:55.376736 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:55 crc kubenswrapper[4821]: I1125 10:32:55.376748 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:55Z","lastTransitionTime":"2025-11-25T10:32:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:55 crc kubenswrapper[4821]: I1125 10:32:55.478821 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:55 crc kubenswrapper[4821]: I1125 10:32:55.478863 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:55 crc kubenswrapper[4821]: I1125 10:32:55.478874 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:55 crc kubenswrapper[4821]: I1125 10:32:55.478889 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:55 crc kubenswrapper[4821]: I1125 10:32:55.478900 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:55Z","lastTransitionTime":"2025-11-25T10:32:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:55 crc kubenswrapper[4821]: I1125 10:32:55.582295 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:55 crc kubenswrapper[4821]: I1125 10:32:55.582345 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:55 crc kubenswrapper[4821]: I1125 10:32:55.582356 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:55 crc kubenswrapper[4821]: I1125 10:32:55.582374 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:55 crc kubenswrapper[4821]: I1125 10:32:55.582385 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:55Z","lastTransitionTime":"2025-11-25T10:32:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:55 crc kubenswrapper[4821]: I1125 10:32:55.685075 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:55 crc kubenswrapper[4821]: I1125 10:32:55.685130 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:55 crc kubenswrapper[4821]: I1125 10:32:55.685142 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:55 crc kubenswrapper[4821]: I1125 10:32:55.685181 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:55 crc kubenswrapper[4821]: I1125 10:32:55.685201 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:55Z","lastTransitionTime":"2025-11-25T10:32:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:55 crc kubenswrapper[4821]: I1125 10:32:55.788103 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:55 crc kubenswrapper[4821]: I1125 10:32:55.788142 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:55 crc kubenswrapper[4821]: I1125 10:32:55.788150 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:55 crc kubenswrapper[4821]: I1125 10:32:55.788194 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:55 crc kubenswrapper[4821]: I1125 10:32:55.788213 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:55Z","lastTransitionTime":"2025-11-25T10:32:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:55 crc kubenswrapper[4821]: I1125 10:32:55.890284 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:55 crc kubenswrapper[4821]: I1125 10:32:55.890327 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:55 crc kubenswrapper[4821]: I1125 10:32:55.890337 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:55 crc kubenswrapper[4821]: I1125 10:32:55.890355 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:55 crc kubenswrapper[4821]: I1125 10:32:55.890365 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:55Z","lastTransitionTime":"2025-11-25T10:32:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:55 crc kubenswrapper[4821]: I1125 10:32:55.993493 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:55 crc kubenswrapper[4821]: I1125 10:32:55.993534 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:55 crc kubenswrapper[4821]: I1125 10:32:55.993544 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:55 crc kubenswrapper[4821]: I1125 10:32:55.993560 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:55 crc kubenswrapper[4821]: I1125 10:32:55.993569 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:55Z","lastTransitionTime":"2025-11-25T10:32:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.095718 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.095760 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.095768 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.095788 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.095800 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:56Z","lastTransitionTime":"2025-11-25T10:32:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.114070 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4msn" Nov 25 10:32:56 crc kubenswrapper[4821]: E1125 10:32:56.114256 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4msn" podUID="1f12cef3-7e40-4f61-836e-23cd0e578e71" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.198122 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.198196 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.198206 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.198221 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.198232 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:56Z","lastTransitionTime":"2025-11-25T10:32:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.294029 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.294082 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.294098 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.294123 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.294137 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:56Z","lastTransitionTime":"2025-11-25T10:32:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:56 crc kubenswrapper[4821]: E1125 10:32:56.308738 4821 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:32:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:32:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:32:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:32:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a5b2b070-39b9-412c-90f1-c7ed2e9e0950\\\",\\\"systemUUID\\\":\\\"ea86e80f-41b7-4190-ad7e-d8c549d535de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:56Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.313065 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.313136 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.313155 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.313203 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.313218 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:56Z","lastTransitionTime":"2025-11-25T10:32:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:56 crc kubenswrapper[4821]: E1125 10:32:56.326393 4821 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:32:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:32:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:32:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:32:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a5b2b070-39b9-412c-90f1-c7ed2e9e0950\\\",\\\"systemUUID\\\":\\\"ea86e80f-41b7-4190-ad7e-d8c549d535de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:56Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.330982 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.331025 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.331037 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.331056 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.331068 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:56Z","lastTransitionTime":"2025-11-25T10:32:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:56 crc kubenswrapper[4821]: E1125 10:32:56.343086 4821 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:32:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:32:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:32:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:32:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a5b2b070-39b9-412c-90f1-c7ed2e9e0950\\\",\\\"systemUUID\\\":\\\"ea86e80f-41b7-4190-ad7e-d8c549d535de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:56Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.347109 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.347143 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.347151 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.347187 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.347201 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:56Z","lastTransitionTime":"2025-11-25T10:32:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.347339 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mzhnw_954085fa-d03a-4e88-9dfd-566257e3558d/ovnkube-controller/2.log" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.347923 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mzhnw_954085fa-d03a-4e88-9dfd-566257e3558d/ovnkube-controller/1.log" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.350953 4821 generic.go:334] "Generic (PLEG): container finished" podID="954085fa-d03a-4e88-9dfd-566257e3558d" containerID="7e6202b51ac13209f3a6de880638a6d142f79755311646c0216f11dfe8bbe3f7" exitCode=1 Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.350996 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" event={"ID":"954085fa-d03a-4e88-9dfd-566257e3558d","Type":"ContainerDied","Data":"7e6202b51ac13209f3a6de880638a6d142f79755311646c0216f11dfe8bbe3f7"} Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.351033 4821 scope.go:117] "RemoveContainer" containerID="5a84f30b6a8ad3333268312e9aadacb17eca12613511195c6f8a7734d5f7af5b" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.351812 4821 scope.go:117] "RemoveContainer" containerID="7e6202b51ac13209f3a6de880638a6d142f79755311646c0216f11dfe8bbe3f7" Nov 25 10:32:56 crc kubenswrapper[4821]: E1125 10:32:56.352007 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-mzhnw_openshift-ovn-kubernetes(954085fa-d03a-4e88-9dfd-566257e3558d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" podUID="954085fa-d03a-4e88-9dfd-566257e3558d" Nov 25 10:32:56 crc kubenswrapper[4821]: E1125 10:32:56.363257 4821 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:32:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:32:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:32:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:32:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a5b2b070-39b9-412c-90f1-c7ed2e9e0950\\\",\\\"systemUUID\\\":\\\"ea86e80f-41b7-4190-ad7e-d8c549d535de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:56Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.364924 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06f949936bac99cfbf89d720c9549c074351d9077bd963fa024d87004c6be23e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:56Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.367652 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.367703 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.367721 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.367746 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.367760 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:56Z","lastTransitionTime":"2025-11-25T10:32:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.378097 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384d8ba58264ab89796d3b53afb0c4c1be33cb4fba719a8beef4afc2be7d66a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:56Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:56 crc kubenswrapper[4821]: E1125 10:32:56.380680 4821 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:32:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:32:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:32:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:32:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a5b2b070-39b9-412c-90f1-c7ed2e9e0950\\\",\\\"systemUUID\\\":\\\"ea86e80f-41b7-4190-ad7e-d8c549d535de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:56Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:56 crc kubenswrapper[4821]: E1125 10:32:56.380800 4821 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.382973 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.383026 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.383038 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.383056 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.383069 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:56Z","lastTransitionTime":"2025-11-25T10:32:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.392275 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:56Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.406856 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17b8714be3bd121e06641873e003f08560b1f71e56d6ea936966df7d4896eb03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85731e6b6dbe1777bc7161a010f39f79523ec4b08d71d5d8a975573aa7fe4957\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:56Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.420408 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zlk96" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d55aa91-9cb9-4c95-bc27-6fec05684c10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4521b5baee45ffc9c7661c781c0b3de458dff325ad525dde283d46af71c0cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vkd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zlk96\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:56Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.431942 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6hbqd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f71b3a2-63cb-443c-9f41-39e57030746c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65473b0e63c2e9c7c09a2bb8f31396daf1d87c2c83d20ed3508ce2fe7358dda3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxdmq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://595f9ea75461f18e763d84799e368becc07b72acdc5337d20d6995bf53c3d27a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxdmq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6hbqd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:56Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.448762 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"954085fa-d03a-4e88-9dfd-566257e3558d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f00746930c4588e548f0e2b60d52bf448aff7401b676a7855cb0f4cce00d329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a14a078ba7d2aed95992b36502008f315f0d3a634b64ed49da70398b2562d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d7254196b3702e48083ae98e829b73c539adf69ed5b2bbdc62189d09ae1762\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d727430108977574365f0e9c04b44e03307bd7ff89cdbb1e48a6e916f504d6b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94c7949f077b1ca2b310bedb090aa72fc4bd7e8f2c8bf438096edfca8568f36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed2e7cad7bb24b5b56a1295098a6a7c230c8f9b9c29a88ef653ae8b6ac31667d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e6202b51ac13209f3a6de880638a6d142f79755311646c0216f11dfe8bbe3f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a84f30b6a8ad3333268312e9aadacb17eca12613511195c6f8a7734d5f7af5b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T10:32:41Z\\\",\\\"message\\\":\\\" 6233 obj_retry.go:365] Adding new object: *v1.Pod openshift-dns/node-resolver-zlk96\\\\nI1125 10:32:41.011897 6233 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-target-xd92c in node crc\\\\nI1125 10:32:41.011901 6233 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-additional-cni-plugins-qppnw after 0 failed attempt(s)\\\\nI1125 10:32:41.011815 6233 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1125 10:32:41.011915 6233 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-additional-cni-plugins-qppnw\\\\nI1125 10:32:41.011902 6233 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-zlk96 in node crc\\\\nI1125 10:32:41.011826 6233 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1125 10:32:41.011923 6233 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-tgdlt after 0 failed attempt(s)\\\\nI1125 10:32:41.011884 6233 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-ingress-operator/metrics]} name:Service_openshift-ingress-operator/metrics_\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e6202b51ac13209f3a6de880638a6d142f79755311646c0216f11dfe8bbe3f7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T10:32:56Z\\\",\\\"message\\\":\\\"to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:56Z is after 2025-08-24T17:21:41Z]\\\\nI1125 10:32:56.048189 6453 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-console/console_TCP_cluster\\\\\\\", UUID:\\\\\\\"d7d7b270-1480-47f8-bdf9-690dbab310cb\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console/console\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]strin\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dbf499f4ae6161752abaa7a466a027c6f874016be90a534e69f07e3dd3f3d9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzhnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:56Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.459522 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-d4msn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f12cef3-7e40-4f61-836e-23cd0e578e71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s7lkv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s7lkv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:44Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-d4msn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:56Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.477673 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d9e58df-1755-4422-8d04-211f98c7c011\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea762f627aa69fb88a9bd82545fa677e9ccfa97f99a5e92c0aa490e619c0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a06b811e5ec52ce984a2403ba2c98332ba60466d08b3580ad9192fac6caef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0385f6ed9597974267f7c45547f6b6725b4ee83a2cc0508ff2d1c1ce8753ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b915b565f7fbb3e4e856e7f1035a519e82bfb3023660c121f6e3813de353bb92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18465f3a6916c36457a305ba5d568c9077f34f1d000532ed518f237d100008b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:56Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.485291 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.485327 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.485337 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.485360 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.485370 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:56Z","lastTransitionTime":"2025-11-25T10:32:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.489825 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25595ccd-762d-414a-9916-4fe38d478b07\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d4b2d8db4d13392d9789e161fd4b8f6eef983b79128ee584d72e4c06565bc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf780f791771ea43a8c8176d02fdfb5204406e023207299369ae46ceb342baaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0251519fd4262f2ea5c64959db49433afdafe36b7ed2404d660eeb7c365d2746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad5c205e0eeecfd43cbdf0eb1a334e3e8faa5e9a9e8dd202c397d5cbb36e081e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:56Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.502418 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:56Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.514081 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ldzbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b84e32ae-4897-4644-a656-6bd82012cc81\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ef433744761f26e8e1f6c0db123821a11923ffd774830744285cd4178d89533\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5pmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ldzbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:56Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.523813 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tgdlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2336955b-a703-4db4-8d0f-d08088ed59b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb0362942bfac635ee6292501b1b851772c01d75dd658eef4b17669a716de72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vjfp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tgdlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:56Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.536801 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62fcb632-f12a-46e2-bb7d-a5311f291ee7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28d43531927e75c3745f100940f7f6bf79fd21026c1309194df8cd7a39361b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69633332d03d2551ff4abb253d358de398eab18486b92db630ce06c0447fb9cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ccb78d9a51ac83ee208c2451685a2512de9deaf742a663bb8c892aa59b76162\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27826e155a9352b1e1866e83fdab683dbcd749cc04b6a4da2db25ecabe3d9a52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://991252e5f56f128827a5054b8e228284bc0266166582cbaedaadc53ca2f5b401\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:56Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.549151 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:56Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.560729 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f948b87-ac86-4de6-ad64-c2ef947f84d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4482f73a45bc82bd55770c3fd788becacec94893fed1c305fed1c81c753d43aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ba68bd093be94d0d40f3851638db5d26bdfc52d175b1ac78afb91dda7affcfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2krbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:56Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.575182 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qppnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"357491b8-cd9c-486e-9898-5ba33678513d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc88d700ca3a449ef1cb32d6124b16d0024d6981ced12e276083f86a1b322ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd9c047c5ce1d896f803ff6cafa9bbf3a18a391b44539e1f1312e08c89e51cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd9c047c5ce1d896f803ff6cafa9bbf3a18a391b44539e1f1312e08c89e51cb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://565a6db63fb69abbba81ffdf6085797900858d263cd6affcb0b5f195474c2516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://565a6db63fb69abbba81ffdf6085797900858d263cd6affcb0b5f195474c2516\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://544885302dffff8336a9aa3c4afa80d69088ffaa2402648f4a329e0a798a8c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://544885302dffff8336a9aa3c4afa80d69088ffaa2402648f4a329e0a798a8c99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8bc193fb3f85b980c31af2e826cee90adb633ad2ee82ff1d46e182ab52e8fab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8bc193fb3f85b980c31af2e826cee90adb633ad2ee82ff1d46e182ab52e8fab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7713e2a35bac8df1e8987fbcd452459fa3541caad0c52ed57fda187bd7c83d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7713e2a35bac8df1e8987fbcd452459fa3541caad0c52ed57fda187bd7c83d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34adfca89026fc20007a73a8cbd4e7c7b3137a581c3476867af17fabe976f71f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34adfca89026fc20007a73a8cbd4e7c7b3137a581c3476867af17fabe976f71f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qppnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:56Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.587890 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.587936 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.587946 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.587964 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.587975 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:56Z","lastTransitionTime":"2025-11-25T10:32:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.690515 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.690571 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.690584 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.690600 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.690610 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:56Z","lastTransitionTime":"2025-11-25T10:32:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.793150 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.793223 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.793233 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.793251 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.793265 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:56Z","lastTransitionTime":"2025-11-25T10:32:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.895723 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.895757 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.895766 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.895780 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.895789 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:56Z","lastTransitionTime":"2025-11-25T10:32:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.998352 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.998457 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.998475 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.998498 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:56 crc kubenswrapper[4821]: I1125 10:32:56.998517 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:56Z","lastTransitionTime":"2025-11-25T10:32:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:57 crc kubenswrapper[4821]: I1125 10:32:57.101661 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:57 crc kubenswrapper[4821]: I1125 10:32:57.101710 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:57 crc kubenswrapper[4821]: I1125 10:32:57.101724 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:57 crc kubenswrapper[4821]: I1125 10:32:57.101746 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:57 crc kubenswrapper[4821]: I1125 10:32:57.101762 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:57Z","lastTransitionTime":"2025-11-25T10:32:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:57 crc kubenswrapper[4821]: I1125 10:32:57.113909 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:32:57 crc kubenswrapper[4821]: I1125 10:32:57.114029 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:32:57 crc kubenswrapper[4821]: E1125 10:32:57.114023 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:32:57 crc kubenswrapper[4821]: I1125 10:32:57.114102 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:32:57 crc kubenswrapper[4821]: E1125 10:32:57.114292 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:32:57 crc kubenswrapper[4821]: E1125 10:32:57.114115 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:32:57 crc kubenswrapper[4821]: I1125 10:32:57.205024 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:57 crc kubenswrapper[4821]: I1125 10:32:57.205083 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:57 crc kubenswrapper[4821]: I1125 10:32:57.205096 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:57 crc kubenswrapper[4821]: I1125 10:32:57.205120 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:57 crc kubenswrapper[4821]: I1125 10:32:57.205133 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:57Z","lastTransitionTime":"2025-11-25T10:32:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:57 crc kubenswrapper[4821]: I1125 10:32:57.308172 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:57 crc kubenswrapper[4821]: I1125 10:32:57.308223 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:57 crc kubenswrapper[4821]: I1125 10:32:57.308234 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:57 crc kubenswrapper[4821]: I1125 10:32:57.308255 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:57 crc kubenswrapper[4821]: I1125 10:32:57.308267 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:57Z","lastTransitionTime":"2025-11-25T10:32:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:57 crc kubenswrapper[4821]: I1125 10:32:57.356468 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mzhnw_954085fa-d03a-4e88-9dfd-566257e3558d/ovnkube-controller/2.log" Nov 25 10:32:57 crc kubenswrapper[4821]: I1125 10:32:57.389122 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 10:32:57 crc kubenswrapper[4821]: I1125 10:32:57.396977 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Nov 25 10:32:57 crc kubenswrapper[4821]: I1125 10:32:57.401398 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25595ccd-762d-414a-9916-4fe38d478b07\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d4b2d8db4d13392d9789e161fd4b8f6eef983b79128ee584d72e4c06565bc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf780f791771ea43a8c8176d02fdfb5204406e023207299369ae46ceb342baaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0251519fd4262f2ea5c64959db49433afdafe36b7ed2404d660eeb7c365d2746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad5c205e0eeecfd43cbdf0eb1a334e3e8faa5e9a9e8dd202c397d5cbb36e081e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:57Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:57 crc kubenswrapper[4821]: I1125 10:32:57.410392 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:57 crc kubenswrapper[4821]: I1125 10:32:57.410418 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:57 crc kubenswrapper[4821]: I1125 10:32:57.410428 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:57 crc kubenswrapper[4821]: I1125 10:32:57.410442 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:57 crc kubenswrapper[4821]: I1125 10:32:57.410452 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:57Z","lastTransitionTime":"2025-11-25T10:32:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:57 crc kubenswrapper[4821]: I1125 10:32:57.414625 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:57Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:57 crc kubenswrapper[4821]: I1125 10:32:57.426977 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ldzbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b84e32ae-4897-4644-a656-6bd82012cc81\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ef433744761f26e8e1f6c0db123821a11923ffd774830744285cd4178d89533\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5pmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ldzbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:57Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:57 crc kubenswrapper[4821]: I1125 10:32:57.445593 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"954085fa-d03a-4e88-9dfd-566257e3558d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f00746930c4588e548f0e2b60d52bf448aff7401b676a7855cb0f4cce00d329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a14a078ba7d2aed95992b36502008f315f0d3a634b64ed49da70398b2562d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d7254196b3702e48083ae98e829b73c539adf69ed5b2bbdc62189d09ae1762\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d727430108977574365f0e9c04b44e03307bd7ff89cdbb1e48a6e916f504d6b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94c7949f077b1ca2b310bedb090aa72fc4bd7e8f2c8bf438096edfca8568f36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed2e7cad7bb24b5b56a1295098a6a7c230c8f9b9c29a88ef653ae8b6ac31667d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e6202b51ac13209f3a6de880638a6d142f79755311646c0216f11dfe8bbe3f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a84f30b6a8ad3333268312e9aadacb17eca12613511195c6f8a7734d5f7af5b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T10:32:41Z\\\",\\\"message\\\":\\\" 6233 obj_retry.go:365] Adding new object: *v1.Pod openshift-dns/node-resolver-zlk96\\\\nI1125 10:32:41.011897 6233 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-target-xd92c in node crc\\\\nI1125 10:32:41.011901 6233 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-additional-cni-plugins-qppnw after 0 failed attempt(s)\\\\nI1125 10:32:41.011815 6233 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1125 10:32:41.011915 6233 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-additional-cni-plugins-qppnw\\\\nI1125 10:32:41.011902 6233 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-zlk96 in node crc\\\\nI1125 10:32:41.011826 6233 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1125 10:32:41.011923 6233 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-tgdlt after 0 failed attempt(s)\\\\nI1125 10:32:41.011884 6233 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-ingress-operator/metrics]} name:Service_openshift-ingress-operator/metrics_\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e6202b51ac13209f3a6de880638a6d142f79755311646c0216f11dfe8bbe3f7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T10:32:56Z\\\",\\\"message\\\":\\\"to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:56Z is after 2025-08-24T17:21:41Z]\\\\nI1125 10:32:56.048189 6453 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-console/console_TCP_cluster\\\\\\\", UUID:\\\\\\\"d7d7b270-1480-47f8-bdf9-690dbab310cb\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console/console\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]strin\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dbf499f4ae6161752abaa7a466a027c6f874016be90a534e69f07e3dd3f3d9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzhnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:57Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:57 crc kubenswrapper[4821]: I1125 10:32:57.457123 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-d4msn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f12cef3-7e40-4f61-836e-23cd0e578e71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s7lkv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s7lkv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:44Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-d4msn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:57Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:57 crc kubenswrapper[4821]: I1125 10:32:57.485984 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d9e58df-1755-4422-8d04-211f98c7c011\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea762f627aa69fb88a9bd82545fa677e9ccfa97f99a5e92c0aa490e619c0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a06b811e5ec52ce984a2403ba2c98332ba60466d08b3580ad9192fac6caef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0385f6ed9597974267f7c45547f6b6725b4ee83a2cc0508ff2d1c1ce8753ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b915b565f7fbb3e4e856e7f1035a519e82bfb3023660c121f6e3813de353bb92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18465f3a6916c36457a305ba5d568c9077f34f1d000532ed518f237d100008b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:57Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:57 crc kubenswrapper[4821]: I1125 10:32:57.500640 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:57Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:57 crc kubenswrapper[4821]: I1125 10:32:57.512602 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:57 crc kubenswrapper[4821]: I1125 10:32:57.512633 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:57 crc kubenswrapper[4821]: I1125 10:32:57.512643 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:57 crc kubenswrapper[4821]: I1125 10:32:57.512660 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:57 crc kubenswrapper[4821]: I1125 10:32:57.512672 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:57Z","lastTransitionTime":"2025-11-25T10:32:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:57 crc kubenswrapper[4821]: I1125 10:32:57.513722 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f948b87-ac86-4de6-ad64-c2ef947f84d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4482f73a45bc82bd55770c3fd788becacec94893fed1c305fed1c81c753d43aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ba68bd093be94d0d40f3851638db5d26bdfc52d175b1ac78afb91dda7affcfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2krbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:57Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:57 crc kubenswrapper[4821]: I1125 10:32:57.533755 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qppnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"357491b8-cd9c-486e-9898-5ba33678513d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc88d700ca3a449ef1cb32d6124b16d0024d6981ced12e276083f86a1b322ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd9c047c5ce1d896f803ff6cafa9bbf3a18a391b44539e1f1312e08c89e51cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd9c047c5ce1d896f803ff6cafa9bbf3a18a391b44539e1f1312e08c89e51cb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://565a6db63fb69abbba81ffdf6085797900858d263cd6affcb0b5f195474c2516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://565a6db63fb69abbba81ffdf6085797900858d263cd6affcb0b5f195474c2516\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://544885302dffff8336a9aa3c4afa80d69088ffaa2402648f4a329e0a798a8c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://544885302dffff8336a9aa3c4afa80d69088ffaa2402648f4a329e0a798a8c99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8bc193fb3f85b980c31af2e826cee90adb633ad2ee82ff1d46e182ab52e8fab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8bc193fb3f85b980c31af2e826cee90adb633ad2ee82ff1d46e182ab52e8fab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7713e2a35bac8df1e8987fbcd452459fa3541caad0c52ed57fda187bd7c83d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7713e2a35bac8df1e8987fbcd452459fa3541caad0c52ed57fda187bd7c83d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34adfca89026fc20007a73a8cbd4e7c7b3137a581c3476867af17fabe976f71f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34adfca89026fc20007a73a8cbd4e7c7b3137a581c3476867af17fabe976f71f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qppnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:57Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:57 crc kubenswrapper[4821]: I1125 10:32:57.545958 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tgdlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2336955b-a703-4db4-8d0f-d08088ed59b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb0362942bfac635ee6292501b1b851772c01d75dd658eef4b17669a716de72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vjfp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tgdlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:57Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:57 crc kubenswrapper[4821]: I1125 10:32:57.562649 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62fcb632-f12a-46e2-bb7d-a5311f291ee7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28d43531927e75c3745f100940f7f6bf79fd21026c1309194df8cd7a39361b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69633332d03d2551ff4abb253d358de398eab18486b92db630ce06c0447fb9cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ccb78d9a51ac83ee208c2451685a2512de9deaf742a663bb8c892aa59b76162\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27826e155a9352b1e1866e83fdab683dbcd749cc04b6a4da2db25ecabe3d9a52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://991252e5f56f128827a5054b8e228284bc0266166582cbaedaadc53ca2f5b401\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:57Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:57 crc kubenswrapper[4821]: I1125 10:32:57.577424 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06f949936bac99cfbf89d720c9549c074351d9077bd963fa024d87004c6be23e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:57Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:57 crc kubenswrapper[4821]: I1125 10:32:57.588596 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384d8ba58264ab89796d3b53afb0c4c1be33cb4fba719a8beef4afc2be7d66a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:57Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:57 crc kubenswrapper[4821]: I1125 10:32:57.601033 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zlk96" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d55aa91-9cb9-4c95-bc27-6fec05684c10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4521b5baee45ffc9c7661c781c0b3de458dff325ad525dde283d46af71c0cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vkd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zlk96\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:57Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:57 crc kubenswrapper[4821]: I1125 10:32:57.613978 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6hbqd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f71b3a2-63cb-443c-9f41-39e57030746c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65473b0e63c2e9c7c09a2bb8f31396daf1d87c2c83d20ed3508ce2fe7358dda3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxdmq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://595f9ea75461f18e763d84799e368becc07b72acdc5337d20d6995bf53c3d27a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxdmq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6hbqd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:57Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:57 crc kubenswrapper[4821]: I1125 10:32:57.615326 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:57 crc kubenswrapper[4821]: I1125 10:32:57.615365 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:57 crc kubenswrapper[4821]: I1125 10:32:57.615382 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:57 crc kubenswrapper[4821]: I1125 10:32:57.615404 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:57 crc kubenswrapper[4821]: I1125 10:32:57.615416 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:57Z","lastTransitionTime":"2025-11-25T10:32:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:57 crc kubenswrapper[4821]: I1125 10:32:57.629746 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:57Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:57 crc kubenswrapper[4821]: I1125 10:32:57.645406 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17b8714be3bd121e06641873e003f08560b1f71e56d6ea936966df7d4896eb03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85731e6b6dbe1777bc7161a010f39f79523ec4b08d71d5d8a975573aa7fe4957\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:57Z is after 2025-08-24T17:21:41Z" Nov 25 10:32:57 crc kubenswrapper[4821]: I1125 10:32:57.717702 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:57 crc kubenswrapper[4821]: I1125 10:32:57.717752 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:57 crc kubenswrapper[4821]: I1125 10:32:57.717763 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:57 crc kubenswrapper[4821]: I1125 10:32:57.717780 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:57 crc kubenswrapper[4821]: I1125 10:32:57.717792 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:57Z","lastTransitionTime":"2025-11-25T10:32:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:57 crc kubenswrapper[4821]: I1125 10:32:57.819875 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:57 crc kubenswrapper[4821]: I1125 10:32:57.819915 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:57 crc kubenswrapper[4821]: I1125 10:32:57.819933 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:57 crc kubenswrapper[4821]: I1125 10:32:57.819951 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:57 crc kubenswrapper[4821]: I1125 10:32:57.819964 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:57Z","lastTransitionTime":"2025-11-25T10:32:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:57 crc kubenswrapper[4821]: I1125 10:32:57.921722 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:57 crc kubenswrapper[4821]: I1125 10:32:57.921757 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:57 crc kubenswrapper[4821]: I1125 10:32:57.921765 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:57 crc kubenswrapper[4821]: I1125 10:32:57.921779 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:57 crc kubenswrapper[4821]: I1125 10:32:57.921787 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:57Z","lastTransitionTime":"2025-11-25T10:32:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:58 crc kubenswrapper[4821]: I1125 10:32:58.023704 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:58 crc kubenswrapper[4821]: I1125 10:32:58.023744 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:58 crc kubenswrapper[4821]: I1125 10:32:58.023752 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:58 crc kubenswrapper[4821]: I1125 10:32:58.023766 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:58 crc kubenswrapper[4821]: I1125 10:32:58.023775 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:58Z","lastTransitionTime":"2025-11-25T10:32:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:58 crc kubenswrapper[4821]: I1125 10:32:58.114221 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4msn" Nov 25 10:32:58 crc kubenswrapper[4821]: E1125 10:32:58.114368 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4msn" podUID="1f12cef3-7e40-4f61-836e-23cd0e578e71" Nov 25 10:32:58 crc kubenswrapper[4821]: I1125 10:32:58.125927 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:58 crc kubenswrapper[4821]: I1125 10:32:58.125963 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:58 crc kubenswrapper[4821]: I1125 10:32:58.125971 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:58 crc kubenswrapper[4821]: I1125 10:32:58.125983 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:58 crc kubenswrapper[4821]: I1125 10:32:58.125994 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:58Z","lastTransitionTime":"2025-11-25T10:32:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:58 crc kubenswrapper[4821]: I1125 10:32:58.228340 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:58 crc kubenswrapper[4821]: I1125 10:32:58.228387 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:58 crc kubenswrapper[4821]: I1125 10:32:58.228399 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:58 crc kubenswrapper[4821]: I1125 10:32:58.228417 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:58 crc kubenswrapper[4821]: I1125 10:32:58.228428 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:58Z","lastTransitionTime":"2025-11-25T10:32:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:58 crc kubenswrapper[4821]: I1125 10:32:58.330803 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:58 crc kubenswrapper[4821]: I1125 10:32:58.330841 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:58 crc kubenswrapper[4821]: I1125 10:32:58.330852 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:58 crc kubenswrapper[4821]: I1125 10:32:58.330867 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:58 crc kubenswrapper[4821]: I1125 10:32:58.330904 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:58Z","lastTransitionTime":"2025-11-25T10:32:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:58 crc kubenswrapper[4821]: I1125 10:32:58.432585 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:58 crc kubenswrapper[4821]: I1125 10:32:58.432620 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:58 crc kubenswrapper[4821]: I1125 10:32:58.432630 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:58 crc kubenswrapper[4821]: I1125 10:32:58.432647 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:58 crc kubenswrapper[4821]: I1125 10:32:58.432658 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:58Z","lastTransitionTime":"2025-11-25T10:32:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:58 crc kubenswrapper[4821]: I1125 10:32:58.535289 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:58 crc kubenswrapper[4821]: I1125 10:32:58.535640 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:58 crc kubenswrapper[4821]: I1125 10:32:58.535844 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:58 crc kubenswrapper[4821]: I1125 10:32:58.536072 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:58 crc kubenswrapper[4821]: I1125 10:32:58.536313 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:58Z","lastTransitionTime":"2025-11-25T10:32:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:58 crc kubenswrapper[4821]: I1125 10:32:58.639960 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:58 crc kubenswrapper[4821]: I1125 10:32:58.639999 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:58 crc kubenswrapper[4821]: I1125 10:32:58.640010 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:58 crc kubenswrapper[4821]: I1125 10:32:58.640027 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:58 crc kubenswrapper[4821]: I1125 10:32:58.640038 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:58Z","lastTransitionTime":"2025-11-25T10:32:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:58 crc kubenswrapper[4821]: I1125 10:32:58.741844 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:58 crc kubenswrapper[4821]: I1125 10:32:58.741884 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:58 crc kubenswrapper[4821]: I1125 10:32:58.741895 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:58 crc kubenswrapper[4821]: I1125 10:32:58.741915 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:58 crc kubenswrapper[4821]: I1125 10:32:58.741927 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:58Z","lastTransitionTime":"2025-11-25T10:32:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:58 crc kubenswrapper[4821]: I1125 10:32:58.844111 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:58 crc kubenswrapper[4821]: I1125 10:32:58.844151 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:58 crc kubenswrapper[4821]: I1125 10:32:58.844172 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:58 crc kubenswrapper[4821]: I1125 10:32:58.844187 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:58 crc kubenswrapper[4821]: I1125 10:32:58.844202 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:58Z","lastTransitionTime":"2025-11-25T10:32:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:58 crc kubenswrapper[4821]: I1125 10:32:58.946606 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:58 crc kubenswrapper[4821]: I1125 10:32:58.946663 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:58 crc kubenswrapper[4821]: I1125 10:32:58.946680 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:58 crc kubenswrapper[4821]: I1125 10:32:58.946701 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:58 crc kubenswrapper[4821]: I1125 10:32:58.946714 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:58Z","lastTransitionTime":"2025-11-25T10:32:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:59 crc kubenswrapper[4821]: I1125 10:32:59.048989 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:59 crc kubenswrapper[4821]: I1125 10:32:59.049024 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:59 crc kubenswrapper[4821]: I1125 10:32:59.049032 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:59 crc kubenswrapper[4821]: I1125 10:32:59.049045 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:59 crc kubenswrapper[4821]: I1125 10:32:59.049055 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:59Z","lastTransitionTime":"2025-11-25T10:32:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:59 crc kubenswrapper[4821]: I1125 10:32:59.113450 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:32:59 crc kubenswrapper[4821]: I1125 10:32:59.113557 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:32:59 crc kubenswrapper[4821]: I1125 10:32:59.113616 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:32:59 crc kubenswrapper[4821]: E1125 10:32:59.113685 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:32:59 crc kubenswrapper[4821]: E1125 10:32:59.113845 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:32:59 crc kubenswrapper[4821]: E1125 10:32:59.113929 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:32:59 crc kubenswrapper[4821]: I1125 10:32:59.152194 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:59 crc kubenswrapper[4821]: I1125 10:32:59.152239 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:59 crc kubenswrapper[4821]: I1125 10:32:59.152251 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:59 crc kubenswrapper[4821]: I1125 10:32:59.152267 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:59 crc kubenswrapper[4821]: I1125 10:32:59.152279 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:59Z","lastTransitionTime":"2025-11-25T10:32:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:59 crc kubenswrapper[4821]: I1125 10:32:59.254814 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:59 crc kubenswrapper[4821]: I1125 10:32:59.254870 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:59 crc kubenswrapper[4821]: I1125 10:32:59.254881 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:59 crc kubenswrapper[4821]: I1125 10:32:59.254900 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:59 crc kubenswrapper[4821]: I1125 10:32:59.254908 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:59Z","lastTransitionTime":"2025-11-25T10:32:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:59 crc kubenswrapper[4821]: I1125 10:32:59.357414 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:59 crc kubenswrapper[4821]: I1125 10:32:59.357464 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:59 crc kubenswrapper[4821]: I1125 10:32:59.357474 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:59 crc kubenswrapper[4821]: I1125 10:32:59.357489 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:59 crc kubenswrapper[4821]: I1125 10:32:59.357499 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:59Z","lastTransitionTime":"2025-11-25T10:32:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:59 crc kubenswrapper[4821]: I1125 10:32:59.459670 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:59 crc kubenswrapper[4821]: I1125 10:32:59.459714 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:59 crc kubenswrapper[4821]: I1125 10:32:59.459725 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:59 crc kubenswrapper[4821]: I1125 10:32:59.459742 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:59 crc kubenswrapper[4821]: I1125 10:32:59.459753 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:59Z","lastTransitionTime":"2025-11-25T10:32:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:59 crc kubenswrapper[4821]: I1125 10:32:59.563094 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:59 crc kubenswrapper[4821]: I1125 10:32:59.563133 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:59 crc kubenswrapper[4821]: I1125 10:32:59.563142 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:59 crc kubenswrapper[4821]: I1125 10:32:59.563187 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:59 crc kubenswrapper[4821]: I1125 10:32:59.563206 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:59Z","lastTransitionTime":"2025-11-25T10:32:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:59 crc kubenswrapper[4821]: I1125 10:32:59.665716 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:59 crc kubenswrapper[4821]: I1125 10:32:59.665755 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:59 crc kubenswrapper[4821]: I1125 10:32:59.665765 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:59 crc kubenswrapper[4821]: I1125 10:32:59.665781 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:59 crc kubenswrapper[4821]: I1125 10:32:59.665791 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:59Z","lastTransitionTime":"2025-11-25T10:32:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:59 crc kubenswrapper[4821]: I1125 10:32:59.771274 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:59 crc kubenswrapper[4821]: I1125 10:32:59.771321 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:59 crc kubenswrapper[4821]: I1125 10:32:59.771329 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:59 crc kubenswrapper[4821]: I1125 10:32:59.771343 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:59 crc kubenswrapper[4821]: I1125 10:32:59.771352 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:59Z","lastTransitionTime":"2025-11-25T10:32:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:59 crc kubenswrapper[4821]: I1125 10:32:59.874305 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:59 crc kubenswrapper[4821]: I1125 10:32:59.874374 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:59 crc kubenswrapper[4821]: I1125 10:32:59.874388 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:59 crc kubenswrapper[4821]: I1125 10:32:59.874410 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:59 crc kubenswrapper[4821]: I1125 10:32:59.874423 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:59Z","lastTransitionTime":"2025-11-25T10:32:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:32:59 crc kubenswrapper[4821]: I1125 10:32:59.977788 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:32:59 crc kubenswrapper[4821]: I1125 10:32:59.977824 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:32:59 crc kubenswrapper[4821]: I1125 10:32:59.977836 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:32:59 crc kubenswrapper[4821]: I1125 10:32:59.977854 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:32:59 crc kubenswrapper[4821]: I1125 10:32:59.977866 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:32:59Z","lastTransitionTime":"2025-11-25T10:32:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:00 crc kubenswrapper[4821]: I1125 10:33:00.080138 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:00 crc kubenswrapper[4821]: I1125 10:33:00.080191 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:00 crc kubenswrapper[4821]: I1125 10:33:00.080204 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:00 crc kubenswrapper[4821]: I1125 10:33:00.080217 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:00 crc kubenswrapper[4821]: I1125 10:33:00.080226 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:00Z","lastTransitionTime":"2025-11-25T10:33:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:00 crc kubenswrapper[4821]: I1125 10:33:00.113997 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4msn" Nov 25 10:33:00 crc kubenswrapper[4821]: E1125 10:33:00.114148 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4msn" podUID="1f12cef3-7e40-4f61-836e-23cd0e578e71" Nov 25 10:33:00 crc kubenswrapper[4821]: I1125 10:33:00.131660 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62fcb632-f12a-46e2-bb7d-a5311f291ee7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28d43531927e75c3745f100940f7f6bf79fd21026c1309194df8cd7a39361b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69633332d03d2551ff4abb253d358de398eab18486b92db630ce06c0447fb9cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ccb78d9a51ac83ee208c2451685a2512de9deaf742a663bb8c892aa59b76162\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27826e155a9352b1e1866e83fdab683dbcd749cc04b6a4da2db25ecabe3d9a52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://991252e5f56f128827a5054b8e228284bc0266166582cbaedaadc53ca2f5b401\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:00Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:00 crc kubenswrapper[4821]: I1125 10:33:00.143305 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63b471fd-5e16-4761-8b07-44a135d5f5f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a68a57fa7fba8f351f264e427a50bed23f57775644e9c268aa7305efb6916794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67df7c609d5a5d658be4c3e873ec3052bd070ad9469f8ff44c29a9571e3e5d95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://269ced131bc0777dc6752070470c95b4bedfe4b4240940eabd769500ddef0404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e48c0e3b58359c6c0a493d58f7d2aec6011747791a3c0d7788102e1fec18865\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e48c0e3b58359c6c0a493d58f7d2aec6011747791a3c0d7788102e1fec18865\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:00Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:00 crc kubenswrapper[4821]: I1125 10:33:00.156606 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:00Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:00 crc kubenswrapper[4821]: I1125 10:33:00.167367 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f948b87-ac86-4de6-ad64-c2ef947f84d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4482f73a45bc82bd55770c3fd788becacec94893fed1c305fed1c81c753d43aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ba68bd093be94d0d40f3851638db5d26bdfc52d175b1ac78afb91dda7affcfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2krbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:00Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:00 crc kubenswrapper[4821]: I1125 10:33:00.183172 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:00 crc kubenswrapper[4821]: I1125 10:33:00.183212 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:00 crc kubenswrapper[4821]: I1125 10:33:00.183224 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:00 crc kubenswrapper[4821]: I1125 10:33:00.183240 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:00 crc kubenswrapper[4821]: I1125 10:33:00.182091 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qppnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"357491b8-cd9c-486e-9898-5ba33678513d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc88d700ca3a449ef1cb32d6124b16d0024d6981ced12e276083f86a1b322ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd9c047c5ce1d896f803ff6cafa9bbf3a18a391b44539e1f1312e08c89e51cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd9c047c5ce1d896f803ff6cafa9bbf3a18a391b44539e1f1312e08c89e51cb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://565a6db63fb69abbba81ffdf6085797900858d263cd6affcb0b5f195474c2516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://565a6db63fb69abbba81ffdf6085797900858d263cd6affcb0b5f195474c2516\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://544885302dffff8336a9aa3c4afa80d69088ffaa2402648f4a329e0a798a8c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://544885302dffff8336a9aa3c4afa80d69088ffaa2402648f4a329e0a798a8c99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8bc193fb3f85b980c31af2e826cee90adb633ad2ee82ff1d46e182ab52e8fab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8bc193fb3f85b980c31af2e826cee90adb633ad2ee82ff1d46e182ab52e8fab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7713e2a35bac8df1e8987fbcd452459fa3541caad0c52ed57fda187bd7c83d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7713e2a35bac8df1e8987fbcd452459fa3541caad0c52ed57fda187bd7c83d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34adfca89026fc20007a73a8cbd4e7c7b3137a581c3476867af17fabe976f71f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34adfca89026fc20007a73a8cbd4e7c7b3137a581c3476867af17fabe976f71f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qppnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:00Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:00 crc kubenswrapper[4821]: I1125 10:33:00.183252 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:00Z","lastTransitionTime":"2025-11-25T10:33:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:00 crc kubenswrapper[4821]: I1125 10:33:00.194522 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tgdlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2336955b-a703-4db4-8d0f-d08088ed59b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb0362942bfac635ee6292501b1b851772c01d75dd658eef4b17669a716de72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vjfp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tgdlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:00Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:00 crc kubenswrapper[4821]: I1125 10:33:00.207956 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06f949936bac99cfbf89d720c9549c074351d9077bd963fa024d87004c6be23e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:00Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:00 crc kubenswrapper[4821]: I1125 10:33:00.219496 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384d8ba58264ab89796d3b53afb0c4c1be33cb4fba719a8beef4afc2be7d66a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:00Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:00 crc kubenswrapper[4821]: I1125 10:33:00.232257 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:00Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:00 crc kubenswrapper[4821]: I1125 10:33:00.246549 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17b8714be3bd121e06641873e003f08560b1f71e56d6ea936966df7d4896eb03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85731e6b6dbe1777bc7161a010f39f79523ec4b08d71d5d8a975573aa7fe4957\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:00Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:00 crc kubenswrapper[4821]: I1125 10:33:00.255271 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zlk96" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d55aa91-9cb9-4c95-bc27-6fec05684c10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4521b5baee45ffc9c7661c781c0b3de458dff325ad525dde283d46af71c0cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vkd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zlk96\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:00Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:00 crc kubenswrapper[4821]: I1125 10:33:00.265439 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6hbqd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f71b3a2-63cb-443c-9f41-39e57030746c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65473b0e63c2e9c7c09a2bb8f31396daf1d87c2c83d20ed3508ce2fe7358dda3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxdmq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://595f9ea75461f18e763d84799e368becc07b72acdc5337d20d6995bf53c3d27a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxdmq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6hbqd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:00Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:00 crc kubenswrapper[4821]: I1125 10:33:00.280681 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1f12cef3-7e40-4f61-836e-23cd0e578e71-metrics-certs\") pod \"network-metrics-daemon-d4msn\" (UID: \"1f12cef3-7e40-4f61-836e-23cd0e578e71\") " pod="openshift-multus/network-metrics-daemon-d4msn" Nov 25 10:33:00 crc kubenswrapper[4821]: E1125 10:33:00.280841 4821 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 10:33:00 crc kubenswrapper[4821]: E1125 10:33:00.280892 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1f12cef3-7e40-4f61-836e-23cd0e578e71-metrics-certs podName:1f12cef3-7e40-4f61-836e-23cd0e578e71 nodeName:}" failed. No retries permitted until 2025-11-25 10:33:16.28087481 +0000 UTC m=+66.817194657 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/1f12cef3-7e40-4f61-836e-23cd0e578e71-metrics-certs") pod "network-metrics-daemon-d4msn" (UID: "1f12cef3-7e40-4f61-836e-23cd0e578e71") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 10:33:00 crc kubenswrapper[4821]: I1125 10:33:00.281938 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d9e58df-1755-4422-8d04-211f98c7c011\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea762f627aa69fb88a9bd82545fa677e9ccfa97f99a5e92c0aa490e619c0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a06b811e5ec52ce984a2403ba2c98332ba60466d08b3580ad9192fac6caef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0385f6ed9597974267f7c45547f6b6725b4ee83a2cc0508ff2d1c1ce8753ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b915b565f7fbb3e4e856e7f1035a519e82bfb3023660c121f6e3813de353bb92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18465f3a6916c36457a305ba5d568c9077f34f1d000532ed518f237d100008b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:00Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:00 crc kubenswrapper[4821]: I1125 10:33:00.285541 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:00 crc kubenswrapper[4821]: I1125 10:33:00.285566 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:00 crc kubenswrapper[4821]: I1125 10:33:00.285575 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:00 crc kubenswrapper[4821]: I1125 10:33:00.285589 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:00 crc kubenswrapper[4821]: I1125 10:33:00.285598 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:00Z","lastTransitionTime":"2025-11-25T10:33:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:00 crc kubenswrapper[4821]: I1125 10:33:00.292302 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25595ccd-762d-414a-9916-4fe38d478b07\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d4b2d8db4d13392d9789e161fd4b8f6eef983b79128ee584d72e4c06565bc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf780f791771ea43a8c8176d02fdfb5204406e023207299369ae46ceb342baaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0251519fd4262f2ea5c64959db49433afdafe36b7ed2404d660eeb7c365d2746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad5c205e0eeecfd43cbdf0eb1a334e3e8faa5e9a9e8dd202c397d5cbb36e081e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:00Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:00 crc kubenswrapper[4821]: I1125 10:33:00.302379 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:00Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:00 crc kubenswrapper[4821]: I1125 10:33:00.316952 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ldzbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b84e32ae-4897-4644-a656-6bd82012cc81\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ef433744761f26e8e1f6c0db123821a11923ffd774830744285cd4178d89533\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5pmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ldzbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:00Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:00 crc kubenswrapper[4821]: I1125 10:33:00.335299 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"954085fa-d03a-4e88-9dfd-566257e3558d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f00746930c4588e548f0e2b60d52bf448aff7401b676a7855cb0f4cce00d329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a14a078ba7d2aed95992b36502008f315f0d3a634b64ed49da70398b2562d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d7254196b3702e48083ae98e829b73c539adf69ed5b2bbdc62189d09ae1762\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d727430108977574365f0e9c04b44e03307bd7ff89cdbb1e48a6e916f504d6b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94c7949f077b1ca2b310bedb090aa72fc4bd7e8f2c8bf438096edfca8568f36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed2e7cad7bb24b5b56a1295098a6a7c230c8f9b9c29a88ef653ae8b6ac31667d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e6202b51ac13209f3a6de880638a6d142f79755311646c0216f11dfe8bbe3f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a84f30b6a8ad3333268312e9aadacb17eca12613511195c6f8a7734d5f7af5b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T10:32:41Z\\\",\\\"message\\\":\\\" 6233 obj_retry.go:365] Adding new object: *v1.Pod openshift-dns/node-resolver-zlk96\\\\nI1125 10:32:41.011897 6233 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-target-xd92c in node crc\\\\nI1125 10:32:41.011901 6233 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-additional-cni-plugins-qppnw after 0 failed attempt(s)\\\\nI1125 10:32:41.011815 6233 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1125 10:32:41.011915 6233 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-additional-cni-plugins-qppnw\\\\nI1125 10:32:41.011902 6233 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-zlk96 in node crc\\\\nI1125 10:32:41.011826 6233 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1125 10:32:41.011923 6233 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-tgdlt after 0 failed attempt(s)\\\\nI1125 10:32:41.011884 6233 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-ingress-operator/metrics]} name:Service_openshift-ingress-operator/metrics_\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e6202b51ac13209f3a6de880638a6d142f79755311646c0216f11dfe8bbe3f7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T10:32:56Z\\\",\\\"message\\\":\\\"to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:56Z is after 2025-08-24T17:21:41Z]\\\\nI1125 10:32:56.048189 6453 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-console/console_TCP_cluster\\\\\\\", UUID:\\\\\\\"d7d7b270-1480-47f8-bdf9-690dbab310cb\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console/console\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]strin\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dbf499f4ae6161752abaa7a466a027c6f874016be90a534e69f07e3dd3f3d9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzhnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:00Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:00 crc kubenswrapper[4821]: I1125 10:33:00.349557 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-d4msn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f12cef3-7e40-4f61-836e-23cd0e578e71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s7lkv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s7lkv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:44Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-d4msn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:00Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:00 crc kubenswrapper[4821]: I1125 10:33:00.387765 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:00 crc kubenswrapper[4821]: I1125 10:33:00.387809 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:00 crc kubenswrapper[4821]: I1125 10:33:00.387823 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:00 crc kubenswrapper[4821]: I1125 10:33:00.387844 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:00 crc kubenswrapper[4821]: I1125 10:33:00.387856 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:00Z","lastTransitionTime":"2025-11-25T10:33:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:00 crc kubenswrapper[4821]: I1125 10:33:00.490276 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:00 crc kubenswrapper[4821]: I1125 10:33:00.490585 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:00 crc kubenswrapper[4821]: I1125 10:33:00.490650 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:00 crc kubenswrapper[4821]: I1125 10:33:00.490725 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:00 crc kubenswrapper[4821]: I1125 10:33:00.490779 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:00Z","lastTransitionTime":"2025-11-25T10:33:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:00 crc kubenswrapper[4821]: I1125 10:33:00.593594 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:00 crc kubenswrapper[4821]: I1125 10:33:00.593858 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:00 crc kubenswrapper[4821]: I1125 10:33:00.593895 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:00 crc kubenswrapper[4821]: I1125 10:33:00.593929 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:00 crc kubenswrapper[4821]: I1125 10:33:00.593951 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:00Z","lastTransitionTime":"2025-11-25T10:33:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:00 crc kubenswrapper[4821]: I1125 10:33:00.696808 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:00 crc kubenswrapper[4821]: I1125 10:33:00.696856 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:00 crc kubenswrapper[4821]: I1125 10:33:00.696869 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:00 crc kubenswrapper[4821]: I1125 10:33:00.696887 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:00 crc kubenswrapper[4821]: I1125 10:33:00.696900 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:00Z","lastTransitionTime":"2025-11-25T10:33:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:00 crc kubenswrapper[4821]: I1125 10:33:00.798879 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:00 crc kubenswrapper[4821]: I1125 10:33:00.799615 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:00 crc kubenswrapper[4821]: I1125 10:33:00.799681 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:00 crc kubenswrapper[4821]: I1125 10:33:00.799742 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:00 crc kubenswrapper[4821]: I1125 10:33:00.799818 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:00Z","lastTransitionTime":"2025-11-25T10:33:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:00 crc kubenswrapper[4821]: I1125 10:33:00.902989 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:00 crc kubenswrapper[4821]: I1125 10:33:00.903036 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:00 crc kubenswrapper[4821]: I1125 10:33:00.903047 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:00 crc kubenswrapper[4821]: I1125 10:33:00.903065 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:00 crc kubenswrapper[4821]: I1125 10:33:00.903078 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:00Z","lastTransitionTime":"2025-11-25T10:33:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:00 crc kubenswrapper[4821]: I1125 10:33:00.989246 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:33:00 crc kubenswrapper[4821]: E1125 10:33:00.989432 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:33:32.989399282 +0000 UTC m=+83.525719139 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:33:00 crc kubenswrapper[4821]: I1125 10:33:00.989664 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:33:00 crc kubenswrapper[4821]: I1125 10:33:00.989745 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:33:00 crc kubenswrapper[4821]: E1125 10:33:00.989842 4821 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 10:33:00 crc kubenswrapper[4821]: E1125 10:33:00.989866 4821 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 10:33:00 crc kubenswrapper[4821]: E1125 10:33:00.989922 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 10:33:32.989907583 +0000 UTC m=+83.526227500 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 10:33:00 crc kubenswrapper[4821]: E1125 10:33:00.989954 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 10:33:32.989936784 +0000 UTC m=+83.526256631 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 10:33:01 crc kubenswrapper[4821]: I1125 10:33:01.005732 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:01 crc kubenswrapper[4821]: I1125 10:33:01.005761 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:01 crc kubenswrapper[4821]: I1125 10:33:01.005769 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:01 crc kubenswrapper[4821]: I1125 10:33:01.005781 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:01 crc kubenswrapper[4821]: I1125 10:33:01.005790 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:01Z","lastTransitionTime":"2025-11-25T10:33:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:01 crc kubenswrapper[4821]: I1125 10:33:01.090921 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:33:01 crc kubenswrapper[4821]: I1125 10:33:01.090959 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:33:01 crc kubenswrapper[4821]: E1125 10:33:01.091086 4821 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 10:33:01 crc kubenswrapper[4821]: E1125 10:33:01.091101 4821 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 10:33:01 crc kubenswrapper[4821]: E1125 10:33:01.091112 4821 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 10:33:01 crc kubenswrapper[4821]: E1125 10:33:01.091149 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 10:33:33.091137145 +0000 UTC m=+83.627456992 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 10:33:01 crc kubenswrapper[4821]: E1125 10:33:01.091456 4821 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 10:33:01 crc kubenswrapper[4821]: E1125 10:33:01.091468 4821 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 10:33:01 crc kubenswrapper[4821]: E1125 10:33:01.091475 4821 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 10:33:01 crc kubenswrapper[4821]: E1125 10:33:01.091497 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 10:33:33.091490023 +0000 UTC m=+83.627809860 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 10:33:01 crc kubenswrapper[4821]: I1125 10:33:01.108910 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:01 crc kubenswrapper[4821]: I1125 10:33:01.109064 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:01 crc kubenswrapper[4821]: I1125 10:33:01.109097 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:01 crc kubenswrapper[4821]: I1125 10:33:01.109123 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:01 crc kubenswrapper[4821]: I1125 10:33:01.109139 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:01Z","lastTransitionTime":"2025-11-25T10:33:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:01 crc kubenswrapper[4821]: I1125 10:33:01.113434 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:33:01 crc kubenswrapper[4821]: E1125 10:33:01.113532 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:33:01 crc kubenswrapper[4821]: I1125 10:33:01.113621 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:33:01 crc kubenswrapper[4821]: E1125 10:33:01.113780 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:33:01 crc kubenswrapper[4821]: I1125 10:33:01.114041 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:33:01 crc kubenswrapper[4821]: E1125 10:33:01.114347 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:33:01 crc kubenswrapper[4821]: I1125 10:33:01.211449 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:01 crc kubenswrapper[4821]: I1125 10:33:01.211481 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:01 crc kubenswrapper[4821]: I1125 10:33:01.211490 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:01 crc kubenswrapper[4821]: I1125 10:33:01.211503 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:01 crc kubenswrapper[4821]: I1125 10:33:01.211515 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:01Z","lastTransitionTime":"2025-11-25T10:33:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:01 crc kubenswrapper[4821]: I1125 10:33:01.313713 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:01 crc kubenswrapper[4821]: I1125 10:33:01.313755 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:01 crc kubenswrapper[4821]: I1125 10:33:01.313764 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:01 crc kubenswrapper[4821]: I1125 10:33:01.313777 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:01 crc kubenswrapper[4821]: I1125 10:33:01.313788 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:01Z","lastTransitionTime":"2025-11-25T10:33:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:01 crc kubenswrapper[4821]: I1125 10:33:01.416393 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:01 crc kubenswrapper[4821]: I1125 10:33:01.416439 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:01 crc kubenswrapper[4821]: I1125 10:33:01.416448 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:01 crc kubenswrapper[4821]: I1125 10:33:01.416463 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:01 crc kubenswrapper[4821]: I1125 10:33:01.416472 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:01Z","lastTransitionTime":"2025-11-25T10:33:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:01 crc kubenswrapper[4821]: I1125 10:33:01.519271 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:01 crc kubenswrapper[4821]: I1125 10:33:01.519364 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:01 crc kubenswrapper[4821]: I1125 10:33:01.519377 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:01 crc kubenswrapper[4821]: I1125 10:33:01.519395 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:01 crc kubenswrapper[4821]: I1125 10:33:01.519408 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:01Z","lastTransitionTime":"2025-11-25T10:33:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:01 crc kubenswrapper[4821]: I1125 10:33:01.621921 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:01 crc kubenswrapper[4821]: I1125 10:33:01.621955 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:01 crc kubenswrapper[4821]: I1125 10:33:01.621963 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:01 crc kubenswrapper[4821]: I1125 10:33:01.621978 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:01 crc kubenswrapper[4821]: I1125 10:33:01.621987 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:01Z","lastTransitionTime":"2025-11-25T10:33:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:01 crc kubenswrapper[4821]: I1125 10:33:01.723633 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:01 crc kubenswrapper[4821]: I1125 10:33:01.723691 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:01 crc kubenswrapper[4821]: I1125 10:33:01.723700 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:01 crc kubenswrapper[4821]: I1125 10:33:01.723715 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:01 crc kubenswrapper[4821]: I1125 10:33:01.723723 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:01Z","lastTransitionTime":"2025-11-25T10:33:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:01 crc kubenswrapper[4821]: I1125 10:33:01.825589 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:01 crc kubenswrapper[4821]: I1125 10:33:01.825631 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:01 crc kubenswrapper[4821]: I1125 10:33:01.825639 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:01 crc kubenswrapper[4821]: I1125 10:33:01.825654 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:01 crc kubenswrapper[4821]: I1125 10:33:01.825669 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:01Z","lastTransitionTime":"2025-11-25T10:33:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:01 crc kubenswrapper[4821]: I1125 10:33:01.927884 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:01 crc kubenswrapper[4821]: I1125 10:33:01.927920 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:01 crc kubenswrapper[4821]: I1125 10:33:01.927929 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:01 crc kubenswrapper[4821]: I1125 10:33:01.927943 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:01 crc kubenswrapper[4821]: I1125 10:33:01.927957 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:01Z","lastTransitionTime":"2025-11-25T10:33:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:02 crc kubenswrapper[4821]: I1125 10:33:02.030260 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:02 crc kubenswrapper[4821]: I1125 10:33:02.030305 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:02 crc kubenswrapper[4821]: I1125 10:33:02.030318 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:02 crc kubenswrapper[4821]: I1125 10:33:02.030335 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:02 crc kubenswrapper[4821]: I1125 10:33:02.030344 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:02Z","lastTransitionTime":"2025-11-25T10:33:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:02 crc kubenswrapper[4821]: I1125 10:33:02.113894 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4msn" Nov 25 10:33:02 crc kubenswrapper[4821]: E1125 10:33:02.114137 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4msn" podUID="1f12cef3-7e40-4f61-836e-23cd0e578e71" Nov 25 10:33:02 crc kubenswrapper[4821]: I1125 10:33:02.133790 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:02 crc kubenswrapper[4821]: I1125 10:33:02.133861 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:02 crc kubenswrapper[4821]: I1125 10:33:02.133879 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:02 crc kubenswrapper[4821]: I1125 10:33:02.133906 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:02 crc kubenswrapper[4821]: I1125 10:33:02.133924 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:02Z","lastTransitionTime":"2025-11-25T10:33:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:02 crc kubenswrapper[4821]: I1125 10:33:02.235715 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:02 crc kubenswrapper[4821]: I1125 10:33:02.235757 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:02 crc kubenswrapper[4821]: I1125 10:33:02.235769 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:02 crc kubenswrapper[4821]: I1125 10:33:02.235787 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:02 crc kubenswrapper[4821]: I1125 10:33:02.235925 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:02Z","lastTransitionTime":"2025-11-25T10:33:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:02 crc kubenswrapper[4821]: I1125 10:33:02.338643 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:02 crc kubenswrapper[4821]: I1125 10:33:02.338682 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:02 crc kubenswrapper[4821]: I1125 10:33:02.338693 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:02 crc kubenswrapper[4821]: I1125 10:33:02.338764 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:02 crc kubenswrapper[4821]: I1125 10:33:02.338846 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:02Z","lastTransitionTime":"2025-11-25T10:33:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:02 crc kubenswrapper[4821]: I1125 10:33:02.440912 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:02 crc kubenswrapper[4821]: I1125 10:33:02.440946 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:02 crc kubenswrapper[4821]: I1125 10:33:02.440955 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:02 crc kubenswrapper[4821]: I1125 10:33:02.440973 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:02 crc kubenswrapper[4821]: I1125 10:33:02.440983 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:02Z","lastTransitionTime":"2025-11-25T10:33:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:02 crc kubenswrapper[4821]: I1125 10:33:02.543071 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:02 crc kubenswrapper[4821]: I1125 10:33:02.543101 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:02 crc kubenswrapper[4821]: I1125 10:33:02.543109 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:02 crc kubenswrapper[4821]: I1125 10:33:02.543122 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:02 crc kubenswrapper[4821]: I1125 10:33:02.543130 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:02Z","lastTransitionTime":"2025-11-25T10:33:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:02 crc kubenswrapper[4821]: I1125 10:33:02.645148 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:02 crc kubenswrapper[4821]: I1125 10:33:02.645208 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:02 crc kubenswrapper[4821]: I1125 10:33:02.645217 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:02 crc kubenswrapper[4821]: I1125 10:33:02.645229 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:02 crc kubenswrapper[4821]: I1125 10:33:02.645238 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:02Z","lastTransitionTime":"2025-11-25T10:33:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:02 crc kubenswrapper[4821]: I1125 10:33:02.747061 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:02 crc kubenswrapper[4821]: I1125 10:33:02.747109 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:02 crc kubenswrapper[4821]: I1125 10:33:02.747118 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:02 crc kubenswrapper[4821]: I1125 10:33:02.747133 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:02 crc kubenswrapper[4821]: I1125 10:33:02.747145 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:02Z","lastTransitionTime":"2025-11-25T10:33:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:02 crc kubenswrapper[4821]: I1125 10:33:02.850082 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:02 crc kubenswrapper[4821]: I1125 10:33:02.850176 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:02 crc kubenswrapper[4821]: I1125 10:33:02.850197 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:02 crc kubenswrapper[4821]: I1125 10:33:02.850214 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:02 crc kubenswrapper[4821]: I1125 10:33:02.850225 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:02Z","lastTransitionTime":"2025-11-25T10:33:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:02 crc kubenswrapper[4821]: I1125 10:33:02.952468 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:02 crc kubenswrapper[4821]: I1125 10:33:02.952505 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:02 crc kubenswrapper[4821]: I1125 10:33:02.952514 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:02 crc kubenswrapper[4821]: I1125 10:33:02.952528 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:02 crc kubenswrapper[4821]: I1125 10:33:02.952541 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:02Z","lastTransitionTime":"2025-11-25T10:33:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:03 crc kubenswrapper[4821]: I1125 10:33:03.054795 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:03 crc kubenswrapper[4821]: I1125 10:33:03.054826 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:03 crc kubenswrapper[4821]: I1125 10:33:03.054834 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:03 crc kubenswrapper[4821]: I1125 10:33:03.054847 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:03 crc kubenswrapper[4821]: I1125 10:33:03.054855 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:03Z","lastTransitionTime":"2025-11-25T10:33:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:03 crc kubenswrapper[4821]: I1125 10:33:03.067444 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" Nov 25 10:33:03 crc kubenswrapper[4821]: I1125 10:33:03.068044 4821 scope.go:117] "RemoveContainer" containerID="7e6202b51ac13209f3a6de880638a6d142f79755311646c0216f11dfe8bbe3f7" Nov 25 10:33:03 crc kubenswrapper[4821]: E1125 10:33:03.068253 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-mzhnw_openshift-ovn-kubernetes(954085fa-d03a-4e88-9dfd-566257e3558d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" podUID="954085fa-d03a-4e88-9dfd-566257e3558d" Nov 25 10:33:03 crc kubenswrapper[4821]: I1125 10:33:03.078388 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tgdlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2336955b-a703-4db4-8d0f-d08088ed59b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb0362942bfac635ee6292501b1b851772c01d75dd658eef4b17669a716de72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vjfp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tgdlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:03Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:03 crc kubenswrapper[4821]: I1125 10:33:03.090522 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62fcb632-f12a-46e2-bb7d-a5311f291ee7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28d43531927e75c3745f100940f7f6bf79fd21026c1309194df8cd7a39361b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69633332d03d2551ff4abb253d358de398eab18486b92db630ce06c0447fb9cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ccb78d9a51ac83ee208c2451685a2512de9deaf742a663bb8c892aa59b76162\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27826e155a9352b1e1866e83fdab683dbcd749cc04b6a4da2db25ecabe3d9a52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://991252e5f56f128827a5054b8e228284bc0266166582cbaedaadc53ca2f5b401\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:03Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:03 crc kubenswrapper[4821]: I1125 10:33:03.101584 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63b471fd-5e16-4761-8b07-44a135d5f5f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a68a57fa7fba8f351f264e427a50bed23f57775644e9c268aa7305efb6916794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67df7c609d5a5d658be4c3e873ec3052bd070ad9469f8ff44c29a9571e3e5d95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://269ced131bc0777dc6752070470c95b4bedfe4b4240940eabd769500ddef0404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e48c0e3b58359c6c0a493d58f7d2aec6011747791a3c0d7788102e1fec18865\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e48c0e3b58359c6c0a493d58f7d2aec6011747791a3c0d7788102e1fec18865\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:03Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:03 crc kubenswrapper[4821]: I1125 10:33:03.114017 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:33:03 crc kubenswrapper[4821]: I1125 10:33:03.114035 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:03Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:03 crc kubenswrapper[4821]: I1125 10:33:03.114098 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:33:03 crc kubenswrapper[4821]: I1125 10:33:03.114027 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:33:03 crc kubenswrapper[4821]: E1125 10:33:03.114230 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:33:03 crc kubenswrapper[4821]: E1125 10:33:03.114154 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:33:03 crc kubenswrapper[4821]: E1125 10:33:03.114418 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:33:03 crc kubenswrapper[4821]: I1125 10:33:03.126033 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f948b87-ac86-4de6-ad64-c2ef947f84d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4482f73a45bc82bd55770c3fd788becacec94893fed1c305fed1c81c753d43aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ba68bd093be94d0d40f3851638db5d26bdfc52d175b1ac78afb91dda7affcfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2krbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:03Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:03 crc kubenswrapper[4821]: I1125 10:33:03.139461 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qppnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"357491b8-cd9c-486e-9898-5ba33678513d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc88d700ca3a449ef1cb32d6124b16d0024d6981ced12e276083f86a1b322ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd9c047c5ce1d896f803ff6cafa9bbf3a18a391b44539e1f1312e08c89e51cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd9c047c5ce1d896f803ff6cafa9bbf3a18a391b44539e1f1312e08c89e51cb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://565a6db63fb69abbba81ffdf6085797900858d263cd6affcb0b5f195474c2516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://565a6db63fb69abbba81ffdf6085797900858d263cd6affcb0b5f195474c2516\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://544885302dffff8336a9aa3c4afa80d69088ffaa2402648f4a329e0a798a8c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://544885302dffff8336a9aa3c4afa80d69088ffaa2402648f4a329e0a798a8c99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8bc193fb3f85b980c31af2e826cee90adb633ad2ee82ff1d46e182ab52e8fab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8bc193fb3f85b980c31af2e826cee90adb633ad2ee82ff1d46e182ab52e8fab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7713e2a35bac8df1e8987fbcd452459fa3541caad0c52ed57fda187bd7c83d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7713e2a35bac8df1e8987fbcd452459fa3541caad0c52ed57fda187bd7c83d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34adfca89026fc20007a73a8cbd4e7c7b3137a581c3476867af17fabe976f71f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34adfca89026fc20007a73a8cbd4e7c7b3137a581c3476867af17fabe976f71f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qppnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:03Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:03 crc kubenswrapper[4821]: I1125 10:33:03.150948 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06f949936bac99cfbf89d720c9549c074351d9077bd963fa024d87004c6be23e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:03Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:03 crc kubenswrapper[4821]: I1125 10:33:03.157720 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:03 crc kubenswrapper[4821]: I1125 10:33:03.157788 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:03 crc kubenswrapper[4821]: I1125 10:33:03.157800 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:03 crc kubenswrapper[4821]: I1125 10:33:03.157841 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:03 crc kubenswrapper[4821]: I1125 10:33:03.157853 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:03Z","lastTransitionTime":"2025-11-25T10:33:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:03 crc kubenswrapper[4821]: I1125 10:33:03.163999 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384d8ba58264ab89796d3b53afb0c4c1be33cb4fba719a8beef4afc2be7d66a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:03Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:03 crc kubenswrapper[4821]: I1125 10:33:03.178944 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:03Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:03 crc kubenswrapper[4821]: I1125 10:33:03.190684 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17b8714be3bd121e06641873e003f08560b1f71e56d6ea936966df7d4896eb03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85731e6b6dbe1777bc7161a010f39f79523ec4b08d71d5d8a975573aa7fe4957\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:03Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:03 crc kubenswrapper[4821]: I1125 10:33:03.204022 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zlk96" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d55aa91-9cb9-4c95-bc27-6fec05684c10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4521b5baee45ffc9c7661c781c0b3de458dff325ad525dde283d46af71c0cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vkd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zlk96\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:03Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:03 crc kubenswrapper[4821]: I1125 10:33:03.214324 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6hbqd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f71b3a2-63cb-443c-9f41-39e57030746c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65473b0e63c2e9c7c09a2bb8f31396daf1d87c2c83d20ed3508ce2fe7358dda3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxdmq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://595f9ea75461f18e763d84799e368becc07b72acdc5337d20d6995bf53c3d27a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxdmq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6hbqd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:03Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:03 crc kubenswrapper[4821]: I1125 10:33:03.232446 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"954085fa-d03a-4e88-9dfd-566257e3558d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f00746930c4588e548f0e2b60d52bf448aff7401b676a7855cb0f4cce00d329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a14a078ba7d2aed95992b36502008f315f0d3a634b64ed49da70398b2562d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d7254196b3702e48083ae98e829b73c539adf69ed5b2bbdc62189d09ae1762\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d727430108977574365f0e9c04b44e03307bd7ff89cdbb1e48a6e916f504d6b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94c7949f077b1ca2b310bedb090aa72fc4bd7e8f2c8bf438096edfca8568f36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed2e7cad7bb24b5b56a1295098a6a7c230c8f9b9c29a88ef653ae8b6ac31667d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e6202b51ac13209f3a6de880638a6d142f79755311646c0216f11dfe8bbe3f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e6202b51ac13209f3a6de880638a6d142f79755311646c0216f11dfe8bbe3f7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T10:32:56Z\\\",\\\"message\\\":\\\"to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:56Z is after 2025-08-24T17:21:41Z]\\\\nI1125 10:32:56.048189 6453 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-console/console_TCP_cluster\\\\\\\", UUID:\\\\\\\"d7d7b270-1480-47f8-bdf9-690dbab310cb\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console/console\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]strin\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-mzhnw_openshift-ovn-kubernetes(954085fa-d03a-4e88-9dfd-566257e3558d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dbf499f4ae6161752abaa7a466a027c6f874016be90a534e69f07e3dd3f3d9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzhnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:03Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:03 crc kubenswrapper[4821]: I1125 10:33:03.241810 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-d4msn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f12cef3-7e40-4f61-836e-23cd0e578e71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s7lkv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s7lkv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:44Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-d4msn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:03Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:03 crc kubenswrapper[4821]: I1125 10:33:03.259134 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d9e58df-1755-4422-8d04-211f98c7c011\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea762f627aa69fb88a9bd82545fa677e9ccfa97f99a5e92c0aa490e619c0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a06b811e5ec52ce984a2403ba2c98332ba60466d08b3580ad9192fac6caef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0385f6ed9597974267f7c45547f6b6725b4ee83a2cc0508ff2d1c1ce8753ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b915b565f7fbb3e4e856e7f1035a519e82bfb3023660c121f6e3813de353bb92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18465f3a6916c36457a305ba5d568c9077f34f1d000532ed518f237d100008b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:03Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:03 crc kubenswrapper[4821]: I1125 10:33:03.260144 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:03 crc kubenswrapper[4821]: I1125 10:33:03.260202 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:03 crc kubenswrapper[4821]: I1125 10:33:03.260221 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:03 crc kubenswrapper[4821]: I1125 10:33:03.260242 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:03 crc kubenswrapper[4821]: I1125 10:33:03.260255 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:03Z","lastTransitionTime":"2025-11-25T10:33:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:03 crc kubenswrapper[4821]: I1125 10:33:03.269708 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25595ccd-762d-414a-9916-4fe38d478b07\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d4b2d8db4d13392d9789e161fd4b8f6eef983b79128ee584d72e4c06565bc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf780f791771ea43a8c8176d02fdfb5204406e023207299369ae46ceb342baaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0251519fd4262f2ea5c64959db49433afdafe36b7ed2404d660eeb7c365d2746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad5c205e0eeecfd43cbdf0eb1a334e3e8faa5e9a9e8dd202c397d5cbb36e081e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:03Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:03 crc kubenswrapper[4821]: I1125 10:33:03.281035 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:03Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:03 crc kubenswrapper[4821]: I1125 10:33:03.292226 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ldzbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b84e32ae-4897-4644-a656-6bd82012cc81\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ef433744761f26e8e1f6c0db123821a11923ffd774830744285cd4178d89533\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5pmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ldzbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:03Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:03 crc kubenswrapper[4821]: I1125 10:33:03.362391 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:03 crc kubenswrapper[4821]: I1125 10:33:03.362438 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:03 crc kubenswrapper[4821]: I1125 10:33:03.362448 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:03 crc kubenswrapper[4821]: I1125 10:33:03.362464 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:03 crc kubenswrapper[4821]: I1125 10:33:03.362478 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:03Z","lastTransitionTime":"2025-11-25T10:33:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:03 crc kubenswrapper[4821]: I1125 10:33:03.464623 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:03 crc kubenswrapper[4821]: I1125 10:33:03.464676 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:03 crc kubenswrapper[4821]: I1125 10:33:03.464689 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:03 crc kubenswrapper[4821]: I1125 10:33:03.464708 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:03 crc kubenswrapper[4821]: I1125 10:33:03.464721 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:03Z","lastTransitionTime":"2025-11-25T10:33:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:03 crc kubenswrapper[4821]: I1125 10:33:03.568025 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:03 crc kubenswrapper[4821]: I1125 10:33:03.568590 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:03 crc kubenswrapper[4821]: I1125 10:33:03.568606 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:03 crc kubenswrapper[4821]: I1125 10:33:03.568625 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:03 crc kubenswrapper[4821]: I1125 10:33:03.568636 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:03Z","lastTransitionTime":"2025-11-25T10:33:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:03 crc kubenswrapper[4821]: I1125 10:33:03.671324 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:03 crc kubenswrapper[4821]: I1125 10:33:03.671365 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:03 crc kubenswrapper[4821]: I1125 10:33:03.671381 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:03 crc kubenswrapper[4821]: I1125 10:33:03.671397 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:03 crc kubenswrapper[4821]: I1125 10:33:03.671408 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:03Z","lastTransitionTime":"2025-11-25T10:33:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:03 crc kubenswrapper[4821]: I1125 10:33:03.773747 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:03 crc kubenswrapper[4821]: I1125 10:33:03.773786 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:03 crc kubenswrapper[4821]: I1125 10:33:03.773793 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:03 crc kubenswrapper[4821]: I1125 10:33:03.773807 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:03 crc kubenswrapper[4821]: I1125 10:33:03.773817 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:03Z","lastTransitionTime":"2025-11-25T10:33:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:03 crc kubenswrapper[4821]: I1125 10:33:03.876148 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:03 crc kubenswrapper[4821]: I1125 10:33:03.876212 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:03 crc kubenswrapper[4821]: I1125 10:33:03.876229 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:03 crc kubenswrapper[4821]: I1125 10:33:03.876247 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:03 crc kubenswrapper[4821]: I1125 10:33:03.876257 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:03Z","lastTransitionTime":"2025-11-25T10:33:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:03 crc kubenswrapper[4821]: I1125 10:33:03.978568 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:03 crc kubenswrapper[4821]: I1125 10:33:03.978647 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:03 crc kubenswrapper[4821]: I1125 10:33:03.978663 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:03 crc kubenswrapper[4821]: I1125 10:33:03.978686 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:03 crc kubenswrapper[4821]: I1125 10:33:03.978699 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:03Z","lastTransitionTime":"2025-11-25T10:33:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:04 crc kubenswrapper[4821]: I1125 10:33:04.081358 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:04 crc kubenswrapper[4821]: I1125 10:33:04.081392 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:04 crc kubenswrapper[4821]: I1125 10:33:04.081400 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:04 crc kubenswrapper[4821]: I1125 10:33:04.081415 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:04 crc kubenswrapper[4821]: I1125 10:33:04.081423 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:04Z","lastTransitionTime":"2025-11-25T10:33:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:04 crc kubenswrapper[4821]: I1125 10:33:04.114275 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4msn" Nov 25 10:33:04 crc kubenswrapper[4821]: E1125 10:33:04.114423 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4msn" podUID="1f12cef3-7e40-4f61-836e-23cd0e578e71" Nov 25 10:33:04 crc kubenswrapper[4821]: I1125 10:33:04.184051 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:04 crc kubenswrapper[4821]: I1125 10:33:04.184111 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:04 crc kubenswrapper[4821]: I1125 10:33:04.184122 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:04 crc kubenswrapper[4821]: I1125 10:33:04.184139 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:04 crc kubenswrapper[4821]: I1125 10:33:04.184150 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:04Z","lastTransitionTime":"2025-11-25T10:33:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:04 crc kubenswrapper[4821]: I1125 10:33:04.286331 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:04 crc kubenswrapper[4821]: I1125 10:33:04.286375 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:04 crc kubenswrapper[4821]: I1125 10:33:04.286385 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:04 crc kubenswrapper[4821]: I1125 10:33:04.286402 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:04 crc kubenswrapper[4821]: I1125 10:33:04.286413 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:04Z","lastTransitionTime":"2025-11-25T10:33:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:04 crc kubenswrapper[4821]: I1125 10:33:04.388614 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:04 crc kubenswrapper[4821]: I1125 10:33:04.388671 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:04 crc kubenswrapper[4821]: I1125 10:33:04.388685 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:04 crc kubenswrapper[4821]: I1125 10:33:04.388703 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:04 crc kubenswrapper[4821]: I1125 10:33:04.388714 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:04Z","lastTransitionTime":"2025-11-25T10:33:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:04 crc kubenswrapper[4821]: I1125 10:33:04.490703 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:04 crc kubenswrapper[4821]: I1125 10:33:04.490745 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:04 crc kubenswrapper[4821]: I1125 10:33:04.490754 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:04 crc kubenswrapper[4821]: I1125 10:33:04.490774 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:04 crc kubenswrapper[4821]: I1125 10:33:04.490790 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:04Z","lastTransitionTime":"2025-11-25T10:33:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:04 crc kubenswrapper[4821]: I1125 10:33:04.593474 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:04 crc kubenswrapper[4821]: I1125 10:33:04.593507 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:04 crc kubenswrapper[4821]: I1125 10:33:04.593515 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:04 crc kubenswrapper[4821]: I1125 10:33:04.593528 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:04 crc kubenswrapper[4821]: I1125 10:33:04.593536 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:04Z","lastTransitionTime":"2025-11-25T10:33:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:04 crc kubenswrapper[4821]: I1125 10:33:04.696152 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:04 crc kubenswrapper[4821]: I1125 10:33:04.696209 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:04 crc kubenswrapper[4821]: I1125 10:33:04.696218 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:04 crc kubenswrapper[4821]: I1125 10:33:04.696231 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:04 crc kubenswrapper[4821]: I1125 10:33:04.696241 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:04Z","lastTransitionTime":"2025-11-25T10:33:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:04 crc kubenswrapper[4821]: I1125 10:33:04.799731 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:04 crc kubenswrapper[4821]: I1125 10:33:04.799774 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:04 crc kubenswrapper[4821]: I1125 10:33:04.799784 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:04 crc kubenswrapper[4821]: I1125 10:33:04.799799 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:04 crc kubenswrapper[4821]: I1125 10:33:04.799809 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:04Z","lastTransitionTime":"2025-11-25T10:33:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:04 crc kubenswrapper[4821]: I1125 10:33:04.902225 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:04 crc kubenswrapper[4821]: I1125 10:33:04.902256 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:04 crc kubenswrapper[4821]: I1125 10:33:04.902263 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:04 crc kubenswrapper[4821]: I1125 10:33:04.902276 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:04 crc kubenswrapper[4821]: I1125 10:33:04.902285 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:04Z","lastTransitionTime":"2025-11-25T10:33:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:05 crc kubenswrapper[4821]: I1125 10:33:05.004111 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:05 crc kubenswrapper[4821]: I1125 10:33:05.004143 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:05 crc kubenswrapper[4821]: I1125 10:33:05.004152 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:05 crc kubenswrapper[4821]: I1125 10:33:05.004182 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:05 crc kubenswrapper[4821]: I1125 10:33:05.004191 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:05Z","lastTransitionTime":"2025-11-25T10:33:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:05 crc kubenswrapper[4821]: I1125 10:33:05.106798 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:05 crc kubenswrapper[4821]: I1125 10:33:05.106835 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:05 crc kubenswrapper[4821]: I1125 10:33:05.106844 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:05 crc kubenswrapper[4821]: I1125 10:33:05.106861 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:05 crc kubenswrapper[4821]: I1125 10:33:05.106872 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:05Z","lastTransitionTime":"2025-11-25T10:33:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:05 crc kubenswrapper[4821]: I1125 10:33:05.113666 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:33:05 crc kubenswrapper[4821]: I1125 10:33:05.113717 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:33:05 crc kubenswrapper[4821]: I1125 10:33:05.113740 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:33:05 crc kubenswrapper[4821]: E1125 10:33:05.113788 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:33:05 crc kubenswrapper[4821]: E1125 10:33:05.113826 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:33:05 crc kubenswrapper[4821]: E1125 10:33:05.113892 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:33:05 crc kubenswrapper[4821]: I1125 10:33:05.209263 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:05 crc kubenswrapper[4821]: I1125 10:33:05.209304 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:05 crc kubenswrapper[4821]: I1125 10:33:05.209312 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:05 crc kubenswrapper[4821]: I1125 10:33:05.209327 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:05 crc kubenswrapper[4821]: I1125 10:33:05.209337 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:05Z","lastTransitionTime":"2025-11-25T10:33:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:05 crc kubenswrapper[4821]: I1125 10:33:05.311367 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:05 crc kubenswrapper[4821]: I1125 10:33:05.311399 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:05 crc kubenswrapper[4821]: I1125 10:33:05.311408 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:05 crc kubenswrapper[4821]: I1125 10:33:05.311422 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:05 crc kubenswrapper[4821]: I1125 10:33:05.311432 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:05Z","lastTransitionTime":"2025-11-25T10:33:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:05 crc kubenswrapper[4821]: I1125 10:33:05.414307 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:05 crc kubenswrapper[4821]: I1125 10:33:05.414376 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:05 crc kubenswrapper[4821]: I1125 10:33:05.414389 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:05 crc kubenswrapper[4821]: I1125 10:33:05.414406 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:05 crc kubenswrapper[4821]: I1125 10:33:05.414419 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:05Z","lastTransitionTime":"2025-11-25T10:33:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:05 crc kubenswrapper[4821]: I1125 10:33:05.517476 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:05 crc kubenswrapper[4821]: I1125 10:33:05.517516 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:05 crc kubenswrapper[4821]: I1125 10:33:05.517527 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:05 crc kubenswrapper[4821]: I1125 10:33:05.517544 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:05 crc kubenswrapper[4821]: I1125 10:33:05.517557 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:05Z","lastTransitionTime":"2025-11-25T10:33:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:05 crc kubenswrapper[4821]: I1125 10:33:05.619442 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:05 crc kubenswrapper[4821]: I1125 10:33:05.619479 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:05 crc kubenswrapper[4821]: I1125 10:33:05.619488 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:05 crc kubenswrapper[4821]: I1125 10:33:05.619501 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:05 crc kubenswrapper[4821]: I1125 10:33:05.619510 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:05Z","lastTransitionTime":"2025-11-25T10:33:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:05 crc kubenswrapper[4821]: I1125 10:33:05.722037 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:05 crc kubenswrapper[4821]: I1125 10:33:05.722087 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:05 crc kubenswrapper[4821]: I1125 10:33:05.722098 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:05 crc kubenswrapper[4821]: I1125 10:33:05.722119 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:05 crc kubenswrapper[4821]: I1125 10:33:05.722134 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:05Z","lastTransitionTime":"2025-11-25T10:33:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:05 crc kubenswrapper[4821]: I1125 10:33:05.824678 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:05 crc kubenswrapper[4821]: I1125 10:33:05.824713 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:05 crc kubenswrapper[4821]: I1125 10:33:05.824722 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:05 crc kubenswrapper[4821]: I1125 10:33:05.824736 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:05 crc kubenswrapper[4821]: I1125 10:33:05.824746 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:05Z","lastTransitionTime":"2025-11-25T10:33:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:05 crc kubenswrapper[4821]: I1125 10:33:05.926820 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:05 crc kubenswrapper[4821]: I1125 10:33:05.926850 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:05 crc kubenswrapper[4821]: I1125 10:33:05.926862 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:05 crc kubenswrapper[4821]: I1125 10:33:05.926884 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:05 crc kubenswrapper[4821]: I1125 10:33:05.926897 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:05Z","lastTransitionTime":"2025-11-25T10:33:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.029056 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.029089 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.029097 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.029111 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.029119 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:06Z","lastTransitionTime":"2025-11-25T10:33:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.113683 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4msn" Nov 25 10:33:06 crc kubenswrapper[4821]: E1125 10:33:06.113878 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4msn" podUID="1f12cef3-7e40-4f61-836e-23cd0e578e71" Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.131099 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.131192 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.131219 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.131238 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.131250 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:06Z","lastTransitionTime":"2025-11-25T10:33:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.234048 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.234109 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.234121 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.234140 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.234152 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:06Z","lastTransitionTime":"2025-11-25T10:33:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.336677 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.336713 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.336722 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.336735 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.336744 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:06Z","lastTransitionTime":"2025-11-25T10:33:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.439112 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.439155 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.439187 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.439202 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.439211 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:06Z","lastTransitionTime":"2025-11-25T10:33:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.542028 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.542083 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.542092 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.542106 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.542114 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:06Z","lastTransitionTime":"2025-11-25T10:33:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.595260 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.595301 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.595311 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.595327 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.595338 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:06Z","lastTransitionTime":"2025-11-25T10:33:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:06 crc kubenswrapper[4821]: E1125 10:33:06.607361 4821 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a5b2b070-39b9-412c-90f1-c7ed2e9e0950\\\",\\\"systemUUID\\\":\\\"ea86e80f-41b7-4190-ad7e-d8c549d535de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:06Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.610637 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.610684 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.610698 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.610715 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.610726 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:06Z","lastTransitionTime":"2025-11-25T10:33:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:06 crc kubenswrapper[4821]: E1125 10:33:06.621290 4821 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a5b2b070-39b9-412c-90f1-c7ed2e9e0950\\\",\\\"systemUUID\\\":\\\"ea86e80f-41b7-4190-ad7e-d8c549d535de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:06Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.623694 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.623718 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.623726 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.623758 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.623767 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:06Z","lastTransitionTime":"2025-11-25T10:33:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:06 crc kubenswrapper[4821]: E1125 10:33:06.634080 4821 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a5b2b070-39b9-412c-90f1-c7ed2e9e0950\\\",\\\"systemUUID\\\":\\\"ea86e80f-41b7-4190-ad7e-d8c549d535de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:06Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.636982 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.637009 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.637035 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.637049 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.637058 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:06Z","lastTransitionTime":"2025-11-25T10:33:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:06 crc kubenswrapper[4821]: E1125 10:33:06.648264 4821 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a5b2b070-39b9-412c-90f1-c7ed2e9e0950\\\",\\\"systemUUID\\\":\\\"ea86e80f-41b7-4190-ad7e-d8c549d535de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:06Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.654309 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.654347 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.654357 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.654370 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.654386 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:06Z","lastTransitionTime":"2025-11-25T10:33:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:06 crc kubenswrapper[4821]: E1125 10:33:06.666532 4821 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a5b2b070-39b9-412c-90f1-c7ed2e9e0950\\\",\\\"systemUUID\\\":\\\"ea86e80f-41b7-4190-ad7e-d8c549d535de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:06Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:06 crc kubenswrapper[4821]: E1125 10:33:06.666712 4821 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.668234 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.668267 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.668279 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.668294 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.668304 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:06Z","lastTransitionTime":"2025-11-25T10:33:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.771370 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.771405 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.771415 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.771431 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.771442 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:06Z","lastTransitionTime":"2025-11-25T10:33:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.873675 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.873710 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.873720 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.873736 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.873748 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:06Z","lastTransitionTime":"2025-11-25T10:33:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.975869 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.975908 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.975918 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.975936 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:06 crc kubenswrapper[4821]: I1125 10:33:06.975946 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:06Z","lastTransitionTime":"2025-11-25T10:33:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:07 crc kubenswrapper[4821]: I1125 10:33:07.078629 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:07 crc kubenswrapper[4821]: I1125 10:33:07.078681 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:07 crc kubenswrapper[4821]: I1125 10:33:07.078692 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:07 crc kubenswrapper[4821]: I1125 10:33:07.078710 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:07 crc kubenswrapper[4821]: I1125 10:33:07.078723 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:07Z","lastTransitionTime":"2025-11-25T10:33:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:07 crc kubenswrapper[4821]: I1125 10:33:07.113347 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:33:07 crc kubenswrapper[4821]: I1125 10:33:07.113389 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:33:07 crc kubenswrapper[4821]: I1125 10:33:07.113423 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:33:07 crc kubenswrapper[4821]: E1125 10:33:07.113521 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:33:07 crc kubenswrapper[4821]: E1125 10:33:07.113607 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:33:07 crc kubenswrapper[4821]: E1125 10:33:07.113725 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:33:07 crc kubenswrapper[4821]: I1125 10:33:07.180694 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:07 crc kubenswrapper[4821]: I1125 10:33:07.180739 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:07 crc kubenswrapper[4821]: I1125 10:33:07.180748 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:07 crc kubenswrapper[4821]: I1125 10:33:07.180763 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:07 crc kubenswrapper[4821]: I1125 10:33:07.180772 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:07Z","lastTransitionTime":"2025-11-25T10:33:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:07 crc kubenswrapper[4821]: I1125 10:33:07.283495 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:07 crc kubenswrapper[4821]: I1125 10:33:07.283544 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:07 crc kubenswrapper[4821]: I1125 10:33:07.283558 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:07 crc kubenswrapper[4821]: I1125 10:33:07.283575 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:07 crc kubenswrapper[4821]: I1125 10:33:07.283586 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:07Z","lastTransitionTime":"2025-11-25T10:33:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:07 crc kubenswrapper[4821]: I1125 10:33:07.385725 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:07 crc kubenswrapper[4821]: I1125 10:33:07.385762 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:07 crc kubenswrapper[4821]: I1125 10:33:07.385773 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:07 crc kubenswrapper[4821]: I1125 10:33:07.385788 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:07 crc kubenswrapper[4821]: I1125 10:33:07.385799 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:07Z","lastTransitionTime":"2025-11-25T10:33:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:07 crc kubenswrapper[4821]: I1125 10:33:07.488403 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:07 crc kubenswrapper[4821]: I1125 10:33:07.488449 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:07 crc kubenswrapper[4821]: I1125 10:33:07.488464 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:07 crc kubenswrapper[4821]: I1125 10:33:07.488486 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:07 crc kubenswrapper[4821]: I1125 10:33:07.488502 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:07Z","lastTransitionTime":"2025-11-25T10:33:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:07 crc kubenswrapper[4821]: I1125 10:33:07.590242 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:07 crc kubenswrapper[4821]: I1125 10:33:07.590288 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:07 crc kubenswrapper[4821]: I1125 10:33:07.590299 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:07 crc kubenswrapper[4821]: I1125 10:33:07.590314 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:07 crc kubenswrapper[4821]: I1125 10:33:07.590324 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:07Z","lastTransitionTime":"2025-11-25T10:33:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:07 crc kubenswrapper[4821]: I1125 10:33:07.692738 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:07 crc kubenswrapper[4821]: I1125 10:33:07.692846 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:07 crc kubenswrapper[4821]: I1125 10:33:07.692862 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:07 crc kubenswrapper[4821]: I1125 10:33:07.692887 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:07 crc kubenswrapper[4821]: I1125 10:33:07.692902 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:07Z","lastTransitionTime":"2025-11-25T10:33:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:07 crc kubenswrapper[4821]: I1125 10:33:07.794678 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:07 crc kubenswrapper[4821]: I1125 10:33:07.794715 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:07 crc kubenswrapper[4821]: I1125 10:33:07.794724 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:07 crc kubenswrapper[4821]: I1125 10:33:07.794738 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:07 crc kubenswrapper[4821]: I1125 10:33:07.794746 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:07Z","lastTransitionTime":"2025-11-25T10:33:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:07 crc kubenswrapper[4821]: I1125 10:33:07.896778 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:07 crc kubenswrapper[4821]: I1125 10:33:07.896817 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:07 crc kubenswrapper[4821]: I1125 10:33:07.896829 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:07 crc kubenswrapper[4821]: I1125 10:33:07.896845 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:07 crc kubenswrapper[4821]: I1125 10:33:07.896857 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:07Z","lastTransitionTime":"2025-11-25T10:33:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:08 crc kubenswrapper[4821]: I1125 10:33:08.000882 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:08 crc kubenswrapper[4821]: I1125 10:33:08.000948 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:08 crc kubenswrapper[4821]: I1125 10:33:08.000960 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:08 crc kubenswrapper[4821]: I1125 10:33:08.000995 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:08 crc kubenswrapper[4821]: I1125 10:33:08.001006 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:08Z","lastTransitionTime":"2025-11-25T10:33:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:08 crc kubenswrapper[4821]: I1125 10:33:08.103581 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:08 crc kubenswrapper[4821]: I1125 10:33:08.103623 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:08 crc kubenswrapper[4821]: I1125 10:33:08.103634 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:08 crc kubenswrapper[4821]: I1125 10:33:08.103651 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:08 crc kubenswrapper[4821]: I1125 10:33:08.103662 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:08Z","lastTransitionTime":"2025-11-25T10:33:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:08 crc kubenswrapper[4821]: I1125 10:33:08.114099 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4msn" Nov 25 10:33:08 crc kubenswrapper[4821]: E1125 10:33:08.114248 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4msn" podUID="1f12cef3-7e40-4f61-836e-23cd0e578e71" Nov 25 10:33:08 crc kubenswrapper[4821]: I1125 10:33:08.206080 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:08 crc kubenswrapper[4821]: I1125 10:33:08.206250 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:08 crc kubenswrapper[4821]: I1125 10:33:08.206280 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:08 crc kubenswrapper[4821]: I1125 10:33:08.206309 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:08 crc kubenswrapper[4821]: I1125 10:33:08.206327 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:08Z","lastTransitionTime":"2025-11-25T10:33:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:08 crc kubenswrapper[4821]: I1125 10:33:08.309601 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:08 crc kubenswrapper[4821]: I1125 10:33:08.309650 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:08 crc kubenswrapper[4821]: I1125 10:33:08.309662 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:08 crc kubenswrapper[4821]: I1125 10:33:08.309681 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:08 crc kubenswrapper[4821]: I1125 10:33:08.309699 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:08Z","lastTransitionTime":"2025-11-25T10:33:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:08 crc kubenswrapper[4821]: I1125 10:33:08.411615 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:08 crc kubenswrapper[4821]: I1125 10:33:08.411652 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:08 crc kubenswrapper[4821]: I1125 10:33:08.411663 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:08 crc kubenswrapper[4821]: I1125 10:33:08.411676 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:08 crc kubenswrapper[4821]: I1125 10:33:08.411686 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:08Z","lastTransitionTime":"2025-11-25T10:33:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:08 crc kubenswrapper[4821]: I1125 10:33:08.514056 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:08 crc kubenswrapper[4821]: I1125 10:33:08.514115 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:08 crc kubenswrapper[4821]: I1125 10:33:08.514125 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:08 crc kubenswrapper[4821]: I1125 10:33:08.514141 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:08 crc kubenswrapper[4821]: I1125 10:33:08.514150 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:08Z","lastTransitionTime":"2025-11-25T10:33:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:08 crc kubenswrapper[4821]: I1125 10:33:08.616704 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:08 crc kubenswrapper[4821]: I1125 10:33:08.616769 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:08 crc kubenswrapper[4821]: I1125 10:33:08.616779 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:08 crc kubenswrapper[4821]: I1125 10:33:08.616797 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:08 crc kubenswrapper[4821]: I1125 10:33:08.616812 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:08Z","lastTransitionTime":"2025-11-25T10:33:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:08 crc kubenswrapper[4821]: I1125 10:33:08.719455 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:08 crc kubenswrapper[4821]: I1125 10:33:08.719496 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:08 crc kubenswrapper[4821]: I1125 10:33:08.719507 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:08 crc kubenswrapper[4821]: I1125 10:33:08.719523 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:08 crc kubenswrapper[4821]: I1125 10:33:08.719535 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:08Z","lastTransitionTime":"2025-11-25T10:33:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:08 crc kubenswrapper[4821]: I1125 10:33:08.821705 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:08 crc kubenswrapper[4821]: I1125 10:33:08.821745 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:08 crc kubenswrapper[4821]: I1125 10:33:08.821759 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:08 crc kubenswrapper[4821]: I1125 10:33:08.821775 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:08 crc kubenswrapper[4821]: I1125 10:33:08.821787 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:08Z","lastTransitionTime":"2025-11-25T10:33:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:08 crc kubenswrapper[4821]: I1125 10:33:08.924034 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:08 crc kubenswrapper[4821]: I1125 10:33:08.924096 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:08 crc kubenswrapper[4821]: I1125 10:33:08.924108 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:08 crc kubenswrapper[4821]: I1125 10:33:08.924126 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:08 crc kubenswrapper[4821]: I1125 10:33:08.924139 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:08Z","lastTransitionTime":"2025-11-25T10:33:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:09 crc kubenswrapper[4821]: I1125 10:33:09.026780 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:09 crc kubenswrapper[4821]: I1125 10:33:09.026839 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:09 crc kubenswrapper[4821]: I1125 10:33:09.026852 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:09 crc kubenswrapper[4821]: I1125 10:33:09.026869 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:09 crc kubenswrapper[4821]: I1125 10:33:09.026882 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:09Z","lastTransitionTime":"2025-11-25T10:33:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:09 crc kubenswrapper[4821]: I1125 10:33:09.113945 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:33:09 crc kubenswrapper[4821]: I1125 10:33:09.114007 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:33:09 crc kubenswrapper[4821]: E1125 10:33:09.114115 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:33:09 crc kubenswrapper[4821]: I1125 10:33:09.114145 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:33:09 crc kubenswrapper[4821]: E1125 10:33:09.114273 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:33:09 crc kubenswrapper[4821]: E1125 10:33:09.114382 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:33:09 crc kubenswrapper[4821]: I1125 10:33:09.134108 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:09 crc kubenswrapper[4821]: I1125 10:33:09.134140 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:09 crc kubenswrapper[4821]: I1125 10:33:09.134175 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:09 crc kubenswrapper[4821]: I1125 10:33:09.134193 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:09 crc kubenswrapper[4821]: I1125 10:33:09.134203 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:09Z","lastTransitionTime":"2025-11-25T10:33:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:09 crc kubenswrapper[4821]: I1125 10:33:09.236840 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:09 crc kubenswrapper[4821]: I1125 10:33:09.236904 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:09 crc kubenswrapper[4821]: I1125 10:33:09.236914 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:09 crc kubenswrapper[4821]: I1125 10:33:09.236950 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:09 crc kubenswrapper[4821]: I1125 10:33:09.236960 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:09Z","lastTransitionTime":"2025-11-25T10:33:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:09 crc kubenswrapper[4821]: I1125 10:33:09.339568 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:09 crc kubenswrapper[4821]: I1125 10:33:09.339613 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:09 crc kubenswrapper[4821]: I1125 10:33:09.339625 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:09 crc kubenswrapper[4821]: I1125 10:33:09.339658 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:09 crc kubenswrapper[4821]: I1125 10:33:09.339670 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:09Z","lastTransitionTime":"2025-11-25T10:33:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:09 crc kubenswrapper[4821]: I1125 10:33:09.442260 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:09 crc kubenswrapper[4821]: I1125 10:33:09.442311 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:09 crc kubenswrapper[4821]: I1125 10:33:09.442323 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:09 crc kubenswrapper[4821]: I1125 10:33:09.442341 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:09 crc kubenswrapper[4821]: I1125 10:33:09.442679 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:09Z","lastTransitionTime":"2025-11-25T10:33:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:09 crc kubenswrapper[4821]: I1125 10:33:09.544449 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:09 crc kubenswrapper[4821]: I1125 10:33:09.544584 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:09 crc kubenswrapper[4821]: I1125 10:33:09.544601 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:09 crc kubenswrapper[4821]: I1125 10:33:09.544621 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:09 crc kubenswrapper[4821]: I1125 10:33:09.544635 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:09Z","lastTransitionTime":"2025-11-25T10:33:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:09 crc kubenswrapper[4821]: I1125 10:33:09.647246 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:09 crc kubenswrapper[4821]: I1125 10:33:09.647274 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:09 crc kubenswrapper[4821]: I1125 10:33:09.647282 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:09 crc kubenswrapper[4821]: I1125 10:33:09.647295 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:09 crc kubenswrapper[4821]: I1125 10:33:09.647304 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:09Z","lastTransitionTime":"2025-11-25T10:33:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:09 crc kubenswrapper[4821]: I1125 10:33:09.749676 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:09 crc kubenswrapper[4821]: I1125 10:33:09.749714 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:09 crc kubenswrapper[4821]: I1125 10:33:09.749723 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:09 crc kubenswrapper[4821]: I1125 10:33:09.749738 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:09 crc kubenswrapper[4821]: I1125 10:33:09.749747 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:09Z","lastTransitionTime":"2025-11-25T10:33:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:09 crc kubenswrapper[4821]: I1125 10:33:09.852216 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:09 crc kubenswrapper[4821]: I1125 10:33:09.852256 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:09 crc kubenswrapper[4821]: I1125 10:33:09.852265 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:09 crc kubenswrapper[4821]: I1125 10:33:09.852278 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:09 crc kubenswrapper[4821]: I1125 10:33:09.852287 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:09Z","lastTransitionTime":"2025-11-25T10:33:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:09 crc kubenswrapper[4821]: I1125 10:33:09.954241 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:09 crc kubenswrapper[4821]: I1125 10:33:09.954270 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:09 crc kubenswrapper[4821]: I1125 10:33:09.954278 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:09 crc kubenswrapper[4821]: I1125 10:33:09.954291 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:09 crc kubenswrapper[4821]: I1125 10:33:09.954299 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:09Z","lastTransitionTime":"2025-11-25T10:33:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:10 crc kubenswrapper[4821]: I1125 10:33:10.057555 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:10 crc kubenswrapper[4821]: I1125 10:33:10.057621 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:10 crc kubenswrapper[4821]: I1125 10:33:10.057642 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:10 crc kubenswrapper[4821]: I1125 10:33:10.057671 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:10 crc kubenswrapper[4821]: I1125 10:33:10.057694 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:10Z","lastTransitionTime":"2025-11-25T10:33:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:10 crc kubenswrapper[4821]: I1125 10:33:10.115235 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4msn" Nov 25 10:33:10 crc kubenswrapper[4821]: E1125 10:33:10.115367 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4msn" podUID="1f12cef3-7e40-4f61-836e-23cd0e578e71" Nov 25 10:33:10 crc kubenswrapper[4821]: I1125 10:33:10.133916 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62fcb632-f12a-46e2-bb7d-a5311f291ee7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28d43531927e75c3745f100940f7f6bf79fd21026c1309194df8cd7a39361b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69633332d03d2551ff4abb253d358de398eab18486b92db630ce06c0447fb9cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ccb78d9a51ac83ee208c2451685a2512de9deaf742a663bb8c892aa59b76162\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27826e155a9352b1e1866e83fdab683dbcd749cc04b6a4da2db25ecabe3d9a52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://991252e5f56f128827a5054b8e228284bc0266166582cbaedaadc53ca2f5b401\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:10Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:10 crc kubenswrapper[4821]: I1125 10:33:10.146502 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63b471fd-5e16-4761-8b07-44a135d5f5f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a68a57fa7fba8f351f264e427a50bed23f57775644e9c268aa7305efb6916794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67df7c609d5a5d658be4c3e873ec3052bd070ad9469f8ff44c29a9571e3e5d95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://269ced131bc0777dc6752070470c95b4bedfe4b4240940eabd769500ddef0404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e48c0e3b58359c6c0a493d58f7d2aec6011747791a3c0d7788102e1fec18865\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e48c0e3b58359c6c0a493d58f7d2aec6011747791a3c0d7788102e1fec18865\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:10Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:10 crc kubenswrapper[4821]: I1125 10:33:10.160599 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:10 crc kubenswrapper[4821]: I1125 10:33:10.160657 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:10 crc kubenswrapper[4821]: I1125 10:33:10.160674 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:10 crc kubenswrapper[4821]: I1125 10:33:10.160697 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:10 crc kubenswrapper[4821]: I1125 10:33:10.160712 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:10Z","lastTransitionTime":"2025-11-25T10:33:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:10 crc kubenswrapper[4821]: I1125 10:33:10.162916 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:10Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:10 crc kubenswrapper[4821]: I1125 10:33:10.173727 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f948b87-ac86-4de6-ad64-c2ef947f84d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4482f73a45bc82bd55770c3fd788becacec94893fed1c305fed1c81c753d43aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ba68bd093be94d0d40f3851638db5d26bdfc52d175b1ac78afb91dda7affcfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2krbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:10Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:10 crc kubenswrapper[4821]: I1125 10:33:10.190612 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qppnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"357491b8-cd9c-486e-9898-5ba33678513d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc88d700ca3a449ef1cb32d6124b16d0024d6981ced12e276083f86a1b322ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd9c047c5ce1d896f803ff6cafa9bbf3a18a391b44539e1f1312e08c89e51cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd9c047c5ce1d896f803ff6cafa9bbf3a18a391b44539e1f1312e08c89e51cb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://565a6db63fb69abbba81ffdf6085797900858d263cd6affcb0b5f195474c2516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://565a6db63fb69abbba81ffdf6085797900858d263cd6affcb0b5f195474c2516\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://544885302dffff8336a9aa3c4afa80d69088ffaa2402648f4a329e0a798a8c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://544885302dffff8336a9aa3c4afa80d69088ffaa2402648f4a329e0a798a8c99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8bc193fb3f85b980c31af2e826cee90adb633ad2ee82ff1d46e182ab52e8fab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8bc193fb3f85b980c31af2e826cee90adb633ad2ee82ff1d46e182ab52e8fab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7713e2a35bac8df1e8987fbcd452459fa3541caad0c52ed57fda187bd7c83d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7713e2a35bac8df1e8987fbcd452459fa3541caad0c52ed57fda187bd7c83d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34adfca89026fc20007a73a8cbd4e7c7b3137a581c3476867af17fabe976f71f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34adfca89026fc20007a73a8cbd4e7c7b3137a581c3476867af17fabe976f71f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qppnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:10Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:10 crc kubenswrapper[4821]: I1125 10:33:10.201624 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tgdlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2336955b-a703-4db4-8d0f-d08088ed59b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb0362942bfac635ee6292501b1b851772c01d75dd658eef4b17669a716de72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vjfp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tgdlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:10Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:10 crc kubenswrapper[4821]: I1125 10:33:10.213005 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06f949936bac99cfbf89d720c9549c074351d9077bd963fa024d87004c6be23e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:10Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:10 crc kubenswrapper[4821]: I1125 10:33:10.222050 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384d8ba58264ab89796d3b53afb0c4c1be33cb4fba719a8beef4afc2be7d66a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:10Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:10 crc kubenswrapper[4821]: I1125 10:33:10.231822 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:10Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:10 crc kubenswrapper[4821]: I1125 10:33:10.246428 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17b8714be3bd121e06641873e003f08560b1f71e56d6ea936966df7d4896eb03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85731e6b6dbe1777bc7161a010f39f79523ec4b08d71d5d8a975573aa7fe4957\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:10Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:10 crc kubenswrapper[4821]: I1125 10:33:10.258489 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zlk96" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d55aa91-9cb9-4c95-bc27-6fec05684c10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4521b5baee45ffc9c7661c781c0b3de458dff325ad525dde283d46af71c0cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vkd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zlk96\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:10Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:10 crc kubenswrapper[4821]: I1125 10:33:10.265349 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:10 crc kubenswrapper[4821]: I1125 10:33:10.265396 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:10 crc kubenswrapper[4821]: I1125 10:33:10.265411 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:10 crc kubenswrapper[4821]: I1125 10:33:10.265431 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:10 crc kubenswrapper[4821]: I1125 10:33:10.265443 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:10Z","lastTransitionTime":"2025-11-25T10:33:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:10 crc kubenswrapper[4821]: I1125 10:33:10.270904 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6hbqd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f71b3a2-63cb-443c-9f41-39e57030746c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65473b0e63c2e9c7c09a2bb8f31396daf1d87c2c83d20ed3508ce2fe7358dda3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxdmq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://595f9ea75461f18e763d84799e368becc07b72acdc5337d20d6995bf53c3d27a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxdmq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6hbqd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:10Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:10 crc kubenswrapper[4821]: I1125 10:33:10.292023 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d9e58df-1755-4422-8d04-211f98c7c011\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea762f627aa69fb88a9bd82545fa677e9ccfa97f99a5e92c0aa490e619c0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a06b811e5ec52ce984a2403ba2c98332ba60466d08b3580ad9192fac6caef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0385f6ed9597974267f7c45547f6b6725b4ee83a2cc0508ff2d1c1ce8753ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b915b565f7fbb3e4e856e7f1035a519e82bfb3023660c121f6e3813de353bb92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18465f3a6916c36457a305ba5d568c9077f34f1d000532ed518f237d100008b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:10Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:10 crc kubenswrapper[4821]: I1125 10:33:10.311398 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25595ccd-762d-414a-9916-4fe38d478b07\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d4b2d8db4d13392d9789e161fd4b8f6eef983b79128ee584d72e4c06565bc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf780f791771ea43a8c8176d02fdfb5204406e023207299369ae46ceb342baaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0251519fd4262f2ea5c64959db49433afdafe36b7ed2404d660eeb7c365d2746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad5c205e0eeecfd43cbdf0eb1a334e3e8faa5e9a9e8dd202c397d5cbb36e081e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:10Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:10 crc kubenswrapper[4821]: I1125 10:33:10.326628 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:10Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:10 crc kubenswrapper[4821]: I1125 10:33:10.340224 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ldzbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b84e32ae-4897-4644-a656-6bd82012cc81\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ef433744761f26e8e1f6c0db123821a11923ffd774830744285cd4178d89533\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5pmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ldzbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:10Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:10 crc kubenswrapper[4821]: I1125 10:33:10.357309 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"954085fa-d03a-4e88-9dfd-566257e3558d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f00746930c4588e548f0e2b60d52bf448aff7401b676a7855cb0f4cce00d329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a14a078ba7d2aed95992b36502008f315f0d3a634b64ed49da70398b2562d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d7254196b3702e48083ae98e829b73c539adf69ed5b2bbdc62189d09ae1762\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d727430108977574365f0e9c04b44e03307bd7ff89cdbb1e48a6e916f504d6b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94c7949f077b1ca2b310bedb090aa72fc4bd7e8f2c8bf438096edfca8568f36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed2e7cad7bb24b5b56a1295098a6a7c230c8f9b9c29a88ef653ae8b6ac31667d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e6202b51ac13209f3a6de880638a6d142f79755311646c0216f11dfe8bbe3f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e6202b51ac13209f3a6de880638a6d142f79755311646c0216f11dfe8bbe3f7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T10:32:56Z\\\",\\\"message\\\":\\\"to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:56Z is after 2025-08-24T17:21:41Z]\\\\nI1125 10:32:56.048189 6453 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-console/console_TCP_cluster\\\\\\\", UUID:\\\\\\\"d7d7b270-1480-47f8-bdf9-690dbab310cb\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console/console\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]strin\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-mzhnw_openshift-ovn-kubernetes(954085fa-d03a-4e88-9dfd-566257e3558d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dbf499f4ae6161752abaa7a466a027c6f874016be90a534e69f07e3dd3f3d9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzhnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:10Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:10 crc kubenswrapper[4821]: I1125 10:33:10.365663 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-d4msn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f12cef3-7e40-4f61-836e-23cd0e578e71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s7lkv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s7lkv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:44Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-d4msn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:10Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:10 crc kubenswrapper[4821]: I1125 10:33:10.367336 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:10 crc kubenswrapper[4821]: I1125 10:33:10.367361 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:10 crc kubenswrapper[4821]: I1125 10:33:10.367369 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:10 crc kubenswrapper[4821]: I1125 10:33:10.367381 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:10 crc kubenswrapper[4821]: I1125 10:33:10.367390 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:10Z","lastTransitionTime":"2025-11-25T10:33:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:10 crc kubenswrapper[4821]: I1125 10:33:10.469620 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:10 crc kubenswrapper[4821]: I1125 10:33:10.469669 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:10 crc kubenswrapper[4821]: I1125 10:33:10.469684 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:10 crc kubenswrapper[4821]: I1125 10:33:10.469702 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:10 crc kubenswrapper[4821]: I1125 10:33:10.469716 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:10Z","lastTransitionTime":"2025-11-25T10:33:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:10 crc kubenswrapper[4821]: I1125 10:33:10.571895 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:10 crc kubenswrapper[4821]: I1125 10:33:10.571940 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:10 crc kubenswrapper[4821]: I1125 10:33:10.571949 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:10 crc kubenswrapper[4821]: I1125 10:33:10.571964 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:10 crc kubenswrapper[4821]: I1125 10:33:10.571974 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:10Z","lastTransitionTime":"2025-11-25T10:33:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:10 crc kubenswrapper[4821]: I1125 10:33:10.674860 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:10 crc kubenswrapper[4821]: I1125 10:33:10.674896 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:10 crc kubenswrapper[4821]: I1125 10:33:10.674906 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:10 crc kubenswrapper[4821]: I1125 10:33:10.675207 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:10 crc kubenswrapper[4821]: I1125 10:33:10.675231 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:10Z","lastTransitionTime":"2025-11-25T10:33:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:10 crc kubenswrapper[4821]: I1125 10:33:10.777743 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:10 crc kubenswrapper[4821]: I1125 10:33:10.777989 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:10 crc kubenswrapper[4821]: I1125 10:33:10.777998 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:10 crc kubenswrapper[4821]: I1125 10:33:10.778010 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:10 crc kubenswrapper[4821]: I1125 10:33:10.778018 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:10Z","lastTransitionTime":"2025-11-25T10:33:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:10 crc kubenswrapper[4821]: I1125 10:33:10.880454 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:10 crc kubenswrapper[4821]: I1125 10:33:10.880696 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:10 crc kubenswrapper[4821]: I1125 10:33:10.880796 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:10 crc kubenswrapper[4821]: I1125 10:33:10.880882 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:10 crc kubenswrapper[4821]: I1125 10:33:10.880953 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:10Z","lastTransitionTime":"2025-11-25T10:33:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:10 crc kubenswrapper[4821]: I1125 10:33:10.983074 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:10 crc kubenswrapper[4821]: I1125 10:33:10.983113 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:10 crc kubenswrapper[4821]: I1125 10:33:10.983125 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:10 crc kubenswrapper[4821]: I1125 10:33:10.983140 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:10 crc kubenswrapper[4821]: I1125 10:33:10.983151 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:10Z","lastTransitionTime":"2025-11-25T10:33:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:11 crc kubenswrapper[4821]: I1125 10:33:11.086559 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:11 crc kubenswrapper[4821]: I1125 10:33:11.086624 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:11 crc kubenswrapper[4821]: I1125 10:33:11.086644 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:11 crc kubenswrapper[4821]: I1125 10:33:11.086671 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:11 crc kubenswrapper[4821]: I1125 10:33:11.086692 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:11Z","lastTransitionTime":"2025-11-25T10:33:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:11 crc kubenswrapper[4821]: I1125 10:33:11.114210 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:33:11 crc kubenswrapper[4821]: I1125 10:33:11.114226 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:33:11 crc kubenswrapper[4821]: I1125 10:33:11.114314 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:33:11 crc kubenswrapper[4821]: E1125 10:33:11.114500 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:33:11 crc kubenswrapper[4821]: E1125 10:33:11.114588 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:33:11 crc kubenswrapper[4821]: E1125 10:33:11.115148 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:33:11 crc kubenswrapper[4821]: I1125 10:33:11.189061 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:11 crc kubenswrapper[4821]: I1125 10:33:11.189111 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:11 crc kubenswrapper[4821]: I1125 10:33:11.189214 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:11 crc kubenswrapper[4821]: I1125 10:33:11.189280 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:11 crc kubenswrapper[4821]: I1125 10:33:11.189300 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:11Z","lastTransitionTime":"2025-11-25T10:33:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:11 crc kubenswrapper[4821]: I1125 10:33:11.292489 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:11 crc kubenswrapper[4821]: I1125 10:33:11.292541 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:11 crc kubenswrapper[4821]: I1125 10:33:11.292554 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:11 crc kubenswrapper[4821]: I1125 10:33:11.292572 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:11 crc kubenswrapper[4821]: I1125 10:33:11.292586 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:11Z","lastTransitionTime":"2025-11-25T10:33:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:11 crc kubenswrapper[4821]: I1125 10:33:11.395025 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:11 crc kubenswrapper[4821]: I1125 10:33:11.395098 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:11 crc kubenswrapper[4821]: I1125 10:33:11.395109 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:11 crc kubenswrapper[4821]: I1125 10:33:11.395126 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:11 crc kubenswrapper[4821]: I1125 10:33:11.395136 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:11Z","lastTransitionTime":"2025-11-25T10:33:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:11 crc kubenswrapper[4821]: I1125 10:33:11.497492 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:11 crc kubenswrapper[4821]: I1125 10:33:11.497538 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:11 crc kubenswrapper[4821]: I1125 10:33:11.497547 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:11 crc kubenswrapper[4821]: I1125 10:33:11.497561 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:11 crc kubenswrapper[4821]: I1125 10:33:11.497571 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:11Z","lastTransitionTime":"2025-11-25T10:33:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:11 crc kubenswrapper[4821]: I1125 10:33:11.599443 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:11 crc kubenswrapper[4821]: I1125 10:33:11.599481 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:11 crc kubenswrapper[4821]: I1125 10:33:11.599491 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:11 crc kubenswrapper[4821]: I1125 10:33:11.599506 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:11 crc kubenswrapper[4821]: I1125 10:33:11.599522 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:11Z","lastTransitionTime":"2025-11-25T10:33:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:11 crc kubenswrapper[4821]: I1125 10:33:11.701423 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:11 crc kubenswrapper[4821]: I1125 10:33:11.701495 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:11 crc kubenswrapper[4821]: I1125 10:33:11.701508 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:11 crc kubenswrapper[4821]: I1125 10:33:11.701526 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:11 crc kubenswrapper[4821]: I1125 10:33:11.701537 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:11Z","lastTransitionTime":"2025-11-25T10:33:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:11 crc kubenswrapper[4821]: I1125 10:33:11.807106 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:11 crc kubenswrapper[4821]: I1125 10:33:11.807155 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:11 crc kubenswrapper[4821]: I1125 10:33:11.807182 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:11 crc kubenswrapper[4821]: I1125 10:33:11.807210 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:11 crc kubenswrapper[4821]: I1125 10:33:11.807223 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:11Z","lastTransitionTime":"2025-11-25T10:33:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:11 crc kubenswrapper[4821]: I1125 10:33:11.909814 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:11 crc kubenswrapper[4821]: I1125 10:33:11.909851 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:11 crc kubenswrapper[4821]: I1125 10:33:11.909861 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:11 crc kubenswrapper[4821]: I1125 10:33:11.909875 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:11 crc kubenswrapper[4821]: I1125 10:33:11.909883 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:11Z","lastTransitionTime":"2025-11-25T10:33:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:12 crc kubenswrapper[4821]: I1125 10:33:12.012714 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:12 crc kubenswrapper[4821]: I1125 10:33:12.012748 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:12 crc kubenswrapper[4821]: I1125 10:33:12.012757 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:12 crc kubenswrapper[4821]: I1125 10:33:12.012773 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:12 crc kubenswrapper[4821]: I1125 10:33:12.012782 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:12Z","lastTransitionTime":"2025-11-25T10:33:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:12 crc kubenswrapper[4821]: I1125 10:33:12.113880 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4msn" Nov 25 10:33:12 crc kubenswrapper[4821]: E1125 10:33:12.114082 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4msn" podUID="1f12cef3-7e40-4f61-836e-23cd0e578e71" Nov 25 10:33:12 crc kubenswrapper[4821]: I1125 10:33:12.115137 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:12 crc kubenswrapper[4821]: I1125 10:33:12.115184 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:12 crc kubenswrapper[4821]: I1125 10:33:12.115201 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:12 crc kubenswrapper[4821]: I1125 10:33:12.115217 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:12 crc kubenswrapper[4821]: I1125 10:33:12.115226 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:12Z","lastTransitionTime":"2025-11-25T10:33:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:12 crc kubenswrapper[4821]: I1125 10:33:12.217790 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:12 crc kubenswrapper[4821]: I1125 10:33:12.217832 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:12 crc kubenswrapper[4821]: I1125 10:33:12.217844 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:12 crc kubenswrapper[4821]: I1125 10:33:12.217862 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:12 crc kubenswrapper[4821]: I1125 10:33:12.217877 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:12Z","lastTransitionTime":"2025-11-25T10:33:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:12 crc kubenswrapper[4821]: I1125 10:33:12.320627 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:12 crc kubenswrapper[4821]: I1125 10:33:12.320669 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:12 crc kubenswrapper[4821]: I1125 10:33:12.320678 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:12 crc kubenswrapper[4821]: I1125 10:33:12.320691 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:12 crc kubenswrapper[4821]: I1125 10:33:12.320702 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:12Z","lastTransitionTime":"2025-11-25T10:33:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:12 crc kubenswrapper[4821]: I1125 10:33:12.423493 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:12 crc kubenswrapper[4821]: I1125 10:33:12.423522 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:12 crc kubenswrapper[4821]: I1125 10:33:12.423532 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:12 crc kubenswrapper[4821]: I1125 10:33:12.423547 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:12 crc kubenswrapper[4821]: I1125 10:33:12.423558 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:12Z","lastTransitionTime":"2025-11-25T10:33:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:12 crc kubenswrapper[4821]: I1125 10:33:12.526527 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:12 crc kubenswrapper[4821]: I1125 10:33:12.526588 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:12 crc kubenswrapper[4821]: I1125 10:33:12.526600 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:12 crc kubenswrapper[4821]: I1125 10:33:12.526619 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:12 crc kubenswrapper[4821]: I1125 10:33:12.526633 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:12Z","lastTransitionTime":"2025-11-25T10:33:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:12 crc kubenswrapper[4821]: I1125 10:33:12.628596 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:12 crc kubenswrapper[4821]: I1125 10:33:12.628639 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:12 crc kubenswrapper[4821]: I1125 10:33:12.628652 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:12 crc kubenswrapper[4821]: I1125 10:33:12.628667 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:12 crc kubenswrapper[4821]: I1125 10:33:12.628678 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:12Z","lastTransitionTime":"2025-11-25T10:33:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:12 crc kubenswrapper[4821]: I1125 10:33:12.731205 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:12 crc kubenswrapper[4821]: I1125 10:33:12.731293 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:12 crc kubenswrapper[4821]: I1125 10:33:12.731305 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:12 crc kubenswrapper[4821]: I1125 10:33:12.731348 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:12 crc kubenswrapper[4821]: I1125 10:33:12.731357 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:12Z","lastTransitionTime":"2025-11-25T10:33:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:12 crc kubenswrapper[4821]: I1125 10:33:12.833577 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:12 crc kubenswrapper[4821]: I1125 10:33:12.833617 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:12 crc kubenswrapper[4821]: I1125 10:33:12.833629 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:12 crc kubenswrapper[4821]: I1125 10:33:12.833646 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:12 crc kubenswrapper[4821]: I1125 10:33:12.833658 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:12Z","lastTransitionTime":"2025-11-25T10:33:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:12 crc kubenswrapper[4821]: I1125 10:33:12.936075 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:12 crc kubenswrapper[4821]: I1125 10:33:12.936131 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:12 crc kubenswrapper[4821]: I1125 10:33:12.936142 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:12 crc kubenswrapper[4821]: I1125 10:33:12.936183 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:12 crc kubenswrapper[4821]: I1125 10:33:12.936195 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:12Z","lastTransitionTime":"2025-11-25T10:33:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:13 crc kubenswrapper[4821]: I1125 10:33:13.038975 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:13 crc kubenswrapper[4821]: I1125 10:33:13.039016 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:13 crc kubenswrapper[4821]: I1125 10:33:13.039028 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:13 crc kubenswrapper[4821]: I1125 10:33:13.039044 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:13 crc kubenswrapper[4821]: I1125 10:33:13.039059 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:13Z","lastTransitionTime":"2025-11-25T10:33:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:13 crc kubenswrapper[4821]: I1125 10:33:13.113755 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:33:13 crc kubenswrapper[4821]: I1125 10:33:13.113754 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:33:13 crc kubenswrapper[4821]: E1125 10:33:13.113889 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:33:13 crc kubenswrapper[4821]: E1125 10:33:13.113963 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:33:13 crc kubenswrapper[4821]: I1125 10:33:13.113773 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:33:13 crc kubenswrapper[4821]: E1125 10:33:13.114028 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:33:13 crc kubenswrapper[4821]: I1125 10:33:13.142080 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:13 crc kubenswrapper[4821]: I1125 10:33:13.142192 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:13 crc kubenswrapper[4821]: I1125 10:33:13.142256 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:13 crc kubenswrapper[4821]: I1125 10:33:13.142296 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:13 crc kubenswrapper[4821]: I1125 10:33:13.142322 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:13Z","lastTransitionTime":"2025-11-25T10:33:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:13 crc kubenswrapper[4821]: I1125 10:33:13.245319 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:13 crc kubenswrapper[4821]: I1125 10:33:13.245362 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:13 crc kubenswrapper[4821]: I1125 10:33:13.245372 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:13 crc kubenswrapper[4821]: I1125 10:33:13.245390 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:13 crc kubenswrapper[4821]: I1125 10:33:13.245401 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:13Z","lastTransitionTime":"2025-11-25T10:33:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:13 crc kubenswrapper[4821]: I1125 10:33:13.347688 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:13 crc kubenswrapper[4821]: I1125 10:33:13.347729 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:13 crc kubenswrapper[4821]: I1125 10:33:13.347740 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:13 crc kubenswrapper[4821]: I1125 10:33:13.347757 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:13 crc kubenswrapper[4821]: I1125 10:33:13.347767 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:13Z","lastTransitionTime":"2025-11-25T10:33:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:13 crc kubenswrapper[4821]: I1125 10:33:13.450042 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:13 crc kubenswrapper[4821]: I1125 10:33:13.450139 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:13 crc kubenswrapper[4821]: I1125 10:33:13.450156 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:13 crc kubenswrapper[4821]: I1125 10:33:13.450207 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:13 crc kubenswrapper[4821]: I1125 10:33:13.450224 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:13Z","lastTransitionTime":"2025-11-25T10:33:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:13 crc kubenswrapper[4821]: I1125 10:33:13.552862 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:13 crc kubenswrapper[4821]: I1125 10:33:13.552901 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:13 crc kubenswrapper[4821]: I1125 10:33:13.552910 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:13 crc kubenswrapper[4821]: I1125 10:33:13.552925 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:13 crc kubenswrapper[4821]: I1125 10:33:13.552935 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:13Z","lastTransitionTime":"2025-11-25T10:33:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:13 crc kubenswrapper[4821]: I1125 10:33:13.655374 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:13 crc kubenswrapper[4821]: I1125 10:33:13.655412 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:13 crc kubenswrapper[4821]: I1125 10:33:13.655421 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:13 crc kubenswrapper[4821]: I1125 10:33:13.655435 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:13 crc kubenswrapper[4821]: I1125 10:33:13.655445 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:13Z","lastTransitionTime":"2025-11-25T10:33:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:13 crc kubenswrapper[4821]: I1125 10:33:13.757628 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:13 crc kubenswrapper[4821]: I1125 10:33:13.757690 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:13 crc kubenswrapper[4821]: I1125 10:33:13.757706 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:13 crc kubenswrapper[4821]: I1125 10:33:13.757725 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:13 crc kubenswrapper[4821]: I1125 10:33:13.757739 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:13Z","lastTransitionTime":"2025-11-25T10:33:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:13 crc kubenswrapper[4821]: I1125 10:33:13.860248 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:13 crc kubenswrapper[4821]: I1125 10:33:13.860299 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:13 crc kubenswrapper[4821]: I1125 10:33:13.860313 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:13 crc kubenswrapper[4821]: I1125 10:33:13.860332 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:13 crc kubenswrapper[4821]: I1125 10:33:13.860343 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:13Z","lastTransitionTime":"2025-11-25T10:33:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:13 crc kubenswrapper[4821]: I1125 10:33:13.962809 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:13 crc kubenswrapper[4821]: I1125 10:33:13.962856 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:13 crc kubenswrapper[4821]: I1125 10:33:13.962867 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:13 crc kubenswrapper[4821]: I1125 10:33:13.962884 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:13 crc kubenswrapper[4821]: I1125 10:33:13.962896 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:13Z","lastTransitionTime":"2025-11-25T10:33:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:14 crc kubenswrapper[4821]: I1125 10:33:14.064970 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:14 crc kubenswrapper[4821]: I1125 10:33:14.065004 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:14 crc kubenswrapper[4821]: I1125 10:33:14.065012 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:14 crc kubenswrapper[4821]: I1125 10:33:14.065025 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:14 crc kubenswrapper[4821]: I1125 10:33:14.065034 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:14Z","lastTransitionTime":"2025-11-25T10:33:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:14 crc kubenswrapper[4821]: I1125 10:33:14.113423 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4msn" Nov 25 10:33:14 crc kubenswrapper[4821]: E1125 10:33:14.113569 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4msn" podUID="1f12cef3-7e40-4f61-836e-23cd0e578e71" Nov 25 10:33:14 crc kubenswrapper[4821]: I1125 10:33:14.174317 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:14 crc kubenswrapper[4821]: I1125 10:33:14.174389 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:14 crc kubenswrapper[4821]: I1125 10:33:14.174407 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:14 crc kubenswrapper[4821]: I1125 10:33:14.174424 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:14 crc kubenswrapper[4821]: I1125 10:33:14.174433 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:14Z","lastTransitionTime":"2025-11-25T10:33:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:14 crc kubenswrapper[4821]: I1125 10:33:14.277430 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:14 crc kubenswrapper[4821]: I1125 10:33:14.277481 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:14 crc kubenswrapper[4821]: I1125 10:33:14.277491 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:14 crc kubenswrapper[4821]: I1125 10:33:14.277506 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:14 crc kubenswrapper[4821]: I1125 10:33:14.277515 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:14Z","lastTransitionTime":"2025-11-25T10:33:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:14 crc kubenswrapper[4821]: I1125 10:33:14.379377 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:14 crc kubenswrapper[4821]: I1125 10:33:14.379424 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:14 crc kubenswrapper[4821]: I1125 10:33:14.379438 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:14 crc kubenswrapper[4821]: I1125 10:33:14.379455 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:14 crc kubenswrapper[4821]: I1125 10:33:14.379466 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:14Z","lastTransitionTime":"2025-11-25T10:33:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:14 crc kubenswrapper[4821]: I1125 10:33:14.482963 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:14 crc kubenswrapper[4821]: I1125 10:33:14.483000 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:14 crc kubenswrapper[4821]: I1125 10:33:14.483011 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:14 crc kubenswrapper[4821]: I1125 10:33:14.483028 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:14 crc kubenswrapper[4821]: I1125 10:33:14.483039 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:14Z","lastTransitionTime":"2025-11-25T10:33:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:14 crc kubenswrapper[4821]: I1125 10:33:14.585881 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:14 crc kubenswrapper[4821]: I1125 10:33:14.585913 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:14 crc kubenswrapper[4821]: I1125 10:33:14.585921 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:14 crc kubenswrapper[4821]: I1125 10:33:14.585934 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:14 crc kubenswrapper[4821]: I1125 10:33:14.585943 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:14Z","lastTransitionTime":"2025-11-25T10:33:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:14 crc kubenswrapper[4821]: I1125 10:33:14.688043 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:14 crc kubenswrapper[4821]: I1125 10:33:14.688104 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:14 crc kubenswrapper[4821]: I1125 10:33:14.688117 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:14 crc kubenswrapper[4821]: I1125 10:33:14.688134 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:14 crc kubenswrapper[4821]: I1125 10:33:14.688390 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:14Z","lastTransitionTime":"2025-11-25T10:33:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:14 crc kubenswrapper[4821]: I1125 10:33:14.790746 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:14 crc kubenswrapper[4821]: I1125 10:33:14.790796 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:14 crc kubenswrapper[4821]: I1125 10:33:14.790805 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:14 crc kubenswrapper[4821]: I1125 10:33:14.790819 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:14 crc kubenswrapper[4821]: I1125 10:33:14.790828 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:14Z","lastTransitionTime":"2025-11-25T10:33:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:14 crc kubenswrapper[4821]: I1125 10:33:14.892948 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:14 crc kubenswrapper[4821]: I1125 10:33:14.892987 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:14 crc kubenswrapper[4821]: I1125 10:33:14.893020 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:14 crc kubenswrapper[4821]: I1125 10:33:14.893040 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:14 crc kubenswrapper[4821]: I1125 10:33:14.893051 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:14Z","lastTransitionTime":"2025-11-25T10:33:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:14 crc kubenswrapper[4821]: I1125 10:33:14.995548 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:14 crc kubenswrapper[4821]: I1125 10:33:14.995610 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:14 crc kubenswrapper[4821]: I1125 10:33:14.995620 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:14 crc kubenswrapper[4821]: I1125 10:33:14.995640 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:14 crc kubenswrapper[4821]: I1125 10:33:14.995649 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:14Z","lastTransitionTime":"2025-11-25T10:33:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:15 crc kubenswrapper[4821]: I1125 10:33:15.097834 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:15 crc kubenswrapper[4821]: I1125 10:33:15.097928 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:15 crc kubenswrapper[4821]: I1125 10:33:15.097944 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:15 crc kubenswrapper[4821]: I1125 10:33:15.097985 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:15 crc kubenswrapper[4821]: I1125 10:33:15.097997 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:15Z","lastTransitionTime":"2025-11-25T10:33:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:15 crc kubenswrapper[4821]: I1125 10:33:15.113536 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:33:15 crc kubenswrapper[4821]: I1125 10:33:15.113607 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:33:15 crc kubenswrapper[4821]: E1125 10:33:15.113688 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:33:15 crc kubenswrapper[4821]: I1125 10:33:15.113541 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:33:15 crc kubenswrapper[4821]: E1125 10:33:15.114264 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:33:15 crc kubenswrapper[4821]: E1125 10:33:15.114351 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:33:15 crc kubenswrapper[4821]: I1125 10:33:15.114770 4821 scope.go:117] "RemoveContainer" containerID="7e6202b51ac13209f3a6de880638a6d142f79755311646c0216f11dfe8bbe3f7" Nov 25 10:33:15 crc kubenswrapper[4821]: E1125 10:33:15.115272 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-mzhnw_openshift-ovn-kubernetes(954085fa-d03a-4e88-9dfd-566257e3558d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" podUID="954085fa-d03a-4e88-9dfd-566257e3558d" Nov 25 10:33:15 crc kubenswrapper[4821]: I1125 10:33:15.200305 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:15 crc kubenswrapper[4821]: I1125 10:33:15.200334 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:15 crc kubenswrapper[4821]: I1125 10:33:15.200342 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:15 crc kubenswrapper[4821]: I1125 10:33:15.200355 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:15 crc kubenswrapper[4821]: I1125 10:33:15.200364 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:15Z","lastTransitionTime":"2025-11-25T10:33:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:15 crc kubenswrapper[4821]: I1125 10:33:15.303138 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:15 crc kubenswrapper[4821]: I1125 10:33:15.303263 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:15 crc kubenswrapper[4821]: I1125 10:33:15.303280 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:15 crc kubenswrapper[4821]: I1125 10:33:15.303310 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:15 crc kubenswrapper[4821]: I1125 10:33:15.303329 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:15Z","lastTransitionTime":"2025-11-25T10:33:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:15 crc kubenswrapper[4821]: I1125 10:33:15.408843 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:15 crc kubenswrapper[4821]: I1125 10:33:15.408898 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:15 crc kubenswrapper[4821]: I1125 10:33:15.408912 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:15 crc kubenswrapper[4821]: I1125 10:33:15.408929 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:15 crc kubenswrapper[4821]: I1125 10:33:15.408941 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:15Z","lastTransitionTime":"2025-11-25T10:33:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:15 crc kubenswrapper[4821]: I1125 10:33:15.512152 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:15 crc kubenswrapper[4821]: I1125 10:33:15.512208 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:15 crc kubenswrapper[4821]: I1125 10:33:15.512241 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:15 crc kubenswrapper[4821]: I1125 10:33:15.512274 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:15 crc kubenswrapper[4821]: I1125 10:33:15.512285 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:15Z","lastTransitionTime":"2025-11-25T10:33:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:15 crc kubenswrapper[4821]: I1125 10:33:15.615369 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:15 crc kubenswrapper[4821]: I1125 10:33:15.615409 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:15 crc kubenswrapper[4821]: I1125 10:33:15.615419 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:15 crc kubenswrapper[4821]: I1125 10:33:15.615433 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:15 crc kubenswrapper[4821]: I1125 10:33:15.615442 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:15Z","lastTransitionTime":"2025-11-25T10:33:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:15 crc kubenswrapper[4821]: I1125 10:33:15.717655 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:15 crc kubenswrapper[4821]: I1125 10:33:15.717694 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:15 crc kubenswrapper[4821]: I1125 10:33:15.717706 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:15 crc kubenswrapper[4821]: I1125 10:33:15.717721 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:15 crc kubenswrapper[4821]: I1125 10:33:15.717733 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:15Z","lastTransitionTime":"2025-11-25T10:33:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:15 crc kubenswrapper[4821]: I1125 10:33:15.820157 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:15 crc kubenswrapper[4821]: I1125 10:33:15.820215 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:15 crc kubenswrapper[4821]: I1125 10:33:15.820226 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:15 crc kubenswrapper[4821]: I1125 10:33:15.820245 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:15 crc kubenswrapper[4821]: I1125 10:33:15.820258 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:15Z","lastTransitionTime":"2025-11-25T10:33:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:15 crc kubenswrapper[4821]: I1125 10:33:15.922771 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:15 crc kubenswrapper[4821]: I1125 10:33:15.922821 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:15 crc kubenswrapper[4821]: I1125 10:33:15.922832 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:15 crc kubenswrapper[4821]: I1125 10:33:15.922853 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:15 crc kubenswrapper[4821]: I1125 10:33:15.922866 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:15Z","lastTransitionTime":"2025-11-25T10:33:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.025534 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.025564 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.025573 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.025587 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.025595 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:16Z","lastTransitionTime":"2025-11-25T10:33:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.114263 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4msn" Nov 25 10:33:16 crc kubenswrapper[4821]: E1125 10:33:16.114401 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4msn" podUID="1f12cef3-7e40-4f61-836e-23cd0e578e71" Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.127497 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.127532 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.127541 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.127556 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.127565 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:16Z","lastTransitionTime":"2025-11-25T10:33:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.229618 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.229661 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.229669 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.229684 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.229693 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:16Z","lastTransitionTime":"2025-11-25T10:33:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.331572 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.331611 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.331619 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.331633 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.331644 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:16Z","lastTransitionTime":"2025-11-25T10:33:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.362133 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1f12cef3-7e40-4f61-836e-23cd0e578e71-metrics-certs\") pod \"network-metrics-daemon-d4msn\" (UID: \"1f12cef3-7e40-4f61-836e-23cd0e578e71\") " pod="openshift-multus/network-metrics-daemon-d4msn" Nov 25 10:33:16 crc kubenswrapper[4821]: E1125 10:33:16.362297 4821 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 10:33:16 crc kubenswrapper[4821]: E1125 10:33:16.362388 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1f12cef3-7e40-4f61-836e-23cd0e578e71-metrics-certs podName:1f12cef3-7e40-4f61-836e-23cd0e578e71 nodeName:}" failed. No retries permitted until 2025-11-25 10:33:48.362366354 +0000 UTC m=+98.898686271 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/1f12cef3-7e40-4f61-836e-23cd0e578e71-metrics-certs") pod "network-metrics-daemon-d4msn" (UID: "1f12cef3-7e40-4f61-836e-23cd0e578e71") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.433580 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.433612 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.433627 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.433640 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.433650 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:16Z","lastTransitionTime":"2025-11-25T10:33:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.535907 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.535940 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.535949 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.535963 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.535973 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:16Z","lastTransitionTime":"2025-11-25T10:33:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.638151 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.638228 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.638244 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.638268 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.638287 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:16Z","lastTransitionTime":"2025-11-25T10:33:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.739995 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.740033 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.740044 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.740061 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.740073 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:16Z","lastTransitionTime":"2025-11-25T10:33:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.799430 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.799471 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.799481 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.799496 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.799505 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:16Z","lastTransitionTime":"2025-11-25T10:33:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:16 crc kubenswrapper[4821]: E1125 10:33:16.815060 4821 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a5b2b070-39b9-412c-90f1-c7ed2e9e0950\\\",\\\"systemUUID\\\":\\\"ea86e80f-41b7-4190-ad7e-d8c549d535de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:16Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.818648 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.818747 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.818758 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.818772 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.818781 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:16Z","lastTransitionTime":"2025-11-25T10:33:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:16 crc kubenswrapper[4821]: E1125 10:33:16.829296 4821 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a5b2b070-39b9-412c-90f1-c7ed2e9e0950\\\",\\\"systemUUID\\\":\\\"ea86e80f-41b7-4190-ad7e-d8c549d535de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:16Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.832428 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.832495 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.832506 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.832524 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.832534 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:16Z","lastTransitionTime":"2025-11-25T10:33:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:16 crc kubenswrapper[4821]: E1125 10:33:16.843462 4821 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a5b2b070-39b9-412c-90f1-c7ed2e9e0950\\\",\\\"systemUUID\\\":\\\"ea86e80f-41b7-4190-ad7e-d8c549d535de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:16Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.846642 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.846692 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.846704 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.846718 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.846727 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:16Z","lastTransitionTime":"2025-11-25T10:33:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:16 crc kubenswrapper[4821]: E1125 10:33:16.858502 4821 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a5b2b070-39b9-412c-90f1-c7ed2e9e0950\\\",\\\"systemUUID\\\":\\\"ea86e80f-41b7-4190-ad7e-d8c549d535de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:16Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.861636 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.861692 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.861702 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.861715 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.861739 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:16Z","lastTransitionTime":"2025-11-25T10:33:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:16 crc kubenswrapper[4821]: E1125 10:33:16.872200 4821 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a5b2b070-39b9-412c-90f1-c7ed2e9e0950\\\",\\\"systemUUID\\\":\\\"ea86e80f-41b7-4190-ad7e-d8c549d535de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:16Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:16 crc kubenswrapper[4821]: E1125 10:33:16.872349 4821 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.874127 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.874178 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.874191 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.874211 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.874224 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:16Z","lastTransitionTime":"2025-11-25T10:33:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.976346 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.976424 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.976435 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.976450 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:16 crc kubenswrapper[4821]: I1125 10:33:16.976461 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:16Z","lastTransitionTime":"2025-11-25T10:33:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:17 crc kubenswrapper[4821]: I1125 10:33:17.079143 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:17 crc kubenswrapper[4821]: I1125 10:33:17.079198 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:17 crc kubenswrapper[4821]: I1125 10:33:17.079207 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:17 crc kubenswrapper[4821]: I1125 10:33:17.079232 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:17 crc kubenswrapper[4821]: I1125 10:33:17.079241 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:17Z","lastTransitionTime":"2025-11-25T10:33:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:17 crc kubenswrapper[4821]: I1125 10:33:17.113755 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:33:17 crc kubenswrapper[4821]: I1125 10:33:17.113790 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:33:17 crc kubenswrapper[4821]: E1125 10:33:17.113886 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:33:17 crc kubenswrapper[4821]: I1125 10:33:17.113765 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:33:17 crc kubenswrapper[4821]: E1125 10:33:17.114016 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:33:17 crc kubenswrapper[4821]: E1125 10:33:17.114059 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:33:17 crc kubenswrapper[4821]: I1125 10:33:17.181638 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:17 crc kubenswrapper[4821]: I1125 10:33:17.181674 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:17 crc kubenswrapper[4821]: I1125 10:33:17.181688 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:17 crc kubenswrapper[4821]: I1125 10:33:17.181704 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:17 crc kubenswrapper[4821]: I1125 10:33:17.181715 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:17Z","lastTransitionTime":"2025-11-25T10:33:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:17 crc kubenswrapper[4821]: I1125 10:33:17.283753 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:17 crc kubenswrapper[4821]: I1125 10:33:17.283801 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:17 crc kubenswrapper[4821]: I1125 10:33:17.283813 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:17 crc kubenswrapper[4821]: I1125 10:33:17.283831 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:17 crc kubenswrapper[4821]: I1125 10:33:17.283843 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:17Z","lastTransitionTime":"2025-11-25T10:33:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:17 crc kubenswrapper[4821]: I1125 10:33:17.385894 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:17 crc kubenswrapper[4821]: I1125 10:33:17.385931 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:17 crc kubenswrapper[4821]: I1125 10:33:17.385941 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:17 crc kubenswrapper[4821]: I1125 10:33:17.385957 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:17 crc kubenswrapper[4821]: I1125 10:33:17.385967 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:17Z","lastTransitionTime":"2025-11-25T10:33:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:17 crc kubenswrapper[4821]: I1125 10:33:17.488608 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:17 crc kubenswrapper[4821]: I1125 10:33:17.488642 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:17 crc kubenswrapper[4821]: I1125 10:33:17.488650 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:17 crc kubenswrapper[4821]: I1125 10:33:17.488664 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:17 crc kubenswrapper[4821]: I1125 10:33:17.488673 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:17Z","lastTransitionTime":"2025-11-25T10:33:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:17 crc kubenswrapper[4821]: I1125 10:33:17.605990 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:17 crc kubenswrapper[4821]: I1125 10:33:17.606028 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:17 crc kubenswrapper[4821]: I1125 10:33:17.606041 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:17 crc kubenswrapper[4821]: I1125 10:33:17.606057 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:17 crc kubenswrapper[4821]: I1125 10:33:17.606070 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:17Z","lastTransitionTime":"2025-11-25T10:33:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:17 crc kubenswrapper[4821]: I1125 10:33:17.708503 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:17 crc kubenswrapper[4821]: I1125 10:33:17.708611 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:17 crc kubenswrapper[4821]: I1125 10:33:17.708628 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:17 crc kubenswrapper[4821]: I1125 10:33:17.708649 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:17 crc kubenswrapper[4821]: I1125 10:33:17.708660 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:17Z","lastTransitionTime":"2025-11-25T10:33:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:17 crc kubenswrapper[4821]: I1125 10:33:17.810780 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:17 crc kubenswrapper[4821]: I1125 10:33:17.810809 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:17 crc kubenswrapper[4821]: I1125 10:33:17.810818 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:17 crc kubenswrapper[4821]: I1125 10:33:17.810830 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:17 crc kubenswrapper[4821]: I1125 10:33:17.810838 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:17Z","lastTransitionTime":"2025-11-25T10:33:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:17 crc kubenswrapper[4821]: I1125 10:33:17.913383 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:17 crc kubenswrapper[4821]: I1125 10:33:17.913423 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:17 crc kubenswrapper[4821]: I1125 10:33:17.913434 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:17 crc kubenswrapper[4821]: I1125 10:33:17.913452 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:17 crc kubenswrapper[4821]: I1125 10:33:17.913463 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:17Z","lastTransitionTime":"2025-11-25T10:33:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:18 crc kubenswrapper[4821]: I1125 10:33:18.015624 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:18 crc kubenswrapper[4821]: I1125 10:33:18.015667 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:18 crc kubenswrapper[4821]: I1125 10:33:18.015677 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:18 crc kubenswrapper[4821]: I1125 10:33:18.015691 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:18 crc kubenswrapper[4821]: I1125 10:33:18.015701 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:18Z","lastTransitionTime":"2025-11-25T10:33:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:18 crc kubenswrapper[4821]: I1125 10:33:18.113876 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4msn" Nov 25 10:33:18 crc kubenswrapper[4821]: E1125 10:33:18.114047 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4msn" podUID="1f12cef3-7e40-4f61-836e-23cd0e578e71" Nov 25 10:33:18 crc kubenswrapper[4821]: I1125 10:33:18.118681 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:18 crc kubenswrapper[4821]: I1125 10:33:18.118703 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:18 crc kubenswrapper[4821]: I1125 10:33:18.118712 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:18 crc kubenswrapper[4821]: I1125 10:33:18.118788 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:18 crc kubenswrapper[4821]: I1125 10:33:18.118801 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:18Z","lastTransitionTime":"2025-11-25T10:33:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:18 crc kubenswrapper[4821]: I1125 10:33:18.221462 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:18 crc kubenswrapper[4821]: I1125 10:33:18.221506 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:18 crc kubenswrapper[4821]: I1125 10:33:18.221518 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:18 crc kubenswrapper[4821]: I1125 10:33:18.221536 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:18 crc kubenswrapper[4821]: I1125 10:33:18.221547 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:18Z","lastTransitionTime":"2025-11-25T10:33:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:18 crc kubenswrapper[4821]: I1125 10:33:18.324711 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:18 crc kubenswrapper[4821]: I1125 10:33:18.324746 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:18 crc kubenswrapper[4821]: I1125 10:33:18.324754 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:18 crc kubenswrapper[4821]: I1125 10:33:18.324769 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:18 crc kubenswrapper[4821]: I1125 10:33:18.324778 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:18Z","lastTransitionTime":"2025-11-25T10:33:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:18 crc kubenswrapper[4821]: I1125 10:33:18.426444 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:18 crc kubenswrapper[4821]: I1125 10:33:18.426493 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:18 crc kubenswrapper[4821]: I1125 10:33:18.426502 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:18 crc kubenswrapper[4821]: I1125 10:33:18.426518 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:18 crc kubenswrapper[4821]: I1125 10:33:18.426532 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:18Z","lastTransitionTime":"2025-11-25T10:33:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:18 crc kubenswrapper[4821]: I1125 10:33:18.528648 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:18 crc kubenswrapper[4821]: I1125 10:33:18.528680 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:18 crc kubenswrapper[4821]: I1125 10:33:18.528688 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:18 crc kubenswrapper[4821]: I1125 10:33:18.528701 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:18 crc kubenswrapper[4821]: I1125 10:33:18.528710 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:18Z","lastTransitionTime":"2025-11-25T10:33:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:18 crc kubenswrapper[4821]: I1125 10:33:18.631373 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:18 crc kubenswrapper[4821]: I1125 10:33:18.631413 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:18 crc kubenswrapper[4821]: I1125 10:33:18.631422 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:18 crc kubenswrapper[4821]: I1125 10:33:18.631436 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:18 crc kubenswrapper[4821]: I1125 10:33:18.631447 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:18Z","lastTransitionTime":"2025-11-25T10:33:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:18 crc kubenswrapper[4821]: I1125 10:33:18.734043 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:18 crc kubenswrapper[4821]: I1125 10:33:18.734084 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:18 crc kubenswrapper[4821]: I1125 10:33:18.734095 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:18 crc kubenswrapper[4821]: I1125 10:33:18.734112 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:18 crc kubenswrapper[4821]: I1125 10:33:18.734124 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:18Z","lastTransitionTime":"2025-11-25T10:33:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:18 crc kubenswrapper[4821]: I1125 10:33:18.836549 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:18 crc kubenswrapper[4821]: I1125 10:33:18.836580 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:18 crc kubenswrapper[4821]: I1125 10:33:18.836591 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:18 crc kubenswrapper[4821]: I1125 10:33:18.836608 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:18 crc kubenswrapper[4821]: I1125 10:33:18.836620 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:18Z","lastTransitionTime":"2025-11-25T10:33:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:18 crc kubenswrapper[4821]: I1125 10:33:18.939003 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:18 crc kubenswrapper[4821]: I1125 10:33:18.939058 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:18 crc kubenswrapper[4821]: I1125 10:33:18.939068 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:18 crc kubenswrapper[4821]: I1125 10:33:18.939083 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:18 crc kubenswrapper[4821]: I1125 10:33:18.939092 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:18Z","lastTransitionTime":"2025-11-25T10:33:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:19 crc kubenswrapper[4821]: I1125 10:33:19.041965 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:19 crc kubenswrapper[4821]: I1125 10:33:19.041995 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:19 crc kubenswrapper[4821]: I1125 10:33:19.042005 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:19 crc kubenswrapper[4821]: I1125 10:33:19.042020 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:19 crc kubenswrapper[4821]: I1125 10:33:19.042031 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:19Z","lastTransitionTime":"2025-11-25T10:33:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:19 crc kubenswrapper[4821]: I1125 10:33:19.114268 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:33:19 crc kubenswrapper[4821]: E1125 10:33:19.114401 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:33:19 crc kubenswrapper[4821]: I1125 10:33:19.114602 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:33:19 crc kubenswrapper[4821]: E1125 10:33:19.114669 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:33:19 crc kubenswrapper[4821]: I1125 10:33:19.114811 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:33:19 crc kubenswrapper[4821]: E1125 10:33:19.114874 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:33:19 crc kubenswrapper[4821]: I1125 10:33:19.144824 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:19 crc kubenswrapper[4821]: I1125 10:33:19.144889 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:19 crc kubenswrapper[4821]: I1125 10:33:19.144909 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:19 crc kubenswrapper[4821]: I1125 10:33:19.144932 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:19 crc kubenswrapper[4821]: I1125 10:33:19.144949 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:19Z","lastTransitionTime":"2025-11-25T10:33:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:19 crc kubenswrapper[4821]: I1125 10:33:19.248076 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:19 crc kubenswrapper[4821]: I1125 10:33:19.248115 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:19 crc kubenswrapper[4821]: I1125 10:33:19.248126 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:19 crc kubenswrapper[4821]: I1125 10:33:19.248141 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:19 crc kubenswrapper[4821]: I1125 10:33:19.248150 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:19Z","lastTransitionTime":"2025-11-25T10:33:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:19 crc kubenswrapper[4821]: I1125 10:33:19.350459 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:19 crc kubenswrapper[4821]: I1125 10:33:19.350508 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:19 crc kubenswrapper[4821]: I1125 10:33:19.350520 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:19 crc kubenswrapper[4821]: I1125 10:33:19.350540 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:19 crc kubenswrapper[4821]: I1125 10:33:19.350551 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:19Z","lastTransitionTime":"2025-11-25T10:33:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:19 crc kubenswrapper[4821]: I1125 10:33:19.452676 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:19 crc kubenswrapper[4821]: I1125 10:33:19.452707 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:19 crc kubenswrapper[4821]: I1125 10:33:19.452716 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:19 crc kubenswrapper[4821]: I1125 10:33:19.452730 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:19 crc kubenswrapper[4821]: I1125 10:33:19.452739 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:19Z","lastTransitionTime":"2025-11-25T10:33:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:19 crc kubenswrapper[4821]: I1125 10:33:19.554883 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:19 crc kubenswrapper[4821]: I1125 10:33:19.554921 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:19 crc kubenswrapper[4821]: I1125 10:33:19.554932 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:19 crc kubenswrapper[4821]: I1125 10:33:19.554951 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:19 crc kubenswrapper[4821]: I1125 10:33:19.554964 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:19Z","lastTransitionTime":"2025-11-25T10:33:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:19 crc kubenswrapper[4821]: I1125 10:33:19.656929 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:19 crc kubenswrapper[4821]: I1125 10:33:19.656978 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:19 crc kubenswrapper[4821]: I1125 10:33:19.656990 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:19 crc kubenswrapper[4821]: I1125 10:33:19.657010 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:19 crc kubenswrapper[4821]: I1125 10:33:19.657021 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:19Z","lastTransitionTime":"2025-11-25T10:33:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:19 crc kubenswrapper[4821]: I1125 10:33:19.759379 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:19 crc kubenswrapper[4821]: I1125 10:33:19.759605 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:19 crc kubenswrapper[4821]: I1125 10:33:19.759673 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:19 crc kubenswrapper[4821]: I1125 10:33:19.759741 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:19 crc kubenswrapper[4821]: I1125 10:33:19.759806 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:19Z","lastTransitionTime":"2025-11-25T10:33:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:19 crc kubenswrapper[4821]: I1125 10:33:19.862232 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:19 crc kubenswrapper[4821]: I1125 10:33:19.862491 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:19 crc kubenswrapper[4821]: I1125 10:33:19.862564 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:19 crc kubenswrapper[4821]: I1125 10:33:19.862630 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:19 crc kubenswrapper[4821]: I1125 10:33:19.862690 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:19Z","lastTransitionTime":"2025-11-25T10:33:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:19 crc kubenswrapper[4821]: I1125 10:33:19.965649 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:19 crc kubenswrapper[4821]: I1125 10:33:19.965708 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:19 crc kubenswrapper[4821]: I1125 10:33:19.965725 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:19 crc kubenswrapper[4821]: I1125 10:33:19.965748 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:19 crc kubenswrapper[4821]: I1125 10:33:19.965763 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:19Z","lastTransitionTime":"2025-11-25T10:33:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.067710 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.067739 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.067747 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.067760 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.067768 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:20Z","lastTransitionTime":"2025-11-25T10:33:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.114485 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4msn" Nov 25 10:33:20 crc kubenswrapper[4821]: E1125 10:33:20.115087 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4msn" podUID="1f12cef3-7e40-4f61-836e-23cd0e578e71" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.130949 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06f949936bac99cfbf89d720c9549c074351d9077bd963fa024d87004c6be23e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:20Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.144073 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384d8ba58264ab89796d3b53afb0c4c1be33cb4fba719a8beef4afc2be7d66a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:20Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.154950 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zlk96" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d55aa91-9cb9-4c95-bc27-6fec05684c10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4521b5baee45ffc9c7661c781c0b3de458dff325ad525dde283d46af71c0cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vkd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zlk96\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:20Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.164670 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6hbqd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f71b3a2-63cb-443c-9f41-39e57030746c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65473b0e63c2e9c7c09a2bb8f31396daf1d87c2c83d20ed3508ce2fe7358dda3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxdmq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://595f9ea75461f18e763d84799e368becc07b72acdc5337d20d6995bf53c3d27a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxdmq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6hbqd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:20Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.170236 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.170265 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.170275 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.170290 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.170301 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:20Z","lastTransitionTime":"2025-11-25T10:33:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.176038 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:20Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.188067 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17b8714be3bd121e06641873e003f08560b1f71e56d6ea936966df7d4896eb03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85731e6b6dbe1777bc7161a010f39f79523ec4b08d71d5d8a975573aa7fe4957\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:20Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.200091 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25595ccd-762d-414a-9916-4fe38d478b07\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d4b2d8db4d13392d9789e161fd4b8f6eef983b79128ee584d72e4c06565bc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf780f791771ea43a8c8176d02fdfb5204406e023207299369ae46ceb342baaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0251519fd4262f2ea5c64959db49433afdafe36b7ed2404d660eeb7c365d2746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad5c205e0eeecfd43cbdf0eb1a334e3e8faa5e9a9e8dd202c397d5cbb36e081e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:20Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.212785 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:20Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.227706 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ldzbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b84e32ae-4897-4644-a656-6bd82012cc81\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ef433744761f26e8e1f6c0db123821a11923ffd774830744285cd4178d89533\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5pmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ldzbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:20Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.248176 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"954085fa-d03a-4e88-9dfd-566257e3558d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f00746930c4588e548f0e2b60d52bf448aff7401b676a7855cb0f4cce00d329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a14a078ba7d2aed95992b36502008f315f0d3a634b64ed49da70398b2562d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d7254196b3702e48083ae98e829b73c539adf69ed5b2bbdc62189d09ae1762\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d727430108977574365f0e9c04b44e03307bd7ff89cdbb1e48a6e916f504d6b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94c7949f077b1ca2b310bedb090aa72fc4bd7e8f2c8bf438096edfca8568f36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed2e7cad7bb24b5b56a1295098a6a7c230c8f9b9c29a88ef653ae8b6ac31667d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e6202b51ac13209f3a6de880638a6d142f79755311646c0216f11dfe8bbe3f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e6202b51ac13209f3a6de880638a6d142f79755311646c0216f11dfe8bbe3f7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T10:32:56Z\\\",\\\"message\\\":\\\"to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:56Z is after 2025-08-24T17:21:41Z]\\\\nI1125 10:32:56.048189 6453 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-console/console_TCP_cluster\\\\\\\", UUID:\\\\\\\"d7d7b270-1480-47f8-bdf9-690dbab310cb\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console/console\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]strin\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-mzhnw_openshift-ovn-kubernetes(954085fa-d03a-4e88-9dfd-566257e3558d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dbf499f4ae6161752abaa7a466a027c6f874016be90a534e69f07e3dd3f3d9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzhnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:20Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.260630 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-d4msn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f12cef3-7e40-4f61-836e-23cd0e578e71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s7lkv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s7lkv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:44Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-d4msn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:20Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.272474 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.272508 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.272517 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.272533 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.272543 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:20Z","lastTransitionTime":"2025-11-25T10:33:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.279039 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d9e58df-1755-4422-8d04-211f98c7c011\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea762f627aa69fb88a9bd82545fa677e9ccfa97f99a5e92c0aa490e619c0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a06b811e5ec52ce984a2403ba2c98332ba60466d08b3580ad9192fac6caef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0385f6ed9597974267f7c45547f6b6725b4ee83a2cc0508ff2d1c1ce8753ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b915b565f7fbb3e4e856e7f1035a519e82bfb3023660c121f6e3813de353bb92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18465f3a6916c36457a305ba5d568c9077f34f1d000532ed518f237d100008b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:20Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.290640 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:20Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.300427 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f948b87-ac86-4de6-ad64-c2ef947f84d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4482f73a45bc82bd55770c3fd788becacec94893fed1c305fed1c81c753d43aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ba68bd093be94d0d40f3851638db5d26bdfc52d175b1ac78afb91dda7affcfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2krbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:20Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.313301 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qppnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"357491b8-cd9c-486e-9898-5ba33678513d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc88d700ca3a449ef1cb32d6124b16d0024d6981ced12e276083f86a1b322ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd9c047c5ce1d896f803ff6cafa9bbf3a18a391b44539e1f1312e08c89e51cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd9c047c5ce1d896f803ff6cafa9bbf3a18a391b44539e1f1312e08c89e51cb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://565a6db63fb69abbba81ffdf6085797900858d263cd6affcb0b5f195474c2516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://565a6db63fb69abbba81ffdf6085797900858d263cd6affcb0b5f195474c2516\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://544885302dffff8336a9aa3c4afa80d69088ffaa2402648f4a329e0a798a8c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://544885302dffff8336a9aa3c4afa80d69088ffaa2402648f4a329e0a798a8c99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8bc193fb3f85b980c31af2e826cee90adb633ad2ee82ff1d46e182ab52e8fab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8bc193fb3f85b980c31af2e826cee90adb633ad2ee82ff1d46e182ab52e8fab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7713e2a35bac8df1e8987fbcd452459fa3541caad0c52ed57fda187bd7c83d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7713e2a35bac8df1e8987fbcd452459fa3541caad0c52ed57fda187bd7c83d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34adfca89026fc20007a73a8cbd4e7c7b3137a581c3476867af17fabe976f71f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34adfca89026fc20007a73a8cbd4e7c7b3137a581c3476867af17fabe976f71f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qppnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:20Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.322494 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tgdlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2336955b-a703-4db4-8d0f-d08088ed59b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb0362942bfac635ee6292501b1b851772c01d75dd658eef4b17669a716de72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vjfp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tgdlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:20Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.332910 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62fcb632-f12a-46e2-bb7d-a5311f291ee7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28d43531927e75c3745f100940f7f6bf79fd21026c1309194df8cd7a39361b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69633332d03d2551ff4abb253d358de398eab18486b92db630ce06c0447fb9cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ccb78d9a51ac83ee208c2451685a2512de9deaf742a663bb8c892aa59b76162\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27826e155a9352b1e1866e83fdab683dbcd749cc04b6a4da2db25ecabe3d9a52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://991252e5f56f128827a5054b8e228284bc0266166582cbaedaadc53ca2f5b401\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:20Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.343086 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63b471fd-5e16-4761-8b07-44a135d5f5f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a68a57fa7fba8f351f264e427a50bed23f57775644e9c268aa7305efb6916794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67df7c609d5a5d658be4c3e873ec3052bd070ad9469f8ff44c29a9571e3e5d95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://269ced131bc0777dc6752070470c95b4bedfe4b4240940eabd769500ddef0404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e48c0e3b58359c6c0a493d58f7d2aec6011747791a3c0d7788102e1fec18865\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e48c0e3b58359c6c0a493d58f7d2aec6011747791a3c0d7788102e1fec18865\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:20Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.374670 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.374720 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.374732 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.374784 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.374797 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:20Z","lastTransitionTime":"2025-11-25T10:33:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.427208 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-ldzbz_b84e32ae-4897-4644-a656-6bd82012cc81/kube-multus/0.log" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.427261 4821 generic.go:334] "Generic (PLEG): container finished" podID="b84e32ae-4897-4644-a656-6bd82012cc81" containerID="8ef433744761f26e8e1f6c0db123821a11923ffd774830744285cd4178d89533" exitCode=1 Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.427291 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-ldzbz" event={"ID":"b84e32ae-4897-4644-a656-6bd82012cc81","Type":"ContainerDied","Data":"8ef433744761f26e8e1f6c0db123821a11923ffd774830744285cd4178d89533"} Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.427554 4821 scope.go:117] "RemoveContainer" containerID="8ef433744761f26e8e1f6c0db123821a11923ffd774830744285cd4178d89533" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.440860 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:20Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.452602 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17b8714be3bd121e06641873e003f08560b1f71e56d6ea936966df7d4896eb03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85731e6b6dbe1777bc7161a010f39f79523ec4b08d71d5d8a975573aa7fe4957\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:20Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.463079 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zlk96" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d55aa91-9cb9-4c95-bc27-6fec05684c10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4521b5baee45ffc9c7661c781c0b3de458dff325ad525dde283d46af71c0cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vkd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zlk96\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:20Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.473515 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6hbqd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f71b3a2-63cb-443c-9f41-39e57030746c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65473b0e63c2e9c7c09a2bb8f31396daf1d87c2c83d20ed3508ce2fe7358dda3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxdmq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://595f9ea75461f18e763d84799e368becc07b72acdc5337d20d6995bf53c3d27a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxdmq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6hbqd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:20Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.477252 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.477488 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.477576 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.477723 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.477845 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:20Z","lastTransitionTime":"2025-11-25T10:33:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.487711 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ldzbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b84e32ae-4897-4644-a656-6bd82012cc81\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ef433744761f26e8e1f6c0db123821a11923ffd774830744285cd4178d89533\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ef433744761f26e8e1f6c0db123821a11923ffd774830744285cd4178d89533\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T10:33:19Z\\\",\\\"message\\\":\\\"2025-11-25T10:32:34+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_60a08f2b-f038-4c96-9f8b-1bd34a1965eb\\\\n2025-11-25T10:32:34+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_60a08f2b-f038-4c96-9f8b-1bd34a1965eb to /host/opt/cni/bin/\\\\n2025-11-25T10:32:34Z [verbose] multus-daemon started\\\\n2025-11-25T10:32:34Z [verbose] Readiness Indicator file check\\\\n2025-11-25T10:33:19Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5pmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ldzbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:20Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.507602 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"954085fa-d03a-4e88-9dfd-566257e3558d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f00746930c4588e548f0e2b60d52bf448aff7401b676a7855cb0f4cce00d329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a14a078ba7d2aed95992b36502008f315f0d3a634b64ed49da70398b2562d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d7254196b3702e48083ae98e829b73c539adf69ed5b2bbdc62189d09ae1762\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d727430108977574365f0e9c04b44e03307bd7ff89cdbb1e48a6e916f504d6b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94c7949f077b1ca2b310bedb090aa72fc4bd7e8f2c8bf438096edfca8568f36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed2e7cad7bb24b5b56a1295098a6a7c230c8f9b9c29a88ef653ae8b6ac31667d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e6202b51ac13209f3a6de880638a6d142f79755311646c0216f11dfe8bbe3f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e6202b51ac13209f3a6de880638a6d142f79755311646c0216f11dfe8bbe3f7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T10:32:56Z\\\",\\\"message\\\":\\\"to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:56Z is after 2025-08-24T17:21:41Z]\\\\nI1125 10:32:56.048189 6453 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-console/console_TCP_cluster\\\\\\\", UUID:\\\\\\\"d7d7b270-1480-47f8-bdf9-690dbab310cb\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console/console\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]strin\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-mzhnw_openshift-ovn-kubernetes(954085fa-d03a-4e88-9dfd-566257e3558d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dbf499f4ae6161752abaa7a466a027c6f874016be90a534e69f07e3dd3f3d9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzhnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:20Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.519766 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-d4msn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f12cef3-7e40-4f61-836e-23cd0e578e71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s7lkv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s7lkv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:44Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-d4msn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:20Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.542074 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d9e58df-1755-4422-8d04-211f98c7c011\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea762f627aa69fb88a9bd82545fa677e9ccfa97f99a5e92c0aa490e619c0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a06b811e5ec52ce984a2403ba2c98332ba60466d08b3580ad9192fac6caef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0385f6ed9597974267f7c45547f6b6725b4ee83a2cc0508ff2d1c1ce8753ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b915b565f7fbb3e4e856e7f1035a519e82bfb3023660c121f6e3813de353bb92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18465f3a6916c36457a305ba5d568c9077f34f1d000532ed518f237d100008b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:20Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.555455 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25595ccd-762d-414a-9916-4fe38d478b07\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d4b2d8db4d13392d9789e161fd4b8f6eef983b79128ee584d72e4c06565bc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf780f791771ea43a8c8176d02fdfb5204406e023207299369ae46ceb342baaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0251519fd4262f2ea5c64959db49433afdafe36b7ed2404d660eeb7c365d2746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad5c205e0eeecfd43cbdf0eb1a334e3e8faa5e9a9e8dd202c397d5cbb36e081e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:20Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.566190 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:20Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.580381 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.580414 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.580423 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.580437 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.580448 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:20Z","lastTransitionTime":"2025-11-25T10:33:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.581293 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qppnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"357491b8-cd9c-486e-9898-5ba33678513d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc88d700ca3a449ef1cb32d6124b16d0024d6981ced12e276083f86a1b322ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd9c047c5ce1d896f803ff6cafa9bbf3a18a391b44539e1f1312e08c89e51cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd9c047c5ce1d896f803ff6cafa9bbf3a18a391b44539e1f1312e08c89e51cb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://565a6db63fb69abbba81ffdf6085797900858d263cd6affcb0b5f195474c2516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://565a6db63fb69abbba81ffdf6085797900858d263cd6affcb0b5f195474c2516\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://544885302dffff8336a9aa3c4afa80d69088ffaa2402648f4a329e0a798a8c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://544885302dffff8336a9aa3c4afa80d69088ffaa2402648f4a329e0a798a8c99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8bc193fb3f85b980c31af2e826cee90adb633ad2ee82ff1d46e182ab52e8fab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8bc193fb3f85b980c31af2e826cee90adb633ad2ee82ff1d46e182ab52e8fab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7713e2a35bac8df1e8987fbcd452459fa3541caad0c52ed57fda187bd7c83d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7713e2a35bac8df1e8987fbcd452459fa3541caad0c52ed57fda187bd7c83d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34adfca89026fc20007a73a8cbd4e7c7b3137a581c3476867af17fabe976f71f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34adfca89026fc20007a73a8cbd4e7c7b3137a581c3476867af17fabe976f71f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qppnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:20Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.592235 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tgdlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2336955b-a703-4db4-8d0f-d08088ed59b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb0362942bfac635ee6292501b1b851772c01d75dd658eef4b17669a716de72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vjfp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tgdlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:20Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.605703 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62fcb632-f12a-46e2-bb7d-a5311f291ee7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28d43531927e75c3745f100940f7f6bf79fd21026c1309194df8cd7a39361b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69633332d03d2551ff4abb253d358de398eab18486b92db630ce06c0447fb9cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ccb78d9a51ac83ee208c2451685a2512de9deaf742a663bb8c892aa59b76162\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27826e155a9352b1e1866e83fdab683dbcd749cc04b6a4da2db25ecabe3d9a52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://991252e5f56f128827a5054b8e228284bc0266166582cbaedaadc53ca2f5b401\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:20Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.615914 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63b471fd-5e16-4761-8b07-44a135d5f5f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a68a57fa7fba8f351f264e427a50bed23f57775644e9c268aa7305efb6916794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67df7c609d5a5d658be4c3e873ec3052bd070ad9469f8ff44c29a9571e3e5d95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://269ced131bc0777dc6752070470c95b4bedfe4b4240940eabd769500ddef0404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e48c0e3b58359c6c0a493d58f7d2aec6011747791a3c0d7788102e1fec18865\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e48c0e3b58359c6c0a493d58f7d2aec6011747791a3c0d7788102e1fec18865\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:20Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.629610 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:20Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.641543 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f948b87-ac86-4de6-ad64-c2ef947f84d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4482f73a45bc82bd55770c3fd788becacec94893fed1c305fed1c81c753d43aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ba68bd093be94d0d40f3851638db5d26bdfc52d175b1ac78afb91dda7affcfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2krbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:20Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.654992 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06f949936bac99cfbf89d720c9549c074351d9077bd963fa024d87004c6be23e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:20Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.666437 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384d8ba58264ab89796d3b53afb0c4c1be33cb4fba719a8beef4afc2be7d66a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:20Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.683096 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.683148 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.683178 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.683192 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.683200 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:20Z","lastTransitionTime":"2025-11-25T10:33:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.785973 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.786009 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.786019 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.786033 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.786044 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:20Z","lastTransitionTime":"2025-11-25T10:33:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.888612 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.888660 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.888674 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.888689 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.888698 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:20Z","lastTransitionTime":"2025-11-25T10:33:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.990581 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.990833 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.990893 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.990964 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:20 crc kubenswrapper[4821]: I1125 10:33:20.991031 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:20Z","lastTransitionTime":"2025-11-25T10:33:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:21 crc kubenswrapper[4821]: I1125 10:33:21.093754 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:21 crc kubenswrapper[4821]: I1125 10:33:21.093805 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:21 crc kubenswrapper[4821]: I1125 10:33:21.093813 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:21 crc kubenswrapper[4821]: I1125 10:33:21.093828 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:21 crc kubenswrapper[4821]: I1125 10:33:21.093837 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:21Z","lastTransitionTime":"2025-11-25T10:33:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:21 crc kubenswrapper[4821]: I1125 10:33:21.114108 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:33:21 crc kubenswrapper[4821]: I1125 10:33:21.114183 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:33:21 crc kubenswrapper[4821]: I1125 10:33:21.114340 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:33:21 crc kubenswrapper[4821]: E1125 10:33:21.114498 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:33:21 crc kubenswrapper[4821]: E1125 10:33:21.114625 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:33:21 crc kubenswrapper[4821]: E1125 10:33:21.114740 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:33:21 crc kubenswrapper[4821]: I1125 10:33:21.196356 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:21 crc kubenswrapper[4821]: I1125 10:33:21.196390 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:21 crc kubenswrapper[4821]: I1125 10:33:21.196403 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:21 crc kubenswrapper[4821]: I1125 10:33:21.196423 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:21 crc kubenswrapper[4821]: I1125 10:33:21.196435 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:21Z","lastTransitionTime":"2025-11-25T10:33:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:21 crc kubenswrapper[4821]: I1125 10:33:21.298753 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:21 crc kubenswrapper[4821]: I1125 10:33:21.298792 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:21 crc kubenswrapper[4821]: I1125 10:33:21.298800 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:21 crc kubenswrapper[4821]: I1125 10:33:21.298814 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:21 crc kubenswrapper[4821]: I1125 10:33:21.298824 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:21Z","lastTransitionTime":"2025-11-25T10:33:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:21 crc kubenswrapper[4821]: I1125 10:33:21.401059 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:21 crc kubenswrapper[4821]: I1125 10:33:21.401103 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:21 crc kubenswrapper[4821]: I1125 10:33:21.401112 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:21 crc kubenswrapper[4821]: I1125 10:33:21.401127 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:21 crc kubenswrapper[4821]: I1125 10:33:21.401136 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:21Z","lastTransitionTime":"2025-11-25T10:33:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:21 crc kubenswrapper[4821]: I1125 10:33:21.433209 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-ldzbz_b84e32ae-4897-4644-a656-6bd82012cc81/kube-multus/0.log" Nov 25 10:33:21 crc kubenswrapper[4821]: I1125 10:33:21.433269 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-ldzbz" event={"ID":"b84e32ae-4897-4644-a656-6bd82012cc81","Type":"ContainerStarted","Data":"418dc8577fdc98324677dbccfcf9317b67cab4408432b30df2818b22f88d426e"} Nov 25 10:33:21 crc kubenswrapper[4821]: I1125 10:33:21.451226 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qppnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"357491b8-cd9c-486e-9898-5ba33678513d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc88d700ca3a449ef1cb32d6124b16d0024d6981ced12e276083f86a1b322ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd9c047c5ce1d896f803ff6cafa9bbf3a18a391b44539e1f1312e08c89e51cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd9c047c5ce1d896f803ff6cafa9bbf3a18a391b44539e1f1312e08c89e51cb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://565a6db63fb69abbba81ffdf6085797900858d263cd6affcb0b5f195474c2516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://565a6db63fb69abbba81ffdf6085797900858d263cd6affcb0b5f195474c2516\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://544885302dffff8336a9aa3c4afa80d69088ffaa2402648f4a329e0a798a8c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://544885302dffff8336a9aa3c4afa80d69088ffaa2402648f4a329e0a798a8c99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8bc193fb3f85b980c31af2e826cee90adb633ad2ee82ff1d46e182ab52e8fab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8bc193fb3f85b980c31af2e826cee90adb633ad2ee82ff1d46e182ab52e8fab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7713e2a35bac8df1e8987fbcd452459fa3541caad0c52ed57fda187bd7c83d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7713e2a35bac8df1e8987fbcd452459fa3541caad0c52ed57fda187bd7c83d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34adfca89026fc20007a73a8cbd4e7c7b3137a581c3476867af17fabe976f71f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34adfca89026fc20007a73a8cbd4e7c7b3137a581c3476867af17fabe976f71f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qppnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:21Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:21 crc kubenswrapper[4821]: I1125 10:33:21.462838 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tgdlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2336955b-a703-4db4-8d0f-d08088ed59b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb0362942bfac635ee6292501b1b851772c01d75dd658eef4b17669a716de72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vjfp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tgdlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:21Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:21 crc kubenswrapper[4821]: I1125 10:33:21.486046 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62fcb632-f12a-46e2-bb7d-a5311f291ee7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28d43531927e75c3745f100940f7f6bf79fd21026c1309194df8cd7a39361b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69633332d03d2551ff4abb253d358de398eab18486b92db630ce06c0447fb9cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ccb78d9a51ac83ee208c2451685a2512de9deaf742a663bb8c892aa59b76162\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27826e155a9352b1e1866e83fdab683dbcd749cc04b6a4da2db25ecabe3d9a52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://991252e5f56f128827a5054b8e228284bc0266166582cbaedaadc53ca2f5b401\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:21Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:21 crc kubenswrapper[4821]: I1125 10:33:21.513201 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:21 crc kubenswrapper[4821]: I1125 10:33:21.513245 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:21 crc kubenswrapper[4821]: I1125 10:33:21.513255 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:21 crc kubenswrapper[4821]: I1125 10:33:21.513274 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:21 crc kubenswrapper[4821]: I1125 10:33:21.513286 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:21Z","lastTransitionTime":"2025-11-25T10:33:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:21 crc kubenswrapper[4821]: I1125 10:33:21.558858 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63b471fd-5e16-4761-8b07-44a135d5f5f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a68a57fa7fba8f351f264e427a50bed23f57775644e9c268aa7305efb6916794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67df7c609d5a5d658be4c3e873ec3052bd070ad9469f8ff44c29a9571e3e5d95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://269ced131bc0777dc6752070470c95b4bedfe4b4240940eabd769500ddef0404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e48c0e3b58359c6c0a493d58f7d2aec6011747791a3c0d7788102e1fec18865\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e48c0e3b58359c6c0a493d58f7d2aec6011747791a3c0d7788102e1fec18865\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:21Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:21 crc kubenswrapper[4821]: I1125 10:33:21.570136 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:21Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:21 crc kubenswrapper[4821]: I1125 10:33:21.581753 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f948b87-ac86-4de6-ad64-c2ef947f84d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4482f73a45bc82bd55770c3fd788becacec94893fed1c305fed1c81c753d43aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ba68bd093be94d0d40f3851638db5d26bdfc52d175b1ac78afb91dda7affcfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2krbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:21Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:21 crc kubenswrapper[4821]: I1125 10:33:21.592782 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06f949936bac99cfbf89d720c9549c074351d9077bd963fa024d87004c6be23e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:21Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:21 crc kubenswrapper[4821]: I1125 10:33:21.604611 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384d8ba58264ab89796d3b53afb0c4c1be33cb4fba719a8beef4afc2be7d66a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:21Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:21 crc kubenswrapper[4821]: I1125 10:33:21.614846 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:21Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:21 crc kubenswrapper[4821]: I1125 10:33:21.615629 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:21 crc kubenswrapper[4821]: I1125 10:33:21.615657 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:21 crc kubenswrapper[4821]: I1125 10:33:21.615667 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:21 crc kubenswrapper[4821]: I1125 10:33:21.615682 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:21 crc kubenswrapper[4821]: I1125 10:33:21.615691 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:21Z","lastTransitionTime":"2025-11-25T10:33:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:21 crc kubenswrapper[4821]: I1125 10:33:21.625370 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17b8714be3bd121e06641873e003f08560b1f71e56d6ea936966df7d4896eb03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85731e6b6dbe1777bc7161a010f39f79523ec4b08d71d5d8a975573aa7fe4957\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:21Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:21 crc kubenswrapper[4821]: I1125 10:33:21.635963 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zlk96" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d55aa91-9cb9-4c95-bc27-6fec05684c10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4521b5baee45ffc9c7661c781c0b3de458dff325ad525dde283d46af71c0cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vkd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zlk96\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:21Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:21 crc kubenswrapper[4821]: I1125 10:33:21.647510 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6hbqd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f71b3a2-63cb-443c-9f41-39e57030746c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65473b0e63c2e9c7c09a2bb8f31396daf1d87c2c83d20ed3508ce2fe7358dda3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxdmq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://595f9ea75461f18e763d84799e368becc07b72acdc5337d20d6995bf53c3d27a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxdmq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6hbqd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:21Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:21 crc kubenswrapper[4821]: I1125 10:33:21.660576 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ldzbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b84e32ae-4897-4644-a656-6bd82012cc81\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://418dc8577fdc98324677dbccfcf9317b67cab4408432b30df2818b22f88d426e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ef433744761f26e8e1f6c0db123821a11923ffd774830744285cd4178d89533\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T10:33:19Z\\\",\\\"message\\\":\\\"2025-11-25T10:32:34+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_60a08f2b-f038-4c96-9f8b-1bd34a1965eb\\\\n2025-11-25T10:32:34+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_60a08f2b-f038-4c96-9f8b-1bd34a1965eb to /host/opt/cni/bin/\\\\n2025-11-25T10:32:34Z [verbose] multus-daemon started\\\\n2025-11-25T10:32:34Z [verbose] Readiness Indicator file check\\\\n2025-11-25T10:33:19Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:33:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5pmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ldzbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:21Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:21 crc kubenswrapper[4821]: I1125 10:33:21.678619 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"954085fa-d03a-4e88-9dfd-566257e3558d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f00746930c4588e548f0e2b60d52bf448aff7401b676a7855cb0f4cce00d329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a14a078ba7d2aed95992b36502008f315f0d3a634b64ed49da70398b2562d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d7254196b3702e48083ae98e829b73c539adf69ed5b2bbdc62189d09ae1762\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d727430108977574365f0e9c04b44e03307bd7ff89cdbb1e48a6e916f504d6b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94c7949f077b1ca2b310bedb090aa72fc4bd7e8f2c8bf438096edfca8568f36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed2e7cad7bb24b5b56a1295098a6a7c230c8f9b9c29a88ef653ae8b6ac31667d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e6202b51ac13209f3a6de880638a6d142f79755311646c0216f11dfe8bbe3f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e6202b51ac13209f3a6de880638a6d142f79755311646c0216f11dfe8bbe3f7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T10:32:56Z\\\",\\\"message\\\":\\\"to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:56Z is after 2025-08-24T17:21:41Z]\\\\nI1125 10:32:56.048189 6453 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-console/console_TCP_cluster\\\\\\\", UUID:\\\\\\\"d7d7b270-1480-47f8-bdf9-690dbab310cb\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console/console\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]strin\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-mzhnw_openshift-ovn-kubernetes(954085fa-d03a-4e88-9dfd-566257e3558d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dbf499f4ae6161752abaa7a466a027c6f874016be90a534e69f07e3dd3f3d9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzhnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:21Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:21 crc kubenswrapper[4821]: I1125 10:33:21.688031 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-d4msn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f12cef3-7e40-4f61-836e-23cd0e578e71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s7lkv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s7lkv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:44Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-d4msn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:21Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:21 crc kubenswrapper[4821]: I1125 10:33:21.707424 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d9e58df-1755-4422-8d04-211f98c7c011\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea762f627aa69fb88a9bd82545fa677e9ccfa97f99a5e92c0aa490e619c0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a06b811e5ec52ce984a2403ba2c98332ba60466d08b3580ad9192fac6caef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0385f6ed9597974267f7c45547f6b6725b4ee83a2cc0508ff2d1c1ce8753ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b915b565f7fbb3e4e856e7f1035a519e82bfb3023660c121f6e3813de353bb92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18465f3a6916c36457a305ba5d568c9077f34f1d000532ed518f237d100008b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:21Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:21 crc kubenswrapper[4821]: I1125 10:33:21.719836 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:21 crc kubenswrapper[4821]: I1125 10:33:21.719877 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:21 crc kubenswrapper[4821]: I1125 10:33:21.719887 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:21 crc kubenswrapper[4821]: I1125 10:33:21.719903 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:21 crc kubenswrapper[4821]: I1125 10:33:21.719914 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:21Z","lastTransitionTime":"2025-11-25T10:33:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:21 crc kubenswrapper[4821]: I1125 10:33:21.719902 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25595ccd-762d-414a-9916-4fe38d478b07\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d4b2d8db4d13392d9789e161fd4b8f6eef983b79128ee584d72e4c06565bc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf780f791771ea43a8c8176d02fdfb5204406e023207299369ae46ceb342baaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0251519fd4262f2ea5c64959db49433afdafe36b7ed2404d660eeb7c365d2746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad5c205e0eeecfd43cbdf0eb1a334e3e8faa5e9a9e8dd202c397d5cbb36e081e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:21Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:21 crc kubenswrapper[4821]: I1125 10:33:21.733273 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:21Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:21 crc kubenswrapper[4821]: I1125 10:33:21.822470 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:21 crc kubenswrapper[4821]: I1125 10:33:21.822518 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:21 crc kubenswrapper[4821]: I1125 10:33:21.822530 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:21 crc kubenswrapper[4821]: I1125 10:33:21.822559 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:21 crc kubenswrapper[4821]: I1125 10:33:21.822572 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:21Z","lastTransitionTime":"2025-11-25T10:33:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:21 crc kubenswrapper[4821]: I1125 10:33:21.924906 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:21 crc kubenswrapper[4821]: I1125 10:33:21.925211 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:21 crc kubenswrapper[4821]: I1125 10:33:21.925508 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:21 crc kubenswrapper[4821]: I1125 10:33:21.925603 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:21 crc kubenswrapper[4821]: I1125 10:33:21.925664 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:21Z","lastTransitionTime":"2025-11-25T10:33:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:22 crc kubenswrapper[4821]: I1125 10:33:22.028043 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:22 crc kubenswrapper[4821]: I1125 10:33:22.028084 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:22 crc kubenswrapper[4821]: I1125 10:33:22.028094 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:22 crc kubenswrapper[4821]: I1125 10:33:22.028109 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:22 crc kubenswrapper[4821]: I1125 10:33:22.028120 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:22Z","lastTransitionTime":"2025-11-25T10:33:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:22 crc kubenswrapper[4821]: I1125 10:33:22.114227 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4msn" Nov 25 10:33:22 crc kubenswrapper[4821]: E1125 10:33:22.114373 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4msn" podUID="1f12cef3-7e40-4f61-836e-23cd0e578e71" Nov 25 10:33:22 crc kubenswrapper[4821]: I1125 10:33:22.130309 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:22 crc kubenswrapper[4821]: I1125 10:33:22.130355 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:22 crc kubenswrapper[4821]: I1125 10:33:22.130368 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:22 crc kubenswrapper[4821]: I1125 10:33:22.130387 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:22 crc kubenswrapper[4821]: I1125 10:33:22.130401 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:22Z","lastTransitionTime":"2025-11-25T10:33:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:22 crc kubenswrapper[4821]: I1125 10:33:22.232300 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:22 crc kubenswrapper[4821]: I1125 10:33:22.232335 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:22 crc kubenswrapper[4821]: I1125 10:33:22.232344 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:22 crc kubenswrapper[4821]: I1125 10:33:22.232359 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:22 crc kubenswrapper[4821]: I1125 10:33:22.232368 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:22Z","lastTransitionTime":"2025-11-25T10:33:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:22 crc kubenswrapper[4821]: I1125 10:33:22.334289 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:22 crc kubenswrapper[4821]: I1125 10:33:22.334337 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:22 crc kubenswrapper[4821]: I1125 10:33:22.334350 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:22 crc kubenswrapper[4821]: I1125 10:33:22.334365 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:22 crc kubenswrapper[4821]: I1125 10:33:22.334377 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:22Z","lastTransitionTime":"2025-11-25T10:33:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:22 crc kubenswrapper[4821]: I1125 10:33:22.435593 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:22 crc kubenswrapper[4821]: I1125 10:33:22.435641 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:22 crc kubenswrapper[4821]: I1125 10:33:22.435650 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:22 crc kubenswrapper[4821]: I1125 10:33:22.435665 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:22 crc kubenswrapper[4821]: I1125 10:33:22.435674 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:22Z","lastTransitionTime":"2025-11-25T10:33:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:22 crc kubenswrapper[4821]: I1125 10:33:22.537680 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:22 crc kubenswrapper[4821]: I1125 10:33:22.537745 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:22 crc kubenswrapper[4821]: I1125 10:33:22.537754 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:22 crc kubenswrapper[4821]: I1125 10:33:22.537774 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:22 crc kubenswrapper[4821]: I1125 10:33:22.537784 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:22Z","lastTransitionTime":"2025-11-25T10:33:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:22 crc kubenswrapper[4821]: I1125 10:33:22.639814 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:22 crc kubenswrapper[4821]: I1125 10:33:22.639850 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:22 crc kubenswrapper[4821]: I1125 10:33:22.639862 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:22 crc kubenswrapper[4821]: I1125 10:33:22.639875 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:22 crc kubenswrapper[4821]: I1125 10:33:22.639890 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:22Z","lastTransitionTime":"2025-11-25T10:33:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:22 crc kubenswrapper[4821]: I1125 10:33:22.742028 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:22 crc kubenswrapper[4821]: I1125 10:33:22.742072 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:22 crc kubenswrapper[4821]: I1125 10:33:22.742081 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:22 crc kubenswrapper[4821]: I1125 10:33:22.742096 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:22 crc kubenswrapper[4821]: I1125 10:33:22.742105 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:22Z","lastTransitionTime":"2025-11-25T10:33:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:22 crc kubenswrapper[4821]: I1125 10:33:22.846258 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:22 crc kubenswrapper[4821]: I1125 10:33:22.846302 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:22 crc kubenswrapper[4821]: I1125 10:33:22.846312 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:22 crc kubenswrapper[4821]: I1125 10:33:22.846327 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:22 crc kubenswrapper[4821]: I1125 10:33:22.846336 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:22Z","lastTransitionTime":"2025-11-25T10:33:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:22 crc kubenswrapper[4821]: I1125 10:33:22.948355 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:22 crc kubenswrapper[4821]: I1125 10:33:22.948409 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:22 crc kubenswrapper[4821]: I1125 10:33:22.948422 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:22 crc kubenswrapper[4821]: I1125 10:33:22.948445 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:22 crc kubenswrapper[4821]: I1125 10:33:22.948457 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:22Z","lastTransitionTime":"2025-11-25T10:33:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:23 crc kubenswrapper[4821]: I1125 10:33:23.050528 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:23 crc kubenswrapper[4821]: I1125 10:33:23.050565 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:23 crc kubenswrapper[4821]: I1125 10:33:23.050575 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:23 crc kubenswrapper[4821]: I1125 10:33:23.050589 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:23 crc kubenswrapper[4821]: I1125 10:33:23.050599 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:23Z","lastTransitionTime":"2025-11-25T10:33:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:23 crc kubenswrapper[4821]: I1125 10:33:23.113497 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:33:23 crc kubenswrapper[4821]: I1125 10:33:23.113548 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:33:23 crc kubenswrapper[4821]: I1125 10:33:23.113578 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:33:23 crc kubenswrapper[4821]: E1125 10:33:23.113634 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:33:23 crc kubenswrapper[4821]: E1125 10:33:23.113796 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:33:23 crc kubenswrapper[4821]: E1125 10:33:23.113830 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:33:23 crc kubenswrapper[4821]: I1125 10:33:23.124573 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Nov 25 10:33:23 crc kubenswrapper[4821]: I1125 10:33:23.152881 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:23 crc kubenswrapper[4821]: I1125 10:33:23.152918 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:23 crc kubenswrapper[4821]: I1125 10:33:23.152926 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:23 crc kubenswrapper[4821]: I1125 10:33:23.152941 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:23 crc kubenswrapper[4821]: I1125 10:33:23.152951 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:23Z","lastTransitionTime":"2025-11-25T10:33:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:23 crc kubenswrapper[4821]: I1125 10:33:23.255783 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:23 crc kubenswrapper[4821]: I1125 10:33:23.255828 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:23 crc kubenswrapper[4821]: I1125 10:33:23.255836 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:23 crc kubenswrapper[4821]: I1125 10:33:23.255852 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:23 crc kubenswrapper[4821]: I1125 10:33:23.255861 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:23Z","lastTransitionTime":"2025-11-25T10:33:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:23 crc kubenswrapper[4821]: I1125 10:33:23.357759 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:23 crc kubenswrapper[4821]: I1125 10:33:23.357792 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:23 crc kubenswrapper[4821]: I1125 10:33:23.357801 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:23 crc kubenswrapper[4821]: I1125 10:33:23.357815 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:23 crc kubenswrapper[4821]: I1125 10:33:23.357824 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:23Z","lastTransitionTime":"2025-11-25T10:33:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:23 crc kubenswrapper[4821]: I1125 10:33:23.460750 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:23 crc kubenswrapper[4821]: I1125 10:33:23.460800 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:23 crc kubenswrapper[4821]: I1125 10:33:23.460812 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:23 crc kubenswrapper[4821]: I1125 10:33:23.460831 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:23 crc kubenswrapper[4821]: I1125 10:33:23.460843 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:23Z","lastTransitionTime":"2025-11-25T10:33:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:23 crc kubenswrapper[4821]: I1125 10:33:23.562562 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:23 crc kubenswrapper[4821]: I1125 10:33:23.562602 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:23 crc kubenswrapper[4821]: I1125 10:33:23.562613 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:23 crc kubenswrapper[4821]: I1125 10:33:23.562632 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:23 crc kubenswrapper[4821]: I1125 10:33:23.562644 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:23Z","lastTransitionTime":"2025-11-25T10:33:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:23 crc kubenswrapper[4821]: I1125 10:33:23.664407 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:23 crc kubenswrapper[4821]: I1125 10:33:23.664443 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:23 crc kubenswrapper[4821]: I1125 10:33:23.664454 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:23 crc kubenswrapper[4821]: I1125 10:33:23.664471 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:23 crc kubenswrapper[4821]: I1125 10:33:23.664482 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:23Z","lastTransitionTime":"2025-11-25T10:33:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:23 crc kubenswrapper[4821]: I1125 10:33:23.766808 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:23 crc kubenswrapper[4821]: I1125 10:33:23.766852 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:23 crc kubenswrapper[4821]: I1125 10:33:23.766866 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:23 crc kubenswrapper[4821]: I1125 10:33:23.766883 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:23 crc kubenswrapper[4821]: I1125 10:33:23.766893 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:23Z","lastTransitionTime":"2025-11-25T10:33:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:23 crc kubenswrapper[4821]: I1125 10:33:23.868417 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:23 crc kubenswrapper[4821]: I1125 10:33:23.868460 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:23 crc kubenswrapper[4821]: I1125 10:33:23.868468 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:23 crc kubenswrapper[4821]: I1125 10:33:23.868487 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:23 crc kubenswrapper[4821]: I1125 10:33:23.868495 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:23Z","lastTransitionTime":"2025-11-25T10:33:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:23 crc kubenswrapper[4821]: I1125 10:33:23.970788 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:23 crc kubenswrapper[4821]: I1125 10:33:23.970832 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:23 crc kubenswrapper[4821]: I1125 10:33:23.970843 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:23 crc kubenswrapper[4821]: I1125 10:33:23.970860 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:23 crc kubenswrapper[4821]: I1125 10:33:23.970872 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:23Z","lastTransitionTime":"2025-11-25T10:33:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:24 crc kubenswrapper[4821]: I1125 10:33:24.073044 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:24 crc kubenswrapper[4821]: I1125 10:33:24.073093 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:24 crc kubenswrapper[4821]: I1125 10:33:24.073101 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:24 crc kubenswrapper[4821]: I1125 10:33:24.073118 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:24 crc kubenswrapper[4821]: I1125 10:33:24.073130 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:24Z","lastTransitionTime":"2025-11-25T10:33:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:24 crc kubenswrapper[4821]: I1125 10:33:24.113705 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4msn" Nov 25 10:33:24 crc kubenswrapper[4821]: E1125 10:33:24.113896 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4msn" podUID="1f12cef3-7e40-4f61-836e-23cd0e578e71" Nov 25 10:33:24 crc kubenswrapper[4821]: I1125 10:33:24.175900 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:24 crc kubenswrapper[4821]: I1125 10:33:24.175951 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:24 crc kubenswrapper[4821]: I1125 10:33:24.175964 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:24 crc kubenswrapper[4821]: I1125 10:33:24.176003 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:24 crc kubenswrapper[4821]: I1125 10:33:24.176015 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:24Z","lastTransitionTime":"2025-11-25T10:33:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:24 crc kubenswrapper[4821]: I1125 10:33:24.277602 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:24 crc kubenswrapper[4821]: I1125 10:33:24.277650 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:24 crc kubenswrapper[4821]: I1125 10:33:24.277662 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:24 crc kubenswrapper[4821]: I1125 10:33:24.277680 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:24 crc kubenswrapper[4821]: I1125 10:33:24.277692 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:24Z","lastTransitionTime":"2025-11-25T10:33:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:24 crc kubenswrapper[4821]: I1125 10:33:24.379445 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:24 crc kubenswrapper[4821]: I1125 10:33:24.379498 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:24 crc kubenswrapper[4821]: I1125 10:33:24.379513 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:24 crc kubenswrapper[4821]: I1125 10:33:24.379531 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:24 crc kubenswrapper[4821]: I1125 10:33:24.379543 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:24Z","lastTransitionTime":"2025-11-25T10:33:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:24 crc kubenswrapper[4821]: I1125 10:33:24.482341 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:24 crc kubenswrapper[4821]: I1125 10:33:24.482370 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:24 crc kubenswrapper[4821]: I1125 10:33:24.482378 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:24 crc kubenswrapper[4821]: I1125 10:33:24.482392 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:24 crc kubenswrapper[4821]: I1125 10:33:24.482400 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:24Z","lastTransitionTime":"2025-11-25T10:33:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:24 crc kubenswrapper[4821]: I1125 10:33:24.584926 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:24 crc kubenswrapper[4821]: I1125 10:33:24.584962 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:24 crc kubenswrapper[4821]: I1125 10:33:24.584971 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:24 crc kubenswrapper[4821]: I1125 10:33:24.584986 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:24 crc kubenswrapper[4821]: I1125 10:33:24.584994 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:24Z","lastTransitionTime":"2025-11-25T10:33:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:24 crc kubenswrapper[4821]: I1125 10:33:24.686480 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:24 crc kubenswrapper[4821]: I1125 10:33:24.686534 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:24 crc kubenswrapper[4821]: I1125 10:33:24.686545 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:24 crc kubenswrapper[4821]: I1125 10:33:24.686560 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:24 crc kubenswrapper[4821]: I1125 10:33:24.686570 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:24Z","lastTransitionTime":"2025-11-25T10:33:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:24 crc kubenswrapper[4821]: I1125 10:33:24.788250 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:24 crc kubenswrapper[4821]: I1125 10:33:24.788338 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:24 crc kubenswrapper[4821]: I1125 10:33:24.788347 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:24 crc kubenswrapper[4821]: I1125 10:33:24.788364 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:24 crc kubenswrapper[4821]: I1125 10:33:24.788376 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:24Z","lastTransitionTime":"2025-11-25T10:33:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:24 crc kubenswrapper[4821]: I1125 10:33:24.890899 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:24 crc kubenswrapper[4821]: I1125 10:33:24.891644 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:24 crc kubenswrapper[4821]: I1125 10:33:24.891678 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:24 crc kubenswrapper[4821]: I1125 10:33:24.891697 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:24 crc kubenswrapper[4821]: I1125 10:33:24.891712 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:24Z","lastTransitionTime":"2025-11-25T10:33:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:24 crc kubenswrapper[4821]: I1125 10:33:24.993478 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:24 crc kubenswrapper[4821]: I1125 10:33:24.993516 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:24 crc kubenswrapper[4821]: I1125 10:33:24.993524 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:24 crc kubenswrapper[4821]: I1125 10:33:24.993539 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:24 crc kubenswrapper[4821]: I1125 10:33:24.993550 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:24Z","lastTransitionTime":"2025-11-25T10:33:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:25 crc kubenswrapper[4821]: I1125 10:33:25.095626 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:25 crc kubenswrapper[4821]: I1125 10:33:25.095673 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:25 crc kubenswrapper[4821]: I1125 10:33:25.095685 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:25 crc kubenswrapper[4821]: I1125 10:33:25.095700 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:25 crc kubenswrapper[4821]: I1125 10:33:25.095712 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:25Z","lastTransitionTime":"2025-11-25T10:33:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:25 crc kubenswrapper[4821]: I1125 10:33:25.113921 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:33:25 crc kubenswrapper[4821]: E1125 10:33:25.114037 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:33:25 crc kubenswrapper[4821]: I1125 10:33:25.114086 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:33:25 crc kubenswrapper[4821]: I1125 10:33:25.114095 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:33:25 crc kubenswrapper[4821]: E1125 10:33:25.114187 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:33:25 crc kubenswrapper[4821]: E1125 10:33:25.114286 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:33:25 crc kubenswrapper[4821]: I1125 10:33:25.198208 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:25 crc kubenswrapper[4821]: I1125 10:33:25.198262 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:25 crc kubenswrapper[4821]: I1125 10:33:25.198273 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:25 crc kubenswrapper[4821]: I1125 10:33:25.198290 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:25 crc kubenswrapper[4821]: I1125 10:33:25.198301 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:25Z","lastTransitionTime":"2025-11-25T10:33:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:25 crc kubenswrapper[4821]: I1125 10:33:25.300653 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:25 crc kubenswrapper[4821]: I1125 10:33:25.300685 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:25 crc kubenswrapper[4821]: I1125 10:33:25.300726 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:25 crc kubenswrapper[4821]: I1125 10:33:25.300740 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:25 crc kubenswrapper[4821]: I1125 10:33:25.300749 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:25Z","lastTransitionTime":"2025-11-25T10:33:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:25 crc kubenswrapper[4821]: I1125 10:33:25.402853 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:25 crc kubenswrapper[4821]: I1125 10:33:25.402905 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:25 crc kubenswrapper[4821]: I1125 10:33:25.402917 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:25 crc kubenswrapper[4821]: I1125 10:33:25.402935 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:25 crc kubenswrapper[4821]: I1125 10:33:25.402948 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:25Z","lastTransitionTime":"2025-11-25T10:33:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:25 crc kubenswrapper[4821]: I1125 10:33:25.506329 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:25 crc kubenswrapper[4821]: I1125 10:33:25.506412 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:25 crc kubenswrapper[4821]: I1125 10:33:25.506448 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:25 crc kubenswrapper[4821]: I1125 10:33:25.506491 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:25 crc kubenswrapper[4821]: I1125 10:33:25.506516 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:25Z","lastTransitionTime":"2025-11-25T10:33:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:25 crc kubenswrapper[4821]: I1125 10:33:25.609739 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:25 crc kubenswrapper[4821]: I1125 10:33:25.609825 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:25 crc kubenswrapper[4821]: I1125 10:33:25.609850 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:25 crc kubenswrapper[4821]: I1125 10:33:25.609883 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:25 crc kubenswrapper[4821]: I1125 10:33:25.609908 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:25Z","lastTransitionTime":"2025-11-25T10:33:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:25 crc kubenswrapper[4821]: I1125 10:33:25.712244 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:25 crc kubenswrapper[4821]: I1125 10:33:25.712280 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:25 crc kubenswrapper[4821]: I1125 10:33:25.712288 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:25 crc kubenswrapper[4821]: I1125 10:33:25.712302 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:25 crc kubenswrapper[4821]: I1125 10:33:25.712311 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:25Z","lastTransitionTime":"2025-11-25T10:33:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:25 crc kubenswrapper[4821]: I1125 10:33:25.814006 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:25 crc kubenswrapper[4821]: I1125 10:33:25.814077 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:25 crc kubenswrapper[4821]: I1125 10:33:25.814087 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:25 crc kubenswrapper[4821]: I1125 10:33:25.814100 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:25 crc kubenswrapper[4821]: I1125 10:33:25.814111 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:25Z","lastTransitionTime":"2025-11-25T10:33:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:25 crc kubenswrapper[4821]: I1125 10:33:25.916044 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:25 crc kubenswrapper[4821]: I1125 10:33:25.916079 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:25 crc kubenswrapper[4821]: I1125 10:33:25.916089 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:25 crc kubenswrapper[4821]: I1125 10:33:25.916131 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:25 crc kubenswrapper[4821]: I1125 10:33:25.916141 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:25Z","lastTransitionTime":"2025-11-25T10:33:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.019044 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.019079 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.019095 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.019116 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.019125 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:26Z","lastTransitionTime":"2025-11-25T10:33:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.113859 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4msn" Nov 25 10:33:26 crc kubenswrapper[4821]: E1125 10:33:26.113996 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4msn" podUID="1f12cef3-7e40-4f61-836e-23cd0e578e71" Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.115753 4821 scope.go:117] "RemoveContainer" containerID="7e6202b51ac13209f3a6de880638a6d142f79755311646c0216f11dfe8bbe3f7" Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.120838 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.120894 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.120913 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.120939 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.120957 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:26Z","lastTransitionTime":"2025-11-25T10:33:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.223785 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.223886 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.223898 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.223914 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.223926 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:26Z","lastTransitionTime":"2025-11-25T10:33:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.326357 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.326398 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.326407 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.326421 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.326430 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:26Z","lastTransitionTime":"2025-11-25T10:33:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.429292 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.429352 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.429368 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.429390 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.429404 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:26Z","lastTransitionTime":"2025-11-25T10:33:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.448652 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mzhnw_954085fa-d03a-4e88-9dfd-566257e3558d/ovnkube-controller/2.log" Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.450730 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" event={"ID":"954085fa-d03a-4e88-9dfd-566257e3558d","Type":"ContainerStarted","Data":"90d8792a79993c04a0eb774498395d955deff5c42c693892426685d157cf9cff"} Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.451240 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.464641 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06f949936bac99cfbf89d720c9549c074351d9077bd963fa024d87004c6be23e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:26Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.482421 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384d8ba58264ab89796d3b53afb0c4c1be33cb4fba719a8beef4afc2be7d66a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:26Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.499575 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:26Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.513337 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17b8714be3bd121e06641873e003f08560b1f71e56d6ea936966df7d4896eb03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85731e6b6dbe1777bc7161a010f39f79523ec4b08d71d5d8a975573aa7fe4957\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:26Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.526217 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zlk96" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d55aa91-9cb9-4c95-bc27-6fec05684c10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4521b5baee45ffc9c7661c781c0b3de458dff325ad525dde283d46af71c0cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vkd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zlk96\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:26Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.531846 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.531880 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.531891 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.531910 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.531921 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:26Z","lastTransitionTime":"2025-11-25T10:33:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.537959 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6hbqd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f71b3a2-63cb-443c-9f41-39e57030746c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65473b0e63c2e9c7c09a2bb8f31396daf1d87c2c83d20ed3508ce2fe7358dda3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxdmq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://595f9ea75461f18e763d84799e368becc07b72acdc5337d20d6995bf53c3d27a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxdmq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6hbqd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:26Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.550587 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ldzbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b84e32ae-4897-4644-a656-6bd82012cc81\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://418dc8577fdc98324677dbccfcf9317b67cab4408432b30df2818b22f88d426e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ef433744761f26e8e1f6c0db123821a11923ffd774830744285cd4178d89533\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T10:33:19Z\\\",\\\"message\\\":\\\"2025-11-25T10:32:34+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_60a08f2b-f038-4c96-9f8b-1bd34a1965eb\\\\n2025-11-25T10:32:34+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_60a08f2b-f038-4c96-9f8b-1bd34a1965eb to /host/opt/cni/bin/\\\\n2025-11-25T10:32:34Z [verbose] multus-daemon started\\\\n2025-11-25T10:32:34Z [verbose] Readiness Indicator file check\\\\n2025-11-25T10:33:19Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:33:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5pmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ldzbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:26Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.566961 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"954085fa-d03a-4e88-9dfd-566257e3558d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f00746930c4588e548f0e2b60d52bf448aff7401b676a7855cb0f4cce00d329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a14a078ba7d2aed95992b36502008f315f0d3a634b64ed49da70398b2562d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d7254196b3702e48083ae98e829b73c539adf69ed5b2bbdc62189d09ae1762\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d727430108977574365f0e9c04b44e03307bd7ff89cdbb1e48a6e916f504d6b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94c7949f077b1ca2b310bedb090aa72fc4bd7e8f2c8bf438096edfca8568f36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed2e7cad7bb24b5b56a1295098a6a7c230c8f9b9c29a88ef653ae8b6ac31667d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90d8792a79993c04a0eb774498395d955deff5c42c693892426685d157cf9cff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e6202b51ac13209f3a6de880638a6d142f79755311646c0216f11dfe8bbe3f7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T10:32:56Z\\\",\\\"message\\\":\\\"to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:56Z is after 2025-08-24T17:21:41Z]\\\\nI1125 10:32:56.048189 6453 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-console/console_TCP_cluster\\\\\\\", UUID:\\\\\\\"d7d7b270-1480-47f8-bdf9-690dbab310cb\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console/console\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]strin\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:33:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dbf499f4ae6161752abaa7a466a027c6f874016be90a534e69f07e3dd3f3d9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzhnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:26Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.577876 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-d4msn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f12cef3-7e40-4f61-836e-23cd0e578e71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s7lkv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s7lkv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:44Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-d4msn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:26Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.586905 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3e50766-ed07-4d22-95d0-3e0a9d822a17\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9525225d46e6f9225786ef8dd0ebe108d41efa26f6c94451c069aa5eb30da58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://733440f5c7493b5e1f4597ebb8f9d3d79507df0a1bc33393fb4534b5a2f2f077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://733440f5c7493b5e1f4597ebb8f9d3d79507df0a1bc33393fb4534b5a2f2f077\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:26Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.605005 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d9e58df-1755-4422-8d04-211f98c7c011\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea762f627aa69fb88a9bd82545fa677e9ccfa97f99a5e92c0aa490e619c0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a06b811e5ec52ce984a2403ba2c98332ba60466d08b3580ad9192fac6caef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0385f6ed9597974267f7c45547f6b6725b4ee83a2cc0508ff2d1c1ce8753ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b915b565f7fbb3e4e856e7f1035a519e82bfb3023660c121f6e3813de353bb92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18465f3a6916c36457a305ba5d568c9077f34f1d000532ed518f237d100008b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:26Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.615794 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25595ccd-762d-414a-9916-4fe38d478b07\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d4b2d8db4d13392d9789e161fd4b8f6eef983b79128ee584d72e4c06565bc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf780f791771ea43a8c8176d02fdfb5204406e023207299369ae46ceb342baaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0251519fd4262f2ea5c64959db49433afdafe36b7ed2404d660eeb7c365d2746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad5c205e0eeecfd43cbdf0eb1a334e3e8faa5e9a9e8dd202c397d5cbb36e081e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:26Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.627457 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:26Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.634394 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.634433 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.634442 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.634456 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.634465 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:26Z","lastTransitionTime":"2025-11-25T10:33:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.640907 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qppnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"357491b8-cd9c-486e-9898-5ba33678513d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc88d700ca3a449ef1cb32d6124b16d0024d6981ced12e276083f86a1b322ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd9c047c5ce1d896f803ff6cafa9bbf3a18a391b44539e1f1312e08c89e51cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd9c047c5ce1d896f803ff6cafa9bbf3a18a391b44539e1f1312e08c89e51cb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://565a6db63fb69abbba81ffdf6085797900858d263cd6affcb0b5f195474c2516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://565a6db63fb69abbba81ffdf6085797900858d263cd6affcb0b5f195474c2516\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://544885302dffff8336a9aa3c4afa80d69088ffaa2402648f4a329e0a798a8c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://544885302dffff8336a9aa3c4afa80d69088ffaa2402648f4a329e0a798a8c99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8bc193fb3f85b980c31af2e826cee90adb633ad2ee82ff1d46e182ab52e8fab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8bc193fb3f85b980c31af2e826cee90adb633ad2ee82ff1d46e182ab52e8fab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7713e2a35bac8df1e8987fbcd452459fa3541caad0c52ed57fda187bd7c83d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7713e2a35bac8df1e8987fbcd452459fa3541caad0c52ed57fda187bd7c83d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34adfca89026fc20007a73a8cbd4e7c7b3137a581c3476867af17fabe976f71f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34adfca89026fc20007a73a8cbd4e7c7b3137a581c3476867af17fabe976f71f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qppnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:26Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.650791 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tgdlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2336955b-a703-4db4-8d0f-d08088ed59b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb0362942bfac635ee6292501b1b851772c01d75dd658eef4b17669a716de72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vjfp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tgdlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:26Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.662281 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62fcb632-f12a-46e2-bb7d-a5311f291ee7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28d43531927e75c3745f100940f7f6bf79fd21026c1309194df8cd7a39361b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69633332d03d2551ff4abb253d358de398eab18486b92db630ce06c0447fb9cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ccb78d9a51ac83ee208c2451685a2512de9deaf742a663bb8c892aa59b76162\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27826e155a9352b1e1866e83fdab683dbcd749cc04b6a4da2db25ecabe3d9a52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://991252e5f56f128827a5054b8e228284bc0266166582cbaedaadc53ca2f5b401\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:26Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.673728 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63b471fd-5e16-4761-8b07-44a135d5f5f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a68a57fa7fba8f351f264e427a50bed23f57775644e9c268aa7305efb6916794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67df7c609d5a5d658be4c3e873ec3052bd070ad9469f8ff44c29a9571e3e5d95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://269ced131bc0777dc6752070470c95b4bedfe4b4240940eabd769500ddef0404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e48c0e3b58359c6c0a493d58f7d2aec6011747791a3c0d7788102e1fec18865\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e48c0e3b58359c6c0a493d58f7d2aec6011747791a3c0d7788102e1fec18865\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:26Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.684462 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:26Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.694941 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f948b87-ac86-4de6-ad64-c2ef947f84d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4482f73a45bc82bd55770c3fd788becacec94893fed1c305fed1c81c753d43aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ba68bd093be94d0d40f3851638db5d26bdfc52d175b1ac78afb91dda7affcfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2krbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:26Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.737051 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.737096 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.737107 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.737127 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.737140 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:26Z","lastTransitionTime":"2025-11-25T10:33:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.838852 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.838893 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.838903 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.838920 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.838932 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:26Z","lastTransitionTime":"2025-11-25T10:33:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.941038 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.941077 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.941085 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.941100 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:26 crc kubenswrapper[4821]: I1125 10:33:26.941110 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:26Z","lastTransitionTime":"2025-11-25T10:33:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.014963 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.015036 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.015049 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.015065 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.015077 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:27Z","lastTransitionTime":"2025-11-25T10:33:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:27 crc kubenswrapper[4821]: E1125 10:33:27.027765 4821 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a5b2b070-39b9-412c-90f1-c7ed2e9e0950\\\",\\\"systemUUID\\\":\\\"ea86e80f-41b7-4190-ad7e-d8c549d535de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:27Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.031624 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.031673 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.031686 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.031703 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.031714 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:27Z","lastTransitionTime":"2025-11-25T10:33:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:27 crc kubenswrapper[4821]: E1125 10:33:27.050299 4821 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a5b2b070-39b9-412c-90f1-c7ed2e9e0950\\\",\\\"systemUUID\\\":\\\"ea86e80f-41b7-4190-ad7e-d8c549d535de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:27Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.057369 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.057412 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.057423 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.057441 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.057452 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:27Z","lastTransitionTime":"2025-11-25T10:33:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:27 crc kubenswrapper[4821]: E1125 10:33:27.068640 4821 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a5b2b070-39b9-412c-90f1-c7ed2e9e0950\\\",\\\"systemUUID\\\":\\\"ea86e80f-41b7-4190-ad7e-d8c549d535de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:27Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.071729 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.071763 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.071772 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.071786 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.071795 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:27Z","lastTransitionTime":"2025-11-25T10:33:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:27 crc kubenswrapper[4821]: E1125 10:33:27.085107 4821 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a5b2b070-39b9-412c-90f1-c7ed2e9e0950\\\",\\\"systemUUID\\\":\\\"ea86e80f-41b7-4190-ad7e-d8c549d535de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:27Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.088419 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.088456 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.088469 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.088486 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.088498 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:27Z","lastTransitionTime":"2025-11-25T10:33:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:27 crc kubenswrapper[4821]: E1125 10:33:27.099451 4821 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a5b2b070-39b9-412c-90f1-c7ed2e9e0950\\\",\\\"systemUUID\\\":\\\"ea86e80f-41b7-4190-ad7e-d8c549d535de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:27Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:27 crc kubenswrapper[4821]: E1125 10:33:27.099566 4821 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.100874 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.100898 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.100956 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.101005 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.101020 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:27Z","lastTransitionTime":"2025-11-25T10:33:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.113687 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.113752 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:33:27 crc kubenswrapper[4821]: E1125 10:33:27.113807 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.113692 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:33:27 crc kubenswrapper[4821]: E1125 10:33:27.113862 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:33:27 crc kubenswrapper[4821]: E1125 10:33:27.113926 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.202702 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.202760 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.202768 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.202779 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.202790 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:27Z","lastTransitionTime":"2025-11-25T10:33:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.306034 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.306106 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.306139 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.306218 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.306243 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:27Z","lastTransitionTime":"2025-11-25T10:33:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.408773 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.408819 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.408833 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.408853 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.408879 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:27Z","lastTransitionTime":"2025-11-25T10:33:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.455035 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mzhnw_954085fa-d03a-4e88-9dfd-566257e3558d/ovnkube-controller/3.log" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.455526 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mzhnw_954085fa-d03a-4e88-9dfd-566257e3558d/ovnkube-controller/2.log" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.459074 4821 generic.go:334] "Generic (PLEG): container finished" podID="954085fa-d03a-4e88-9dfd-566257e3558d" containerID="90d8792a79993c04a0eb774498395d955deff5c42c693892426685d157cf9cff" exitCode=1 Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.459105 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" event={"ID":"954085fa-d03a-4e88-9dfd-566257e3558d","Type":"ContainerDied","Data":"90d8792a79993c04a0eb774498395d955deff5c42c693892426685d157cf9cff"} Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.459187 4821 scope.go:117] "RemoveContainer" containerID="7e6202b51ac13209f3a6de880638a6d142f79755311646c0216f11dfe8bbe3f7" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.460118 4821 scope.go:117] "RemoveContainer" containerID="90d8792a79993c04a0eb774498395d955deff5c42c693892426685d157cf9cff" Nov 25 10:33:27 crc kubenswrapper[4821]: E1125 10:33:27.460454 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-mzhnw_openshift-ovn-kubernetes(954085fa-d03a-4e88-9dfd-566257e3558d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" podUID="954085fa-d03a-4e88-9dfd-566257e3558d" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.473971 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17b8714be3bd121e06641873e003f08560b1f71e56d6ea936966df7d4896eb03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85731e6b6dbe1777bc7161a010f39f79523ec4b08d71d5d8a975573aa7fe4957\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:27Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.484500 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zlk96" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d55aa91-9cb9-4c95-bc27-6fec05684c10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4521b5baee45ffc9c7661c781c0b3de458dff325ad525dde283d46af71c0cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vkd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zlk96\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:27Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.496777 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6hbqd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f71b3a2-63cb-443c-9f41-39e57030746c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65473b0e63c2e9c7c09a2bb8f31396daf1d87c2c83d20ed3508ce2fe7358dda3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxdmq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://595f9ea75461f18e763d84799e368becc07b72acdc5337d20d6995bf53c3d27a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxdmq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6hbqd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:27Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.508575 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:27Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.511233 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.511262 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.511270 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.511284 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.511293 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:27Z","lastTransitionTime":"2025-11-25T10:33:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.525679 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d9e58df-1755-4422-8d04-211f98c7c011\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea762f627aa69fb88a9bd82545fa677e9ccfa97f99a5e92c0aa490e619c0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a06b811e5ec52ce984a2403ba2c98332ba60466d08b3580ad9192fac6caef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0385f6ed9597974267f7c45547f6b6725b4ee83a2cc0508ff2d1c1ce8753ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b915b565f7fbb3e4e856e7f1035a519e82bfb3023660c121f6e3813de353bb92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18465f3a6916c36457a305ba5d568c9077f34f1d000532ed518f237d100008b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:27Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.537160 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25595ccd-762d-414a-9916-4fe38d478b07\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d4b2d8db4d13392d9789e161fd4b8f6eef983b79128ee584d72e4c06565bc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf780f791771ea43a8c8176d02fdfb5204406e023207299369ae46ceb342baaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0251519fd4262f2ea5c64959db49433afdafe36b7ed2404d660eeb7c365d2746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad5c205e0eeecfd43cbdf0eb1a334e3e8faa5e9a9e8dd202c397d5cbb36e081e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:27Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.548258 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:27Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.558756 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ldzbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b84e32ae-4897-4644-a656-6bd82012cc81\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://418dc8577fdc98324677dbccfcf9317b67cab4408432b30df2818b22f88d426e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ef433744761f26e8e1f6c0db123821a11923ffd774830744285cd4178d89533\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T10:33:19Z\\\",\\\"message\\\":\\\"2025-11-25T10:32:34+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_60a08f2b-f038-4c96-9f8b-1bd34a1965eb\\\\n2025-11-25T10:32:34+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_60a08f2b-f038-4c96-9f8b-1bd34a1965eb to /host/opt/cni/bin/\\\\n2025-11-25T10:32:34Z [verbose] multus-daemon started\\\\n2025-11-25T10:32:34Z [verbose] Readiness Indicator file check\\\\n2025-11-25T10:33:19Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:33:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5pmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ldzbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:27Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.574120 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"954085fa-d03a-4e88-9dfd-566257e3558d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f00746930c4588e548f0e2b60d52bf448aff7401b676a7855cb0f4cce00d329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a14a078ba7d2aed95992b36502008f315f0d3a634b64ed49da70398b2562d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d7254196b3702e48083ae98e829b73c539adf69ed5b2bbdc62189d09ae1762\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d727430108977574365f0e9c04b44e03307bd7ff89cdbb1e48a6e916f504d6b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94c7949f077b1ca2b310bedb090aa72fc4bd7e8f2c8bf438096edfca8568f36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed2e7cad7bb24b5b56a1295098a6a7c230c8f9b9c29a88ef653ae8b6ac31667d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90d8792a79993c04a0eb774498395d955deff5c42c693892426685d157cf9cff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e6202b51ac13209f3a6de880638a6d142f79755311646c0216f11dfe8bbe3f7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T10:32:56Z\\\",\\\"message\\\":\\\"to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:32:56Z is after 2025-08-24T17:21:41Z]\\\\nI1125 10:32:56.048189 6453 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-console/console_TCP_cluster\\\\\\\", UUID:\\\\\\\"d7d7b270-1480-47f8-bdf9-690dbab310cb\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console/console\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]strin\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d8792a79993c04a0eb774498395d955deff5c42c693892426685d157cf9cff\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T10:33:26Z\\\",\\\"message\\\":\\\" success event on pod openshift-ovn-kubernetes/ovnkube-node-mzhnw\\\\nI1125 10:33:26.842402 6839 ovn.go:134] Ensuring zone local for Pod openshift-network-node-identity/network-node-identity-vrzqb in node crc\\\\nI1125 10:33:26.842411 6839 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-apiserver/kube-apiserver-crc\\\\nF1125 10:33:26.842413 6839 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:26Z is after 2025-08-24T17:21:41Z]\\\\nI1125 10:33:26.842420 6839 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI1125 10:33:26.842421 6839 loadbalancer.go:304] Deleted 0 stale LB\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:33:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dbf499f4ae6161752abaa7a466a027c6f874016be90a534e69f07e3dd3f3d9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzhnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:27Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.583879 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-d4msn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f12cef3-7e40-4f61-836e-23cd0e578e71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s7lkv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s7lkv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:44Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-d4msn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:27Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.593010 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3e50766-ed07-4d22-95d0-3e0a9d822a17\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9525225d46e6f9225786ef8dd0ebe108d41efa26f6c94451c069aa5eb30da58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://733440f5c7493b5e1f4597ebb8f9d3d79507df0a1bc33393fb4534b5a2f2f077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://733440f5c7493b5e1f4597ebb8f9d3d79507df0a1bc33393fb4534b5a2f2f077\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:27Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.603296 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63b471fd-5e16-4761-8b07-44a135d5f5f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a68a57fa7fba8f351f264e427a50bed23f57775644e9c268aa7305efb6916794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67df7c609d5a5d658be4c3e873ec3052bd070ad9469f8ff44c29a9571e3e5d95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://269ced131bc0777dc6752070470c95b4bedfe4b4240940eabd769500ddef0404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e48c0e3b58359c6c0a493d58f7d2aec6011747791a3c0d7788102e1fec18865\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e48c0e3b58359c6c0a493d58f7d2aec6011747791a3c0d7788102e1fec18865\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:27Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.613642 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.613673 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.613682 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.613698 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.613710 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:27Z","lastTransitionTime":"2025-11-25T10:33:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.615257 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:27Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.624363 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f948b87-ac86-4de6-ad64-c2ef947f84d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4482f73a45bc82bd55770c3fd788becacec94893fed1c305fed1c81c753d43aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ba68bd093be94d0d40f3851638db5d26bdfc52d175b1ac78afb91dda7affcfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2krbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:27Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.635893 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qppnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"357491b8-cd9c-486e-9898-5ba33678513d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc88d700ca3a449ef1cb32d6124b16d0024d6981ced12e276083f86a1b322ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd9c047c5ce1d896f803ff6cafa9bbf3a18a391b44539e1f1312e08c89e51cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd9c047c5ce1d896f803ff6cafa9bbf3a18a391b44539e1f1312e08c89e51cb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://565a6db63fb69abbba81ffdf6085797900858d263cd6affcb0b5f195474c2516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://565a6db63fb69abbba81ffdf6085797900858d263cd6affcb0b5f195474c2516\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://544885302dffff8336a9aa3c4afa80d69088ffaa2402648f4a329e0a798a8c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://544885302dffff8336a9aa3c4afa80d69088ffaa2402648f4a329e0a798a8c99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8bc193fb3f85b980c31af2e826cee90adb633ad2ee82ff1d46e182ab52e8fab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8bc193fb3f85b980c31af2e826cee90adb633ad2ee82ff1d46e182ab52e8fab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7713e2a35bac8df1e8987fbcd452459fa3541caad0c52ed57fda187bd7c83d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7713e2a35bac8df1e8987fbcd452459fa3541caad0c52ed57fda187bd7c83d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34adfca89026fc20007a73a8cbd4e7c7b3137a581c3476867af17fabe976f71f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34adfca89026fc20007a73a8cbd4e7c7b3137a581c3476867af17fabe976f71f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qppnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:27Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.644473 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tgdlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2336955b-a703-4db4-8d0f-d08088ed59b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb0362942bfac635ee6292501b1b851772c01d75dd658eef4b17669a716de72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vjfp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tgdlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:27Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.659622 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62fcb632-f12a-46e2-bb7d-a5311f291ee7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28d43531927e75c3745f100940f7f6bf79fd21026c1309194df8cd7a39361b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69633332d03d2551ff4abb253d358de398eab18486b92db630ce06c0447fb9cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ccb78d9a51ac83ee208c2451685a2512de9deaf742a663bb8c892aa59b76162\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27826e155a9352b1e1866e83fdab683dbcd749cc04b6a4da2db25ecabe3d9a52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://991252e5f56f128827a5054b8e228284bc0266166582cbaedaadc53ca2f5b401\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:27Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.674280 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384d8ba58264ab89796d3b53afb0c4c1be33cb4fba719a8beef4afc2be7d66a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:27Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.684756 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06f949936bac99cfbf89d720c9549c074351d9077bd963fa024d87004c6be23e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:27Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.715667 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.715702 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.715712 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.715729 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.715739 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:27Z","lastTransitionTime":"2025-11-25T10:33:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.817681 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.817727 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.817741 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.817756 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.817766 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:27Z","lastTransitionTime":"2025-11-25T10:33:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.920042 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.920075 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.920084 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.920100 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:27 crc kubenswrapper[4821]: I1125 10:33:27.920112 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:27Z","lastTransitionTime":"2025-11-25T10:33:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:28 crc kubenswrapper[4821]: I1125 10:33:28.022122 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:28 crc kubenswrapper[4821]: I1125 10:33:28.022153 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:28 crc kubenswrapper[4821]: I1125 10:33:28.022186 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:28 crc kubenswrapper[4821]: I1125 10:33:28.022202 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:28 crc kubenswrapper[4821]: I1125 10:33:28.022210 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:28Z","lastTransitionTime":"2025-11-25T10:33:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:28 crc kubenswrapper[4821]: I1125 10:33:28.113822 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4msn" Nov 25 10:33:28 crc kubenswrapper[4821]: E1125 10:33:28.113942 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4msn" podUID="1f12cef3-7e40-4f61-836e-23cd0e578e71" Nov 25 10:33:28 crc kubenswrapper[4821]: I1125 10:33:28.123658 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:28 crc kubenswrapper[4821]: I1125 10:33:28.123717 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:28 crc kubenswrapper[4821]: I1125 10:33:28.123733 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:28 crc kubenswrapper[4821]: I1125 10:33:28.123757 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:28 crc kubenswrapper[4821]: I1125 10:33:28.123768 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:28Z","lastTransitionTime":"2025-11-25T10:33:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:28 crc kubenswrapper[4821]: I1125 10:33:28.226210 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:28 crc kubenswrapper[4821]: I1125 10:33:28.226246 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:28 crc kubenswrapper[4821]: I1125 10:33:28.226254 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:28 crc kubenswrapper[4821]: I1125 10:33:28.226268 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:28 crc kubenswrapper[4821]: I1125 10:33:28.226277 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:28Z","lastTransitionTime":"2025-11-25T10:33:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:28 crc kubenswrapper[4821]: I1125 10:33:28.328784 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:28 crc kubenswrapper[4821]: I1125 10:33:28.328822 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:28 crc kubenswrapper[4821]: I1125 10:33:28.328832 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:28 crc kubenswrapper[4821]: I1125 10:33:28.328848 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:28 crc kubenswrapper[4821]: I1125 10:33:28.328867 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:28Z","lastTransitionTime":"2025-11-25T10:33:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:28 crc kubenswrapper[4821]: I1125 10:33:28.431217 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:28 crc kubenswrapper[4821]: I1125 10:33:28.431299 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:28 crc kubenswrapper[4821]: I1125 10:33:28.431314 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:28 crc kubenswrapper[4821]: I1125 10:33:28.431332 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:28 crc kubenswrapper[4821]: I1125 10:33:28.431343 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:28Z","lastTransitionTime":"2025-11-25T10:33:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:28 crc kubenswrapper[4821]: I1125 10:33:28.464158 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mzhnw_954085fa-d03a-4e88-9dfd-566257e3558d/ovnkube-controller/3.log" Nov 25 10:33:28 crc kubenswrapper[4821]: I1125 10:33:28.468226 4821 scope.go:117] "RemoveContainer" containerID="90d8792a79993c04a0eb774498395d955deff5c42c693892426685d157cf9cff" Nov 25 10:33:28 crc kubenswrapper[4821]: E1125 10:33:28.468411 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-mzhnw_openshift-ovn-kubernetes(954085fa-d03a-4e88-9dfd-566257e3558d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" podUID="954085fa-d03a-4e88-9dfd-566257e3558d" Nov 25 10:33:28 crc kubenswrapper[4821]: I1125 10:33:28.485257 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62fcb632-f12a-46e2-bb7d-a5311f291ee7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28d43531927e75c3745f100940f7f6bf79fd21026c1309194df8cd7a39361b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69633332d03d2551ff4abb253d358de398eab18486b92db630ce06c0447fb9cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ccb78d9a51ac83ee208c2451685a2512de9deaf742a663bb8c892aa59b76162\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27826e155a9352b1e1866e83fdab683dbcd749cc04b6a4da2db25ecabe3d9a52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://991252e5f56f128827a5054b8e228284bc0266166582cbaedaadc53ca2f5b401\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:28Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:28 crc kubenswrapper[4821]: I1125 10:33:28.497918 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63b471fd-5e16-4761-8b07-44a135d5f5f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a68a57fa7fba8f351f264e427a50bed23f57775644e9c268aa7305efb6916794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67df7c609d5a5d658be4c3e873ec3052bd070ad9469f8ff44c29a9571e3e5d95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://269ced131bc0777dc6752070470c95b4bedfe4b4240940eabd769500ddef0404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e48c0e3b58359c6c0a493d58f7d2aec6011747791a3c0d7788102e1fec18865\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e48c0e3b58359c6c0a493d58f7d2aec6011747791a3c0d7788102e1fec18865\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:28Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:28 crc kubenswrapper[4821]: I1125 10:33:28.511570 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:28Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:28 crc kubenswrapper[4821]: I1125 10:33:28.524124 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f948b87-ac86-4de6-ad64-c2ef947f84d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4482f73a45bc82bd55770c3fd788becacec94893fed1c305fed1c81c753d43aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ba68bd093be94d0d40f3851638db5d26bdfc52d175b1ac78afb91dda7affcfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2krbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:28Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:28 crc kubenswrapper[4821]: I1125 10:33:28.534110 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:28 crc kubenswrapper[4821]: I1125 10:33:28.534148 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:28 crc kubenswrapper[4821]: I1125 10:33:28.534157 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:28 crc kubenswrapper[4821]: I1125 10:33:28.534171 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:28 crc kubenswrapper[4821]: I1125 10:33:28.534180 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:28Z","lastTransitionTime":"2025-11-25T10:33:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:28 crc kubenswrapper[4821]: I1125 10:33:28.540189 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qppnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"357491b8-cd9c-486e-9898-5ba33678513d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc88d700ca3a449ef1cb32d6124b16d0024d6981ced12e276083f86a1b322ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd9c047c5ce1d896f803ff6cafa9bbf3a18a391b44539e1f1312e08c89e51cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd9c047c5ce1d896f803ff6cafa9bbf3a18a391b44539e1f1312e08c89e51cb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://565a6db63fb69abbba81ffdf6085797900858d263cd6affcb0b5f195474c2516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://565a6db63fb69abbba81ffdf6085797900858d263cd6affcb0b5f195474c2516\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://544885302dffff8336a9aa3c4afa80d69088ffaa2402648f4a329e0a798a8c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://544885302dffff8336a9aa3c4afa80d69088ffaa2402648f4a329e0a798a8c99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8bc193fb3f85b980c31af2e826cee90adb633ad2ee82ff1d46e182ab52e8fab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8bc193fb3f85b980c31af2e826cee90adb633ad2ee82ff1d46e182ab52e8fab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7713e2a35bac8df1e8987fbcd452459fa3541caad0c52ed57fda187bd7c83d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7713e2a35bac8df1e8987fbcd452459fa3541caad0c52ed57fda187bd7c83d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34adfca89026fc20007a73a8cbd4e7c7b3137a581c3476867af17fabe976f71f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34adfca89026fc20007a73a8cbd4e7c7b3137a581c3476867af17fabe976f71f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qppnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:28Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:28 crc kubenswrapper[4821]: I1125 10:33:28.551425 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tgdlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2336955b-a703-4db4-8d0f-d08088ed59b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb0362942bfac635ee6292501b1b851772c01d75dd658eef4b17669a716de72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vjfp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tgdlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:28Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:28 crc kubenswrapper[4821]: I1125 10:33:28.564255 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06f949936bac99cfbf89d720c9549c074351d9077bd963fa024d87004c6be23e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:28Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:28 crc kubenswrapper[4821]: I1125 10:33:28.575768 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384d8ba58264ab89796d3b53afb0c4c1be33cb4fba719a8beef4afc2be7d66a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:28Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:28 crc kubenswrapper[4821]: I1125 10:33:28.587104 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:28Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:28 crc kubenswrapper[4821]: I1125 10:33:28.598487 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17b8714be3bd121e06641873e003f08560b1f71e56d6ea936966df7d4896eb03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85731e6b6dbe1777bc7161a010f39f79523ec4b08d71d5d8a975573aa7fe4957\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:28Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:28 crc kubenswrapper[4821]: I1125 10:33:28.606601 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zlk96" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d55aa91-9cb9-4c95-bc27-6fec05684c10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4521b5baee45ffc9c7661c781c0b3de458dff325ad525dde283d46af71c0cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vkd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zlk96\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:28Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:28 crc kubenswrapper[4821]: I1125 10:33:28.615855 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6hbqd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f71b3a2-63cb-443c-9f41-39e57030746c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65473b0e63c2e9c7c09a2bb8f31396daf1d87c2c83d20ed3508ce2fe7358dda3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxdmq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://595f9ea75461f18e763d84799e368becc07b72acdc5337d20d6995bf53c3d27a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxdmq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6hbqd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:28Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:28 crc kubenswrapper[4821]: I1125 10:33:28.624895 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-d4msn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f12cef3-7e40-4f61-836e-23cd0e578e71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s7lkv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s7lkv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:44Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-d4msn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:28Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:28 crc kubenswrapper[4821]: I1125 10:33:28.634286 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3e50766-ed07-4d22-95d0-3e0a9d822a17\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9525225d46e6f9225786ef8dd0ebe108d41efa26f6c94451c069aa5eb30da58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://733440f5c7493b5e1f4597ebb8f9d3d79507df0a1bc33393fb4534b5a2f2f077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://733440f5c7493b5e1f4597ebb8f9d3d79507df0a1bc33393fb4534b5a2f2f077\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:28Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:28 crc kubenswrapper[4821]: I1125 10:33:28.635947 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:28 crc kubenswrapper[4821]: I1125 10:33:28.635975 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:28 crc kubenswrapper[4821]: I1125 10:33:28.635985 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:28 crc kubenswrapper[4821]: I1125 10:33:28.635998 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:28 crc kubenswrapper[4821]: I1125 10:33:28.636007 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:28Z","lastTransitionTime":"2025-11-25T10:33:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:28 crc kubenswrapper[4821]: I1125 10:33:28.651937 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d9e58df-1755-4422-8d04-211f98c7c011\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea762f627aa69fb88a9bd82545fa677e9ccfa97f99a5e92c0aa490e619c0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a06b811e5ec52ce984a2403ba2c98332ba60466d08b3580ad9192fac6caef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0385f6ed9597974267f7c45547f6b6725b4ee83a2cc0508ff2d1c1ce8753ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b915b565f7fbb3e4e856e7f1035a519e82bfb3023660c121f6e3813de353bb92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18465f3a6916c36457a305ba5d568c9077f34f1d000532ed518f237d100008b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:28Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:28 crc kubenswrapper[4821]: I1125 10:33:28.663156 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25595ccd-762d-414a-9916-4fe38d478b07\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d4b2d8db4d13392d9789e161fd4b8f6eef983b79128ee584d72e4c06565bc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf780f791771ea43a8c8176d02fdfb5204406e023207299369ae46ceb342baaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0251519fd4262f2ea5c64959db49433afdafe36b7ed2404d660eeb7c365d2746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad5c205e0eeecfd43cbdf0eb1a334e3e8faa5e9a9e8dd202c397d5cbb36e081e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:28Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:28 crc kubenswrapper[4821]: I1125 10:33:28.673919 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:28Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:28 crc kubenswrapper[4821]: I1125 10:33:28.688606 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ldzbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b84e32ae-4897-4644-a656-6bd82012cc81\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://418dc8577fdc98324677dbccfcf9317b67cab4408432b30df2818b22f88d426e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ef433744761f26e8e1f6c0db123821a11923ffd774830744285cd4178d89533\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T10:33:19Z\\\",\\\"message\\\":\\\"2025-11-25T10:32:34+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_60a08f2b-f038-4c96-9f8b-1bd34a1965eb\\\\n2025-11-25T10:32:34+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_60a08f2b-f038-4c96-9f8b-1bd34a1965eb to /host/opt/cni/bin/\\\\n2025-11-25T10:32:34Z [verbose] multus-daemon started\\\\n2025-11-25T10:32:34Z [verbose] Readiness Indicator file check\\\\n2025-11-25T10:33:19Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:33:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5pmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ldzbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:28Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:28 crc kubenswrapper[4821]: I1125 10:33:28.708375 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"954085fa-d03a-4e88-9dfd-566257e3558d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f00746930c4588e548f0e2b60d52bf448aff7401b676a7855cb0f4cce00d329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a14a078ba7d2aed95992b36502008f315f0d3a634b64ed49da70398b2562d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d7254196b3702e48083ae98e829b73c539adf69ed5b2bbdc62189d09ae1762\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d727430108977574365f0e9c04b44e03307bd7ff89cdbb1e48a6e916f504d6b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94c7949f077b1ca2b310bedb090aa72fc4bd7e8f2c8bf438096edfca8568f36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed2e7cad7bb24b5b56a1295098a6a7c230c8f9b9c29a88ef653ae8b6ac31667d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90d8792a79993c04a0eb774498395d955deff5c42c693892426685d157cf9cff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d8792a79993c04a0eb774498395d955deff5c42c693892426685d157cf9cff\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T10:33:26Z\\\",\\\"message\\\":\\\" success event on pod openshift-ovn-kubernetes/ovnkube-node-mzhnw\\\\nI1125 10:33:26.842402 6839 ovn.go:134] Ensuring zone local for Pod openshift-network-node-identity/network-node-identity-vrzqb in node crc\\\\nI1125 10:33:26.842411 6839 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-apiserver/kube-apiserver-crc\\\\nF1125 10:33:26.842413 6839 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:26Z is after 2025-08-24T17:21:41Z]\\\\nI1125 10:33:26.842420 6839 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI1125 10:33:26.842421 6839 loadbalancer.go:304] Deleted 0 stale LB\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:33:26Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-mzhnw_openshift-ovn-kubernetes(954085fa-d03a-4e88-9dfd-566257e3558d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dbf499f4ae6161752abaa7a466a027c6f874016be90a534e69f07e3dd3f3d9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzhnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:28Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:28 crc kubenswrapper[4821]: I1125 10:33:28.738713 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:28 crc kubenswrapper[4821]: I1125 10:33:28.738761 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:28 crc kubenswrapper[4821]: I1125 10:33:28.738774 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:28 crc kubenswrapper[4821]: I1125 10:33:28.738791 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:28 crc kubenswrapper[4821]: I1125 10:33:28.738803 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:28Z","lastTransitionTime":"2025-11-25T10:33:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:28 crc kubenswrapper[4821]: I1125 10:33:28.841322 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:28 crc kubenswrapper[4821]: I1125 10:33:28.841371 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:28 crc kubenswrapper[4821]: I1125 10:33:28.841383 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:28 crc kubenswrapper[4821]: I1125 10:33:28.841398 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:28 crc kubenswrapper[4821]: I1125 10:33:28.841409 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:28Z","lastTransitionTime":"2025-11-25T10:33:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:28 crc kubenswrapper[4821]: I1125 10:33:28.943544 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:28 crc kubenswrapper[4821]: I1125 10:33:28.943582 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:28 crc kubenswrapper[4821]: I1125 10:33:28.943594 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:28 crc kubenswrapper[4821]: I1125 10:33:28.943610 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:28 crc kubenswrapper[4821]: I1125 10:33:28.943619 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:28Z","lastTransitionTime":"2025-11-25T10:33:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:29 crc kubenswrapper[4821]: I1125 10:33:29.046387 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:29 crc kubenswrapper[4821]: I1125 10:33:29.046501 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:29 crc kubenswrapper[4821]: I1125 10:33:29.046521 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:29 crc kubenswrapper[4821]: I1125 10:33:29.046540 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:29 crc kubenswrapper[4821]: I1125 10:33:29.046552 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:29Z","lastTransitionTime":"2025-11-25T10:33:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:29 crc kubenswrapper[4821]: I1125 10:33:29.113971 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:33:29 crc kubenswrapper[4821]: I1125 10:33:29.113994 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:33:29 crc kubenswrapper[4821]: I1125 10:33:29.114011 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:33:29 crc kubenswrapper[4821]: E1125 10:33:29.114094 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:33:29 crc kubenswrapper[4821]: E1125 10:33:29.114204 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:33:29 crc kubenswrapper[4821]: E1125 10:33:29.114284 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:33:29 crc kubenswrapper[4821]: I1125 10:33:29.149413 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:29 crc kubenswrapper[4821]: I1125 10:33:29.149448 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:29 crc kubenswrapper[4821]: I1125 10:33:29.149456 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:29 crc kubenswrapper[4821]: I1125 10:33:29.149472 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:29 crc kubenswrapper[4821]: I1125 10:33:29.149484 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:29Z","lastTransitionTime":"2025-11-25T10:33:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:29 crc kubenswrapper[4821]: I1125 10:33:29.251963 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:29 crc kubenswrapper[4821]: I1125 10:33:29.252024 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:29 crc kubenswrapper[4821]: I1125 10:33:29.252040 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:29 crc kubenswrapper[4821]: I1125 10:33:29.252065 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:29 crc kubenswrapper[4821]: I1125 10:33:29.252083 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:29Z","lastTransitionTime":"2025-11-25T10:33:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:29 crc kubenswrapper[4821]: I1125 10:33:29.354444 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:29 crc kubenswrapper[4821]: I1125 10:33:29.354469 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:29 crc kubenswrapper[4821]: I1125 10:33:29.354477 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:29 crc kubenswrapper[4821]: I1125 10:33:29.354494 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:29 crc kubenswrapper[4821]: I1125 10:33:29.354502 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:29Z","lastTransitionTime":"2025-11-25T10:33:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:29 crc kubenswrapper[4821]: I1125 10:33:29.456651 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:29 crc kubenswrapper[4821]: I1125 10:33:29.456687 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:29 crc kubenswrapper[4821]: I1125 10:33:29.456696 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:29 crc kubenswrapper[4821]: I1125 10:33:29.456711 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:29 crc kubenswrapper[4821]: I1125 10:33:29.456721 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:29Z","lastTransitionTime":"2025-11-25T10:33:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:29 crc kubenswrapper[4821]: I1125 10:33:29.559212 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:29 crc kubenswrapper[4821]: I1125 10:33:29.559271 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:29 crc kubenswrapper[4821]: I1125 10:33:29.559285 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:29 crc kubenswrapper[4821]: I1125 10:33:29.559300 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:29 crc kubenswrapper[4821]: I1125 10:33:29.559310 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:29Z","lastTransitionTime":"2025-11-25T10:33:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:29 crc kubenswrapper[4821]: I1125 10:33:29.662304 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:29 crc kubenswrapper[4821]: I1125 10:33:29.662341 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:29 crc kubenswrapper[4821]: I1125 10:33:29.662351 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:29 crc kubenswrapper[4821]: I1125 10:33:29.662365 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:29 crc kubenswrapper[4821]: I1125 10:33:29.662376 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:29Z","lastTransitionTime":"2025-11-25T10:33:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:29 crc kubenswrapper[4821]: I1125 10:33:29.764554 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:29 crc kubenswrapper[4821]: I1125 10:33:29.764586 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:29 crc kubenswrapper[4821]: I1125 10:33:29.764595 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:29 crc kubenswrapper[4821]: I1125 10:33:29.764608 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:29 crc kubenswrapper[4821]: I1125 10:33:29.764617 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:29Z","lastTransitionTime":"2025-11-25T10:33:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:29 crc kubenswrapper[4821]: I1125 10:33:29.866407 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:29 crc kubenswrapper[4821]: I1125 10:33:29.866449 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:29 crc kubenswrapper[4821]: I1125 10:33:29.866459 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:29 crc kubenswrapper[4821]: I1125 10:33:29.866474 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:29 crc kubenswrapper[4821]: I1125 10:33:29.866487 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:29Z","lastTransitionTime":"2025-11-25T10:33:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:29 crc kubenswrapper[4821]: I1125 10:33:29.969380 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:29 crc kubenswrapper[4821]: I1125 10:33:29.969457 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:29 crc kubenswrapper[4821]: I1125 10:33:29.969471 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:29 crc kubenswrapper[4821]: I1125 10:33:29.969491 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:29 crc kubenswrapper[4821]: I1125 10:33:29.969503 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:29Z","lastTransitionTime":"2025-11-25T10:33:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:30 crc kubenswrapper[4821]: I1125 10:33:30.071528 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:30 crc kubenswrapper[4821]: I1125 10:33:30.071593 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:30 crc kubenswrapper[4821]: I1125 10:33:30.071609 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:30 crc kubenswrapper[4821]: I1125 10:33:30.071628 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:30 crc kubenswrapper[4821]: I1125 10:33:30.071642 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:30Z","lastTransitionTime":"2025-11-25T10:33:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:30 crc kubenswrapper[4821]: I1125 10:33:30.114092 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4msn" Nov 25 10:33:30 crc kubenswrapper[4821]: E1125 10:33:30.114270 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4msn" podUID="1f12cef3-7e40-4f61-836e-23cd0e578e71" Nov 25 10:33:30 crc kubenswrapper[4821]: I1125 10:33:30.129811 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06f949936bac99cfbf89d720c9549c074351d9077bd963fa024d87004c6be23e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:30Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:30 crc kubenswrapper[4821]: I1125 10:33:30.144469 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384d8ba58264ab89796d3b53afb0c4c1be33cb4fba719a8beef4afc2be7d66a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:30Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:30 crc kubenswrapper[4821]: I1125 10:33:30.154453 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zlk96" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d55aa91-9cb9-4c95-bc27-6fec05684c10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4521b5baee45ffc9c7661c781c0b3de458dff325ad525dde283d46af71c0cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vkd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zlk96\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:30Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:30 crc kubenswrapper[4821]: I1125 10:33:30.172439 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6hbqd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f71b3a2-63cb-443c-9f41-39e57030746c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65473b0e63c2e9c7c09a2bb8f31396daf1d87c2c83d20ed3508ce2fe7358dda3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxdmq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://595f9ea75461f18e763d84799e368becc07b72acdc5337d20d6995bf53c3d27a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxdmq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6hbqd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:30Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:30 crc kubenswrapper[4821]: I1125 10:33:30.174844 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:30 crc kubenswrapper[4821]: I1125 10:33:30.174938 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:30 crc kubenswrapper[4821]: I1125 10:33:30.174959 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:30 crc kubenswrapper[4821]: I1125 10:33:30.174992 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:30 crc kubenswrapper[4821]: I1125 10:33:30.175012 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:30Z","lastTransitionTime":"2025-11-25T10:33:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:30 crc kubenswrapper[4821]: I1125 10:33:30.186804 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:30Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:30 crc kubenswrapper[4821]: I1125 10:33:30.202069 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17b8714be3bd121e06641873e003f08560b1f71e56d6ea936966df7d4896eb03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85731e6b6dbe1777bc7161a010f39f79523ec4b08d71d5d8a975573aa7fe4957\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:30Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:30 crc kubenswrapper[4821]: I1125 10:33:30.217033 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25595ccd-762d-414a-9916-4fe38d478b07\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d4b2d8db4d13392d9789e161fd4b8f6eef983b79128ee584d72e4c06565bc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf780f791771ea43a8c8176d02fdfb5204406e023207299369ae46ceb342baaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0251519fd4262f2ea5c64959db49433afdafe36b7ed2404d660eeb7c365d2746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad5c205e0eeecfd43cbdf0eb1a334e3e8faa5e9a9e8dd202c397d5cbb36e081e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:30Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:30 crc kubenswrapper[4821]: I1125 10:33:30.231222 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:30Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:30 crc kubenswrapper[4821]: I1125 10:33:30.243711 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ldzbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b84e32ae-4897-4644-a656-6bd82012cc81\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://418dc8577fdc98324677dbccfcf9317b67cab4408432b30df2818b22f88d426e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ef433744761f26e8e1f6c0db123821a11923ffd774830744285cd4178d89533\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T10:33:19Z\\\",\\\"message\\\":\\\"2025-11-25T10:32:34+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_60a08f2b-f038-4c96-9f8b-1bd34a1965eb\\\\n2025-11-25T10:32:34+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_60a08f2b-f038-4c96-9f8b-1bd34a1965eb to /host/opt/cni/bin/\\\\n2025-11-25T10:32:34Z [verbose] multus-daemon started\\\\n2025-11-25T10:32:34Z [verbose] Readiness Indicator file check\\\\n2025-11-25T10:33:19Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:33:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5pmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ldzbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:30Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:30 crc kubenswrapper[4821]: I1125 10:33:30.261500 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"954085fa-d03a-4e88-9dfd-566257e3558d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f00746930c4588e548f0e2b60d52bf448aff7401b676a7855cb0f4cce00d329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a14a078ba7d2aed95992b36502008f315f0d3a634b64ed49da70398b2562d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d7254196b3702e48083ae98e829b73c539adf69ed5b2bbdc62189d09ae1762\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d727430108977574365f0e9c04b44e03307bd7ff89cdbb1e48a6e916f504d6b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94c7949f077b1ca2b310bedb090aa72fc4bd7e8f2c8bf438096edfca8568f36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed2e7cad7bb24b5b56a1295098a6a7c230c8f9b9c29a88ef653ae8b6ac31667d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90d8792a79993c04a0eb774498395d955deff5c42c693892426685d157cf9cff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d8792a79993c04a0eb774498395d955deff5c42c693892426685d157cf9cff\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T10:33:26Z\\\",\\\"message\\\":\\\" success event on pod openshift-ovn-kubernetes/ovnkube-node-mzhnw\\\\nI1125 10:33:26.842402 6839 ovn.go:134] Ensuring zone local for Pod openshift-network-node-identity/network-node-identity-vrzqb in node crc\\\\nI1125 10:33:26.842411 6839 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-apiserver/kube-apiserver-crc\\\\nF1125 10:33:26.842413 6839 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:26Z is after 2025-08-24T17:21:41Z]\\\\nI1125 10:33:26.842420 6839 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI1125 10:33:26.842421 6839 loadbalancer.go:304] Deleted 0 stale LB\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:33:26Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-mzhnw_openshift-ovn-kubernetes(954085fa-d03a-4e88-9dfd-566257e3558d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dbf499f4ae6161752abaa7a466a027c6f874016be90a534e69f07e3dd3f3d9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzhnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:30Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:30 crc kubenswrapper[4821]: I1125 10:33:30.276020 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-d4msn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f12cef3-7e40-4f61-836e-23cd0e578e71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s7lkv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s7lkv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:44Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-d4msn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:30Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:30 crc kubenswrapper[4821]: I1125 10:33:30.277194 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:30 crc kubenswrapper[4821]: I1125 10:33:30.277240 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:30 crc kubenswrapper[4821]: I1125 10:33:30.277253 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:30 crc kubenswrapper[4821]: I1125 10:33:30.277271 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:30 crc kubenswrapper[4821]: I1125 10:33:30.277283 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:30Z","lastTransitionTime":"2025-11-25T10:33:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:30 crc kubenswrapper[4821]: I1125 10:33:30.287118 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3e50766-ed07-4d22-95d0-3e0a9d822a17\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9525225d46e6f9225786ef8dd0ebe108d41efa26f6c94451c069aa5eb30da58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://733440f5c7493b5e1f4597ebb8f9d3d79507df0a1bc33393fb4534b5a2f2f077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://733440f5c7493b5e1f4597ebb8f9d3d79507df0a1bc33393fb4534b5a2f2f077\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:30Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:30 crc kubenswrapper[4821]: I1125 10:33:30.307770 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d9e58df-1755-4422-8d04-211f98c7c011\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea762f627aa69fb88a9bd82545fa677e9ccfa97f99a5e92c0aa490e619c0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a06b811e5ec52ce984a2403ba2c98332ba60466d08b3580ad9192fac6caef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0385f6ed9597974267f7c45547f6b6725b4ee83a2cc0508ff2d1c1ce8753ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b915b565f7fbb3e4e856e7f1035a519e82bfb3023660c121f6e3813de353bb92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18465f3a6916c36457a305ba5d568c9077f34f1d000532ed518f237d100008b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:30Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:30 crc kubenswrapper[4821]: I1125 10:33:30.322954 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:30Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:30 crc kubenswrapper[4821]: I1125 10:33:30.336704 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f948b87-ac86-4de6-ad64-c2ef947f84d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4482f73a45bc82bd55770c3fd788becacec94893fed1c305fed1c81c753d43aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ba68bd093be94d0d40f3851638db5d26bdfc52d175b1ac78afb91dda7affcfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2krbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:30Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:30 crc kubenswrapper[4821]: I1125 10:33:30.351343 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qppnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"357491b8-cd9c-486e-9898-5ba33678513d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc88d700ca3a449ef1cb32d6124b16d0024d6981ced12e276083f86a1b322ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd9c047c5ce1d896f803ff6cafa9bbf3a18a391b44539e1f1312e08c89e51cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd9c047c5ce1d896f803ff6cafa9bbf3a18a391b44539e1f1312e08c89e51cb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://565a6db63fb69abbba81ffdf6085797900858d263cd6affcb0b5f195474c2516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://565a6db63fb69abbba81ffdf6085797900858d263cd6affcb0b5f195474c2516\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://544885302dffff8336a9aa3c4afa80d69088ffaa2402648f4a329e0a798a8c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://544885302dffff8336a9aa3c4afa80d69088ffaa2402648f4a329e0a798a8c99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8bc193fb3f85b980c31af2e826cee90adb633ad2ee82ff1d46e182ab52e8fab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8bc193fb3f85b980c31af2e826cee90adb633ad2ee82ff1d46e182ab52e8fab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7713e2a35bac8df1e8987fbcd452459fa3541caad0c52ed57fda187bd7c83d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7713e2a35bac8df1e8987fbcd452459fa3541caad0c52ed57fda187bd7c83d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34adfca89026fc20007a73a8cbd4e7c7b3137a581c3476867af17fabe976f71f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34adfca89026fc20007a73a8cbd4e7c7b3137a581c3476867af17fabe976f71f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qppnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:30Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:30 crc kubenswrapper[4821]: I1125 10:33:30.367554 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tgdlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2336955b-a703-4db4-8d0f-d08088ed59b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb0362942bfac635ee6292501b1b851772c01d75dd658eef4b17669a716de72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vjfp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tgdlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:30Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:30 crc kubenswrapper[4821]: I1125 10:33:30.379536 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:30 crc kubenswrapper[4821]: I1125 10:33:30.379592 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:30 crc kubenswrapper[4821]: I1125 10:33:30.379605 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:30 crc kubenswrapper[4821]: I1125 10:33:30.379624 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:30 crc kubenswrapper[4821]: I1125 10:33:30.379636 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:30Z","lastTransitionTime":"2025-11-25T10:33:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:30 crc kubenswrapper[4821]: I1125 10:33:30.382896 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62fcb632-f12a-46e2-bb7d-a5311f291ee7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28d43531927e75c3745f100940f7f6bf79fd21026c1309194df8cd7a39361b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69633332d03d2551ff4abb253d358de398eab18486b92db630ce06c0447fb9cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ccb78d9a51ac83ee208c2451685a2512de9deaf742a663bb8c892aa59b76162\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27826e155a9352b1e1866e83fdab683dbcd749cc04b6a4da2db25ecabe3d9a52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://991252e5f56f128827a5054b8e228284bc0266166582cbaedaadc53ca2f5b401\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:30Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:30 crc kubenswrapper[4821]: I1125 10:33:30.397606 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63b471fd-5e16-4761-8b07-44a135d5f5f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a68a57fa7fba8f351f264e427a50bed23f57775644e9c268aa7305efb6916794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67df7c609d5a5d658be4c3e873ec3052bd070ad9469f8ff44c29a9571e3e5d95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://269ced131bc0777dc6752070470c95b4bedfe4b4240940eabd769500ddef0404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e48c0e3b58359c6c0a493d58f7d2aec6011747791a3c0d7788102e1fec18865\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e48c0e3b58359c6c0a493d58f7d2aec6011747791a3c0d7788102e1fec18865\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:30Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:30 crc kubenswrapper[4821]: I1125 10:33:30.481576 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:30 crc kubenswrapper[4821]: I1125 10:33:30.481628 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:30 crc kubenswrapper[4821]: I1125 10:33:30.481638 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:30 crc kubenswrapper[4821]: I1125 10:33:30.481654 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:30 crc kubenswrapper[4821]: I1125 10:33:30.481668 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:30Z","lastTransitionTime":"2025-11-25T10:33:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:30 crc kubenswrapper[4821]: I1125 10:33:30.584878 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:30 crc kubenswrapper[4821]: I1125 10:33:30.584921 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:30 crc kubenswrapper[4821]: I1125 10:33:30.584930 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:30 crc kubenswrapper[4821]: I1125 10:33:30.584946 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:30 crc kubenswrapper[4821]: I1125 10:33:30.584956 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:30Z","lastTransitionTime":"2025-11-25T10:33:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:30 crc kubenswrapper[4821]: I1125 10:33:30.686672 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:30 crc kubenswrapper[4821]: I1125 10:33:30.686717 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:30 crc kubenswrapper[4821]: I1125 10:33:30.686725 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:30 crc kubenswrapper[4821]: I1125 10:33:30.686739 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:30 crc kubenswrapper[4821]: I1125 10:33:30.686748 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:30Z","lastTransitionTime":"2025-11-25T10:33:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:30 crc kubenswrapper[4821]: I1125 10:33:30.789272 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:30 crc kubenswrapper[4821]: I1125 10:33:30.789335 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:30 crc kubenswrapper[4821]: I1125 10:33:30.789345 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:30 crc kubenswrapper[4821]: I1125 10:33:30.789360 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:30 crc kubenswrapper[4821]: I1125 10:33:30.789372 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:30Z","lastTransitionTime":"2025-11-25T10:33:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:30 crc kubenswrapper[4821]: I1125 10:33:30.892584 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:30 crc kubenswrapper[4821]: I1125 10:33:30.892643 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:30 crc kubenswrapper[4821]: I1125 10:33:30.892656 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:30 crc kubenswrapper[4821]: I1125 10:33:30.892676 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:30 crc kubenswrapper[4821]: I1125 10:33:30.892689 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:30Z","lastTransitionTime":"2025-11-25T10:33:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:30 crc kubenswrapper[4821]: I1125 10:33:30.995366 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:30 crc kubenswrapper[4821]: I1125 10:33:30.995410 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:30 crc kubenswrapper[4821]: I1125 10:33:30.995419 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:30 crc kubenswrapper[4821]: I1125 10:33:30.995435 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:30 crc kubenswrapper[4821]: I1125 10:33:30.995444 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:30Z","lastTransitionTime":"2025-11-25T10:33:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:31 crc kubenswrapper[4821]: I1125 10:33:31.097787 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:31 crc kubenswrapper[4821]: I1125 10:33:31.097835 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:31 crc kubenswrapper[4821]: I1125 10:33:31.097845 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:31 crc kubenswrapper[4821]: I1125 10:33:31.097863 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:31 crc kubenswrapper[4821]: I1125 10:33:31.097873 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:31Z","lastTransitionTime":"2025-11-25T10:33:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:31 crc kubenswrapper[4821]: I1125 10:33:31.114203 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:33:31 crc kubenswrapper[4821]: I1125 10:33:31.114247 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:33:31 crc kubenswrapper[4821]: I1125 10:33:31.114272 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:33:31 crc kubenswrapper[4821]: E1125 10:33:31.114363 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:33:31 crc kubenswrapper[4821]: E1125 10:33:31.114473 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:33:31 crc kubenswrapper[4821]: E1125 10:33:31.114553 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:33:31 crc kubenswrapper[4821]: I1125 10:33:31.200514 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:31 crc kubenswrapper[4821]: I1125 10:33:31.200608 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:31 crc kubenswrapper[4821]: I1125 10:33:31.200620 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:31 crc kubenswrapper[4821]: I1125 10:33:31.200653 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:31 crc kubenswrapper[4821]: I1125 10:33:31.200665 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:31Z","lastTransitionTime":"2025-11-25T10:33:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:31 crc kubenswrapper[4821]: I1125 10:33:31.303400 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:31 crc kubenswrapper[4821]: I1125 10:33:31.303430 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:31 crc kubenswrapper[4821]: I1125 10:33:31.303438 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:31 crc kubenswrapper[4821]: I1125 10:33:31.303451 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:31 crc kubenswrapper[4821]: I1125 10:33:31.303460 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:31Z","lastTransitionTime":"2025-11-25T10:33:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:31 crc kubenswrapper[4821]: I1125 10:33:31.405336 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:31 crc kubenswrapper[4821]: I1125 10:33:31.405370 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:31 crc kubenswrapper[4821]: I1125 10:33:31.405395 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:31 crc kubenswrapper[4821]: I1125 10:33:31.405411 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:31 crc kubenswrapper[4821]: I1125 10:33:31.405420 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:31Z","lastTransitionTime":"2025-11-25T10:33:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:31 crc kubenswrapper[4821]: I1125 10:33:31.508740 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:31 crc kubenswrapper[4821]: I1125 10:33:31.509096 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:31 crc kubenswrapper[4821]: I1125 10:33:31.509108 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:31 crc kubenswrapper[4821]: I1125 10:33:31.509125 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:31 crc kubenswrapper[4821]: I1125 10:33:31.509139 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:31Z","lastTransitionTime":"2025-11-25T10:33:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:31 crc kubenswrapper[4821]: I1125 10:33:31.611668 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:31 crc kubenswrapper[4821]: I1125 10:33:31.611730 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:31 crc kubenswrapper[4821]: I1125 10:33:31.611748 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:31 crc kubenswrapper[4821]: I1125 10:33:31.611765 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:31 crc kubenswrapper[4821]: I1125 10:33:31.611776 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:31Z","lastTransitionTime":"2025-11-25T10:33:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:31 crc kubenswrapper[4821]: I1125 10:33:31.714051 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:31 crc kubenswrapper[4821]: I1125 10:33:31.714101 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:31 crc kubenswrapper[4821]: I1125 10:33:31.714112 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:31 crc kubenswrapper[4821]: I1125 10:33:31.714133 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:31 crc kubenswrapper[4821]: I1125 10:33:31.714147 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:31Z","lastTransitionTime":"2025-11-25T10:33:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:31 crc kubenswrapper[4821]: I1125 10:33:31.816151 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:31 crc kubenswrapper[4821]: I1125 10:33:31.816232 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:31 crc kubenswrapper[4821]: I1125 10:33:31.816249 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:31 crc kubenswrapper[4821]: I1125 10:33:31.816271 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:31 crc kubenswrapper[4821]: I1125 10:33:31.816287 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:31Z","lastTransitionTime":"2025-11-25T10:33:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:31 crc kubenswrapper[4821]: I1125 10:33:31.918248 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:31 crc kubenswrapper[4821]: I1125 10:33:31.918295 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:31 crc kubenswrapper[4821]: I1125 10:33:31.918310 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:31 crc kubenswrapper[4821]: I1125 10:33:31.918330 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:31 crc kubenswrapper[4821]: I1125 10:33:31.918347 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:31Z","lastTransitionTime":"2025-11-25T10:33:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:32 crc kubenswrapper[4821]: I1125 10:33:32.020848 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:32 crc kubenswrapper[4821]: I1125 10:33:32.021128 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:32 crc kubenswrapper[4821]: I1125 10:33:32.021223 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:32 crc kubenswrapper[4821]: I1125 10:33:32.021295 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:32 crc kubenswrapper[4821]: I1125 10:33:32.021356 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:32Z","lastTransitionTime":"2025-11-25T10:33:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:32 crc kubenswrapper[4821]: I1125 10:33:32.113459 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4msn" Nov 25 10:33:32 crc kubenswrapper[4821]: E1125 10:33:32.113657 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4msn" podUID="1f12cef3-7e40-4f61-836e-23cd0e578e71" Nov 25 10:33:32 crc kubenswrapper[4821]: I1125 10:33:32.123621 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:32 crc kubenswrapper[4821]: I1125 10:33:32.123662 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:32 crc kubenswrapper[4821]: I1125 10:33:32.123673 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:32 crc kubenswrapper[4821]: I1125 10:33:32.123687 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:32 crc kubenswrapper[4821]: I1125 10:33:32.123699 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:32Z","lastTransitionTime":"2025-11-25T10:33:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:32 crc kubenswrapper[4821]: I1125 10:33:32.225839 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:32 crc kubenswrapper[4821]: I1125 10:33:32.225880 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:32 crc kubenswrapper[4821]: I1125 10:33:32.225892 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:32 crc kubenswrapper[4821]: I1125 10:33:32.225909 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:32 crc kubenswrapper[4821]: I1125 10:33:32.225923 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:32Z","lastTransitionTime":"2025-11-25T10:33:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:32 crc kubenswrapper[4821]: I1125 10:33:32.328622 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:32 crc kubenswrapper[4821]: I1125 10:33:32.328660 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:32 crc kubenswrapper[4821]: I1125 10:33:32.328672 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:32 crc kubenswrapper[4821]: I1125 10:33:32.328689 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:32 crc kubenswrapper[4821]: I1125 10:33:32.328700 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:32Z","lastTransitionTime":"2025-11-25T10:33:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:32 crc kubenswrapper[4821]: I1125 10:33:32.431106 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:32 crc kubenswrapper[4821]: I1125 10:33:32.431153 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:32 crc kubenswrapper[4821]: I1125 10:33:32.431186 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:32 crc kubenswrapper[4821]: I1125 10:33:32.431208 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:32 crc kubenswrapper[4821]: I1125 10:33:32.431221 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:32Z","lastTransitionTime":"2025-11-25T10:33:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:32 crc kubenswrapper[4821]: I1125 10:33:32.534318 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:32 crc kubenswrapper[4821]: I1125 10:33:32.534371 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:32 crc kubenswrapper[4821]: I1125 10:33:32.534379 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:32 crc kubenswrapper[4821]: I1125 10:33:32.534393 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:32 crc kubenswrapper[4821]: I1125 10:33:32.534405 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:32Z","lastTransitionTime":"2025-11-25T10:33:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:32 crc kubenswrapper[4821]: I1125 10:33:32.636628 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:32 crc kubenswrapper[4821]: I1125 10:33:32.636668 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:32 crc kubenswrapper[4821]: I1125 10:33:32.636677 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:32 crc kubenswrapper[4821]: I1125 10:33:32.636690 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:32 crc kubenswrapper[4821]: I1125 10:33:32.636698 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:32Z","lastTransitionTime":"2025-11-25T10:33:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:32 crc kubenswrapper[4821]: I1125 10:33:32.738378 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:32 crc kubenswrapper[4821]: I1125 10:33:32.738415 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:32 crc kubenswrapper[4821]: I1125 10:33:32.738438 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:32 crc kubenswrapper[4821]: I1125 10:33:32.738455 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:32 crc kubenswrapper[4821]: I1125 10:33:32.738464 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:32Z","lastTransitionTime":"2025-11-25T10:33:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:32 crc kubenswrapper[4821]: I1125 10:33:32.840408 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:32 crc kubenswrapper[4821]: I1125 10:33:32.840460 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:32 crc kubenswrapper[4821]: I1125 10:33:32.840471 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:32 crc kubenswrapper[4821]: I1125 10:33:32.840489 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:32 crc kubenswrapper[4821]: I1125 10:33:32.840501 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:32Z","lastTransitionTime":"2025-11-25T10:33:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:32 crc kubenswrapper[4821]: I1125 10:33:32.943237 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:32 crc kubenswrapper[4821]: I1125 10:33:32.943282 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:32 crc kubenswrapper[4821]: I1125 10:33:32.943301 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:32 crc kubenswrapper[4821]: I1125 10:33:32.943317 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:32 crc kubenswrapper[4821]: I1125 10:33:32.943330 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:32Z","lastTransitionTime":"2025-11-25T10:33:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:33 crc kubenswrapper[4821]: I1125 10:33:33.046213 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:33 crc kubenswrapper[4821]: I1125 10:33:33.046255 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:33 crc kubenswrapper[4821]: I1125 10:33:33.046265 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:33 crc kubenswrapper[4821]: I1125 10:33:33.046281 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:33 crc kubenswrapper[4821]: I1125 10:33:33.046296 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:33Z","lastTransitionTime":"2025-11-25T10:33:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:33 crc kubenswrapper[4821]: I1125 10:33:33.075484 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:33:33 crc kubenswrapper[4821]: E1125 10:33:33.075682 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:37.075652584 +0000 UTC m=+147.611972441 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:33:33 crc kubenswrapper[4821]: I1125 10:33:33.075776 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:33:33 crc kubenswrapper[4821]: I1125 10:33:33.075814 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:33:33 crc kubenswrapper[4821]: E1125 10:33:33.075927 4821 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 10:33:33 crc kubenswrapper[4821]: E1125 10:33:33.075934 4821 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 10:33:33 crc kubenswrapper[4821]: E1125 10:33:33.075968 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 10:34:37.075959011 +0000 UTC m=+147.612278938 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 10:33:33 crc kubenswrapper[4821]: E1125 10:33:33.075985 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 10:34:37.075977002 +0000 UTC m=+147.612296949 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 10:33:33 crc kubenswrapper[4821]: I1125 10:33:33.113592 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:33:33 crc kubenswrapper[4821]: I1125 10:33:33.113637 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:33:33 crc kubenswrapper[4821]: I1125 10:33:33.113725 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:33:33 crc kubenswrapper[4821]: E1125 10:33:33.113716 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:33:33 crc kubenswrapper[4821]: E1125 10:33:33.113866 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:33:33 crc kubenswrapper[4821]: E1125 10:33:33.113926 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:33:33 crc kubenswrapper[4821]: I1125 10:33:33.148809 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:33 crc kubenswrapper[4821]: I1125 10:33:33.148845 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:33 crc kubenswrapper[4821]: I1125 10:33:33.148853 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:33 crc kubenswrapper[4821]: I1125 10:33:33.148867 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:33 crc kubenswrapper[4821]: I1125 10:33:33.148876 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:33Z","lastTransitionTime":"2025-11-25T10:33:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:33 crc kubenswrapper[4821]: I1125 10:33:33.176894 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:33:33 crc kubenswrapper[4821]: I1125 10:33:33.176956 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:33:33 crc kubenswrapper[4821]: E1125 10:33:33.177138 4821 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 10:33:33 crc kubenswrapper[4821]: E1125 10:33:33.177160 4821 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 10:33:33 crc kubenswrapper[4821]: E1125 10:33:33.177205 4821 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 10:33:33 crc kubenswrapper[4821]: E1125 10:33:33.177200 4821 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 10:33:33 crc kubenswrapper[4821]: E1125 10:33:33.177237 4821 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 10:33:33 crc kubenswrapper[4821]: E1125 10:33:33.177250 4821 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 10:33:33 crc kubenswrapper[4821]: E1125 10:33:33.177283 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 10:34:37.177261482 +0000 UTC m=+147.713581329 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 10:33:33 crc kubenswrapper[4821]: E1125 10:33:33.177311 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 10:34:37.177293322 +0000 UTC m=+147.713613169 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 10:33:33 crc kubenswrapper[4821]: I1125 10:33:33.250997 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:33 crc kubenswrapper[4821]: I1125 10:33:33.251044 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:33 crc kubenswrapper[4821]: I1125 10:33:33.251057 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:33 crc kubenswrapper[4821]: I1125 10:33:33.251074 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:33 crc kubenswrapper[4821]: I1125 10:33:33.251092 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:33Z","lastTransitionTime":"2025-11-25T10:33:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:33 crc kubenswrapper[4821]: I1125 10:33:33.353240 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:33 crc kubenswrapper[4821]: I1125 10:33:33.353284 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:33 crc kubenswrapper[4821]: I1125 10:33:33.353293 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:33 crc kubenswrapper[4821]: I1125 10:33:33.353308 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:33 crc kubenswrapper[4821]: I1125 10:33:33.353319 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:33Z","lastTransitionTime":"2025-11-25T10:33:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:33 crc kubenswrapper[4821]: I1125 10:33:33.454840 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:33 crc kubenswrapper[4821]: I1125 10:33:33.454877 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:33 crc kubenswrapper[4821]: I1125 10:33:33.454886 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:33 crc kubenswrapper[4821]: I1125 10:33:33.454900 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:33 crc kubenswrapper[4821]: I1125 10:33:33.454908 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:33Z","lastTransitionTime":"2025-11-25T10:33:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:33 crc kubenswrapper[4821]: I1125 10:33:33.557480 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:33 crc kubenswrapper[4821]: I1125 10:33:33.557543 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:33 crc kubenswrapper[4821]: I1125 10:33:33.557559 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:33 crc kubenswrapper[4821]: I1125 10:33:33.557580 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:33 crc kubenswrapper[4821]: I1125 10:33:33.557596 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:33Z","lastTransitionTime":"2025-11-25T10:33:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:33 crc kubenswrapper[4821]: I1125 10:33:33.660443 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:33 crc kubenswrapper[4821]: I1125 10:33:33.660494 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:33 crc kubenswrapper[4821]: I1125 10:33:33.660503 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:33 crc kubenswrapper[4821]: I1125 10:33:33.660522 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:33 crc kubenswrapper[4821]: I1125 10:33:33.660532 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:33Z","lastTransitionTime":"2025-11-25T10:33:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:33 crc kubenswrapper[4821]: I1125 10:33:33.763267 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:33 crc kubenswrapper[4821]: I1125 10:33:33.763305 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:33 crc kubenswrapper[4821]: I1125 10:33:33.763317 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:33 crc kubenswrapper[4821]: I1125 10:33:33.763333 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:33 crc kubenswrapper[4821]: I1125 10:33:33.763344 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:33Z","lastTransitionTime":"2025-11-25T10:33:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:33 crc kubenswrapper[4821]: I1125 10:33:33.865944 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:33 crc kubenswrapper[4821]: I1125 10:33:33.865994 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:33 crc kubenswrapper[4821]: I1125 10:33:33.866007 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:33 crc kubenswrapper[4821]: I1125 10:33:33.866025 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:33 crc kubenswrapper[4821]: I1125 10:33:33.866037 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:33Z","lastTransitionTime":"2025-11-25T10:33:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:33 crc kubenswrapper[4821]: I1125 10:33:33.969648 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:33 crc kubenswrapper[4821]: I1125 10:33:33.969697 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:33 crc kubenswrapper[4821]: I1125 10:33:33.969706 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:33 crc kubenswrapper[4821]: I1125 10:33:33.969722 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:33 crc kubenswrapper[4821]: I1125 10:33:33.969734 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:33Z","lastTransitionTime":"2025-11-25T10:33:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:34 crc kubenswrapper[4821]: I1125 10:33:34.072269 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:34 crc kubenswrapper[4821]: I1125 10:33:34.072337 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:34 crc kubenswrapper[4821]: I1125 10:33:34.072360 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:34 crc kubenswrapper[4821]: I1125 10:33:34.072388 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:34 crc kubenswrapper[4821]: I1125 10:33:34.072407 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:34Z","lastTransitionTime":"2025-11-25T10:33:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:34 crc kubenswrapper[4821]: I1125 10:33:34.114315 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4msn" Nov 25 10:33:34 crc kubenswrapper[4821]: E1125 10:33:34.114669 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4msn" podUID="1f12cef3-7e40-4f61-836e-23cd0e578e71" Nov 25 10:33:34 crc kubenswrapper[4821]: I1125 10:33:34.175301 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:34 crc kubenswrapper[4821]: I1125 10:33:34.175344 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:34 crc kubenswrapper[4821]: I1125 10:33:34.175352 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:34 crc kubenswrapper[4821]: I1125 10:33:34.175367 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:34 crc kubenswrapper[4821]: I1125 10:33:34.175378 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:34Z","lastTransitionTime":"2025-11-25T10:33:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:34 crc kubenswrapper[4821]: I1125 10:33:34.278038 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:34 crc kubenswrapper[4821]: I1125 10:33:34.278076 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:34 crc kubenswrapper[4821]: I1125 10:33:34.278086 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:34 crc kubenswrapper[4821]: I1125 10:33:34.278101 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:34 crc kubenswrapper[4821]: I1125 10:33:34.278112 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:34Z","lastTransitionTime":"2025-11-25T10:33:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:34 crc kubenswrapper[4821]: I1125 10:33:34.379553 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:34 crc kubenswrapper[4821]: I1125 10:33:34.379602 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:34 crc kubenswrapper[4821]: I1125 10:33:34.379617 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:34 crc kubenswrapper[4821]: I1125 10:33:34.379635 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:34 crc kubenswrapper[4821]: I1125 10:33:34.379648 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:34Z","lastTransitionTime":"2025-11-25T10:33:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:34 crc kubenswrapper[4821]: I1125 10:33:34.481647 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:34 crc kubenswrapper[4821]: I1125 10:33:34.481690 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:34 crc kubenswrapper[4821]: I1125 10:33:34.481699 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:34 crc kubenswrapper[4821]: I1125 10:33:34.481713 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:34 crc kubenswrapper[4821]: I1125 10:33:34.481723 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:34Z","lastTransitionTime":"2025-11-25T10:33:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:34 crc kubenswrapper[4821]: I1125 10:33:34.584057 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:34 crc kubenswrapper[4821]: I1125 10:33:34.584118 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:34 crc kubenswrapper[4821]: I1125 10:33:34.584127 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:34 crc kubenswrapper[4821]: I1125 10:33:34.584142 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:34 crc kubenswrapper[4821]: I1125 10:33:34.584154 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:34Z","lastTransitionTime":"2025-11-25T10:33:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:34 crc kubenswrapper[4821]: I1125 10:33:34.686986 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:34 crc kubenswrapper[4821]: I1125 10:33:34.687033 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:34 crc kubenswrapper[4821]: I1125 10:33:34.687044 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:34 crc kubenswrapper[4821]: I1125 10:33:34.687060 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:34 crc kubenswrapper[4821]: I1125 10:33:34.687072 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:34Z","lastTransitionTime":"2025-11-25T10:33:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:34 crc kubenswrapper[4821]: I1125 10:33:34.790025 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:34 crc kubenswrapper[4821]: I1125 10:33:34.790083 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:34 crc kubenswrapper[4821]: I1125 10:33:34.790099 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:34 crc kubenswrapper[4821]: I1125 10:33:34.790122 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:34 crc kubenswrapper[4821]: I1125 10:33:34.790138 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:34Z","lastTransitionTime":"2025-11-25T10:33:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:34 crc kubenswrapper[4821]: I1125 10:33:34.892542 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:34 crc kubenswrapper[4821]: I1125 10:33:34.892574 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:34 crc kubenswrapper[4821]: I1125 10:33:34.892583 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:34 crc kubenswrapper[4821]: I1125 10:33:34.892598 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:34 crc kubenswrapper[4821]: I1125 10:33:34.892610 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:34Z","lastTransitionTime":"2025-11-25T10:33:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:34 crc kubenswrapper[4821]: I1125 10:33:34.994409 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:34 crc kubenswrapper[4821]: I1125 10:33:34.994448 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:34 crc kubenswrapper[4821]: I1125 10:33:34.994457 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:34 crc kubenswrapper[4821]: I1125 10:33:34.994472 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:34 crc kubenswrapper[4821]: I1125 10:33:34.994481 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:34Z","lastTransitionTime":"2025-11-25T10:33:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:35 crc kubenswrapper[4821]: I1125 10:33:35.096757 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:35 crc kubenswrapper[4821]: I1125 10:33:35.096824 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:35 crc kubenswrapper[4821]: I1125 10:33:35.096843 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:35 crc kubenswrapper[4821]: I1125 10:33:35.096870 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:35 crc kubenswrapper[4821]: I1125 10:33:35.096890 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:35Z","lastTransitionTime":"2025-11-25T10:33:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:35 crc kubenswrapper[4821]: I1125 10:33:35.113258 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:33:35 crc kubenswrapper[4821]: I1125 10:33:35.113318 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:33:35 crc kubenswrapper[4821]: E1125 10:33:35.113388 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:33:35 crc kubenswrapper[4821]: I1125 10:33:35.113318 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:33:35 crc kubenswrapper[4821]: E1125 10:33:35.113491 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:33:35 crc kubenswrapper[4821]: E1125 10:33:35.113640 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:33:35 crc kubenswrapper[4821]: I1125 10:33:35.198428 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:35 crc kubenswrapper[4821]: I1125 10:33:35.198460 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:35 crc kubenswrapper[4821]: I1125 10:33:35.198471 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:35 crc kubenswrapper[4821]: I1125 10:33:35.198488 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:35 crc kubenswrapper[4821]: I1125 10:33:35.198499 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:35Z","lastTransitionTime":"2025-11-25T10:33:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:35 crc kubenswrapper[4821]: I1125 10:33:35.300408 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:35 crc kubenswrapper[4821]: I1125 10:33:35.300445 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:35 crc kubenswrapper[4821]: I1125 10:33:35.300454 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:35 crc kubenswrapper[4821]: I1125 10:33:35.300467 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:35 crc kubenswrapper[4821]: I1125 10:33:35.300477 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:35Z","lastTransitionTime":"2025-11-25T10:33:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:35 crc kubenswrapper[4821]: I1125 10:33:35.403591 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:35 crc kubenswrapper[4821]: I1125 10:33:35.403626 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:35 crc kubenswrapper[4821]: I1125 10:33:35.403636 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:35 crc kubenswrapper[4821]: I1125 10:33:35.403651 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:35 crc kubenswrapper[4821]: I1125 10:33:35.403660 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:35Z","lastTransitionTime":"2025-11-25T10:33:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:35 crc kubenswrapper[4821]: I1125 10:33:35.506110 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:35 crc kubenswrapper[4821]: I1125 10:33:35.506148 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:35 crc kubenswrapper[4821]: I1125 10:33:35.506156 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:35 crc kubenswrapper[4821]: I1125 10:33:35.506196 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:35 crc kubenswrapper[4821]: I1125 10:33:35.506208 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:35Z","lastTransitionTime":"2025-11-25T10:33:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:35 crc kubenswrapper[4821]: I1125 10:33:35.608972 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:35 crc kubenswrapper[4821]: I1125 10:33:35.609316 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:35 crc kubenswrapper[4821]: I1125 10:33:35.609331 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:35 crc kubenswrapper[4821]: I1125 10:33:35.609348 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:35 crc kubenswrapper[4821]: I1125 10:33:35.609359 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:35Z","lastTransitionTime":"2025-11-25T10:33:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:35 crc kubenswrapper[4821]: I1125 10:33:35.710981 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:35 crc kubenswrapper[4821]: I1125 10:33:35.711035 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:35 crc kubenswrapper[4821]: I1125 10:33:35.711048 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:35 crc kubenswrapper[4821]: I1125 10:33:35.711065 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:35 crc kubenswrapper[4821]: I1125 10:33:35.711081 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:35Z","lastTransitionTime":"2025-11-25T10:33:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:35 crc kubenswrapper[4821]: I1125 10:33:35.813213 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:35 crc kubenswrapper[4821]: I1125 10:33:35.813251 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:35 crc kubenswrapper[4821]: I1125 10:33:35.813260 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:35 crc kubenswrapper[4821]: I1125 10:33:35.813275 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:35 crc kubenswrapper[4821]: I1125 10:33:35.813285 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:35Z","lastTransitionTime":"2025-11-25T10:33:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:35 crc kubenswrapper[4821]: I1125 10:33:35.917178 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:35 crc kubenswrapper[4821]: I1125 10:33:35.917244 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:35 crc kubenswrapper[4821]: I1125 10:33:35.917258 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:35 crc kubenswrapper[4821]: I1125 10:33:35.917275 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:35 crc kubenswrapper[4821]: I1125 10:33:35.917286 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:35Z","lastTransitionTime":"2025-11-25T10:33:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:36 crc kubenswrapper[4821]: I1125 10:33:36.019798 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:36 crc kubenswrapper[4821]: I1125 10:33:36.019836 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:36 crc kubenswrapper[4821]: I1125 10:33:36.019849 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:36 crc kubenswrapper[4821]: I1125 10:33:36.019867 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:36 crc kubenswrapper[4821]: I1125 10:33:36.019879 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:36Z","lastTransitionTime":"2025-11-25T10:33:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:36 crc kubenswrapper[4821]: I1125 10:33:36.113307 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4msn" Nov 25 10:33:36 crc kubenswrapper[4821]: E1125 10:33:36.113566 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4msn" podUID="1f12cef3-7e40-4f61-836e-23cd0e578e71" Nov 25 10:33:36 crc kubenswrapper[4821]: I1125 10:33:36.121519 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:36 crc kubenswrapper[4821]: I1125 10:33:36.121562 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:36 crc kubenswrapper[4821]: I1125 10:33:36.121575 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:36 crc kubenswrapper[4821]: I1125 10:33:36.121590 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:36 crc kubenswrapper[4821]: I1125 10:33:36.121602 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:36Z","lastTransitionTime":"2025-11-25T10:33:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:36 crc kubenswrapper[4821]: I1125 10:33:36.224275 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:36 crc kubenswrapper[4821]: I1125 10:33:36.224564 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:36 crc kubenswrapper[4821]: I1125 10:33:36.224647 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:36 crc kubenswrapper[4821]: I1125 10:33:36.224726 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:36 crc kubenswrapper[4821]: I1125 10:33:36.224797 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:36Z","lastTransitionTime":"2025-11-25T10:33:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:36 crc kubenswrapper[4821]: I1125 10:33:36.327103 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:36 crc kubenswrapper[4821]: I1125 10:33:36.327149 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:36 crc kubenswrapper[4821]: I1125 10:33:36.327181 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:36 crc kubenswrapper[4821]: I1125 10:33:36.327201 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:36 crc kubenswrapper[4821]: I1125 10:33:36.327213 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:36Z","lastTransitionTime":"2025-11-25T10:33:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:36 crc kubenswrapper[4821]: I1125 10:33:36.429459 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:36 crc kubenswrapper[4821]: I1125 10:33:36.429528 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:36 crc kubenswrapper[4821]: I1125 10:33:36.429548 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:36 crc kubenswrapper[4821]: I1125 10:33:36.429573 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:36 crc kubenswrapper[4821]: I1125 10:33:36.429587 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:36Z","lastTransitionTime":"2025-11-25T10:33:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:36 crc kubenswrapper[4821]: I1125 10:33:36.531692 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:36 crc kubenswrapper[4821]: I1125 10:33:36.531728 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:36 crc kubenswrapper[4821]: I1125 10:33:36.531738 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:36 crc kubenswrapper[4821]: I1125 10:33:36.531755 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:36 crc kubenswrapper[4821]: I1125 10:33:36.531765 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:36Z","lastTransitionTime":"2025-11-25T10:33:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:36 crc kubenswrapper[4821]: I1125 10:33:36.634471 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:36 crc kubenswrapper[4821]: I1125 10:33:36.634498 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:36 crc kubenswrapper[4821]: I1125 10:33:36.634505 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:36 crc kubenswrapper[4821]: I1125 10:33:36.634518 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:36 crc kubenswrapper[4821]: I1125 10:33:36.634527 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:36Z","lastTransitionTime":"2025-11-25T10:33:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:36 crc kubenswrapper[4821]: I1125 10:33:36.737271 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:36 crc kubenswrapper[4821]: I1125 10:33:36.737335 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:36 crc kubenswrapper[4821]: I1125 10:33:36.737347 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:36 crc kubenswrapper[4821]: I1125 10:33:36.737370 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:36 crc kubenswrapper[4821]: I1125 10:33:36.737382 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:36Z","lastTransitionTime":"2025-11-25T10:33:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:36 crc kubenswrapper[4821]: I1125 10:33:36.839784 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:36 crc kubenswrapper[4821]: I1125 10:33:36.839861 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:36 crc kubenswrapper[4821]: I1125 10:33:36.839878 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:36 crc kubenswrapper[4821]: I1125 10:33:36.839927 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:36 crc kubenswrapper[4821]: I1125 10:33:36.839946 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:36Z","lastTransitionTime":"2025-11-25T10:33:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:36 crc kubenswrapper[4821]: I1125 10:33:36.942044 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:36 crc kubenswrapper[4821]: I1125 10:33:36.942087 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:36 crc kubenswrapper[4821]: I1125 10:33:36.942100 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:36 crc kubenswrapper[4821]: I1125 10:33:36.942117 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:36 crc kubenswrapper[4821]: I1125 10:33:36.942127 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:36Z","lastTransitionTime":"2025-11-25T10:33:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:37 crc kubenswrapper[4821]: I1125 10:33:37.045079 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:37 crc kubenswrapper[4821]: I1125 10:33:37.045160 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:37 crc kubenswrapper[4821]: I1125 10:33:37.045212 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:37 crc kubenswrapper[4821]: I1125 10:33:37.045242 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:37 crc kubenswrapper[4821]: I1125 10:33:37.045256 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:37Z","lastTransitionTime":"2025-11-25T10:33:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:37 crc kubenswrapper[4821]: I1125 10:33:37.114074 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:33:37 crc kubenswrapper[4821]: I1125 10:33:37.114099 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:33:37 crc kubenswrapper[4821]: E1125 10:33:37.114377 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:33:37 crc kubenswrapper[4821]: I1125 10:33:37.114153 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:33:37 crc kubenswrapper[4821]: E1125 10:33:37.114519 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:33:37 crc kubenswrapper[4821]: E1125 10:33:37.114671 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:33:37 crc kubenswrapper[4821]: I1125 10:33:37.146847 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:37 crc kubenswrapper[4821]: I1125 10:33:37.146897 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:37 crc kubenswrapper[4821]: I1125 10:33:37.146910 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:37 crc kubenswrapper[4821]: I1125 10:33:37.146926 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:37 crc kubenswrapper[4821]: I1125 10:33:37.146936 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:37Z","lastTransitionTime":"2025-11-25T10:33:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:37 crc kubenswrapper[4821]: I1125 10:33:37.249725 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:37 crc kubenswrapper[4821]: I1125 10:33:37.249770 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:37 crc kubenswrapper[4821]: I1125 10:33:37.249781 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:37 crc kubenswrapper[4821]: I1125 10:33:37.249797 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:37 crc kubenswrapper[4821]: I1125 10:33:37.249809 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:37Z","lastTransitionTime":"2025-11-25T10:33:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:37 crc kubenswrapper[4821]: I1125 10:33:37.352930 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:37 crc kubenswrapper[4821]: I1125 10:33:37.352980 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:37 crc kubenswrapper[4821]: I1125 10:33:37.352990 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:37 crc kubenswrapper[4821]: I1125 10:33:37.353008 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:37 crc kubenswrapper[4821]: I1125 10:33:37.353019 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:37Z","lastTransitionTime":"2025-11-25T10:33:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:37 crc kubenswrapper[4821]: I1125 10:33:37.414280 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:37 crc kubenswrapper[4821]: I1125 10:33:37.414329 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:37 crc kubenswrapper[4821]: I1125 10:33:37.414340 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:37 crc kubenswrapper[4821]: I1125 10:33:37.414358 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:37 crc kubenswrapper[4821]: I1125 10:33:37.414369 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:37Z","lastTransitionTime":"2025-11-25T10:33:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:37 crc kubenswrapper[4821]: E1125 10:33:37.425874 4821 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a5b2b070-39b9-412c-90f1-c7ed2e9e0950\\\",\\\"systemUUID\\\":\\\"ea86e80f-41b7-4190-ad7e-d8c549d535de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:37Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:37 crc kubenswrapper[4821]: I1125 10:33:37.429511 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:37 crc kubenswrapper[4821]: I1125 10:33:37.429579 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:37 crc kubenswrapper[4821]: I1125 10:33:37.429594 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:37 crc kubenswrapper[4821]: I1125 10:33:37.429619 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:37 crc kubenswrapper[4821]: I1125 10:33:37.429631 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:37Z","lastTransitionTime":"2025-11-25T10:33:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:37 crc kubenswrapper[4821]: E1125 10:33:37.443795 4821 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a5b2b070-39b9-412c-90f1-c7ed2e9e0950\\\",\\\"systemUUID\\\":\\\"ea86e80f-41b7-4190-ad7e-d8c549d535de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:37Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:37 crc kubenswrapper[4821]: I1125 10:33:37.447351 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:37 crc kubenswrapper[4821]: I1125 10:33:37.447408 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:37 crc kubenswrapper[4821]: I1125 10:33:37.447419 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:37 crc kubenswrapper[4821]: I1125 10:33:37.447439 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:37 crc kubenswrapper[4821]: I1125 10:33:37.447451 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:37Z","lastTransitionTime":"2025-11-25T10:33:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:37 crc kubenswrapper[4821]: E1125 10:33:37.460406 4821 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a5b2b070-39b9-412c-90f1-c7ed2e9e0950\\\",\\\"systemUUID\\\":\\\"ea86e80f-41b7-4190-ad7e-d8c549d535de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:37Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:37 crc kubenswrapper[4821]: I1125 10:33:37.464908 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:37 crc kubenswrapper[4821]: I1125 10:33:37.464954 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:37 crc kubenswrapper[4821]: I1125 10:33:37.464964 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:37 crc kubenswrapper[4821]: I1125 10:33:37.464979 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:37 crc kubenswrapper[4821]: I1125 10:33:37.464988 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:37Z","lastTransitionTime":"2025-11-25T10:33:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:37 crc kubenswrapper[4821]: E1125 10:33:37.479501 4821 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a5b2b070-39b9-412c-90f1-c7ed2e9e0950\\\",\\\"systemUUID\\\":\\\"ea86e80f-41b7-4190-ad7e-d8c549d535de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:37Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:37 crc kubenswrapper[4821]: I1125 10:33:37.484761 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:37 crc kubenswrapper[4821]: I1125 10:33:37.484814 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:37 crc kubenswrapper[4821]: I1125 10:33:37.484827 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:37 crc kubenswrapper[4821]: I1125 10:33:37.484875 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:37 crc kubenswrapper[4821]: I1125 10:33:37.484890 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:37Z","lastTransitionTime":"2025-11-25T10:33:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:37 crc kubenswrapper[4821]: E1125 10:33:37.497779 4821 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a5b2b070-39b9-412c-90f1-c7ed2e9e0950\\\",\\\"systemUUID\\\":\\\"ea86e80f-41b7-4190-ad7e-d8c549d535de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:37Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:37 crc kubenswrapper[4821]: E1125 10:33:37.497896 4821 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 10:33:37 crc kubenswrapper[4821]: I1125 10:33:37.499668 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:37 crc kubenswrapper[4821]: I1125 10:33:37.499727 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:37 crc kubenswrapper[4821]: I1125 10:33:37.499737 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:37 crc kubenswrapper[4821]: I1125 10:33:37.499752 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:37 crc kubenswrapper[4821]: I1125 10:33:37.499761 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:37Z","lastTransitionTime":"2025-11-25T10:33:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:37 crc kubenswrapper[4821]: I1125 10:33:37.602079 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:37 crc kubenswrapper[4821]: I1125 10:33:37.602119 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:37 crc kubenswrapper[4821]: I1125 10:33:37.602130 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:37 crc kubenswrapper[4821]: I1125 10:33:37.602145 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:37 crc kubenswrapper[4821]: I1125 10:33:37.602156 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:37Z","lastTransitionTime":"2025-11-25T10:33:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:37 crc kubenswrapper[4821]: I1125 10:33:37.704394 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:37 crc kubenswrapper[4821]: I1125 10:33:37.704437 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:37 crc kubenswrapper[4821]: I1125 10:33:37.704453 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:37 crc kubenswrapper[4821]: I1125 10:33:37.704469 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:37 crc kubenswrapper[4821]: I1125 10:33:37.704481 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:37Z","lastTransitionTime":"2025-11-25T10:33:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:37 crc kubenswrapper[4821]: I1125 10:33:37.806934 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:37 crc kubenswrapper[4821]: I1125 10:33:37.806996 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:37 crc kubenswrapper[4821]: I1125 10:33:37.807009 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:37 crc kubenswrapper[4821]: I1125 10:33:37.807025 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:37 crc kubenswrapper[4821]: I1125 10:33:37.807037 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:37Z","lastTransitionTime":"2025-11-25T10:33:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:37 crc kubenswrapper[4821]: I1125 10:33:37.908708 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:37 crc kubenswrapper[4821]: I1125 10:33:37.908752 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:37 crc kubenswrapper[4821]: I1125 10:33:37.908763 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:37 crc kubenswrapper[4821]: I1125 10:33:37.908780 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:37 crc kubenswrapper[4821]: I1125 10:33:37.908792 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:37Z","lastTransitionTime":"2025-11-25T10:33:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:38 crc kubenswrapper[4821]: I1125 10:33:38.010743 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:38 crc kubenswrapper[4821]: I1125 10:33:38.010787 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:38 crc kubenswrapper[4821]: I1125 10:33:38.010816 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:38 crc kubenswrapper[4821]: I1125 10:33:38.010831 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:38 crc kubenswrapper[4821]: I1125 10:33:38.010840 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:38Z","lastTransitionTime":"2025-11-25T10:33:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:38 crc kubenswrapper[4821]: I1125 10:33:38.112836 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:38 crc kubenswrapper[4821]: I1125 10:33:38.112882 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:38 crc kubenswrapper[4821]: I1125 10:33:38.112896 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:38 crc kubenswrapper[4821]: I1125 10:33:38.112913 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:38 crc kubenswrapper[4821]: I1125 10:33:38.112925 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:38Z","lastTransitionTime":"2025-11-25T10:33:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:38 crc kubenswrapper[4821]: I1125 10:33:38.113530 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4msn" Nov 25 10:33:38 crc kubenswrapper[4821]: E1125 10:33:38.113650 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4msn" podUID="1f12cef3-7e40-4f61-836e-23cd0e578e71" Nov 25 10:33:38 crc kubenswrapper[4821]: I1125 10:33:38.215461 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:38 crc kubenswrapper[4821]: I1125 10:33:38.215498 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:38 crc kubenswrapper[4821]: I1125 10:33:38.215507 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:38 crc kubenswrapper[4821]: I1125 10:33:38.215530 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:38 crc kubenswrapper[4821]: I1125 10:33:38.215544 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:38Z","lastTransitionTime":"2025-11-25T10:33:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:38 crc kubenswrapper[4821]: I1125 10:33:38.317406 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:38 crc kubenswrapper[4821]: I1125 10:33:38.317446 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:38 crc kubenswrapper[4821]: I1125 10:33:38.317457 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:38 crc kubenswrapper[4821]: I1125 10:33:38.317472 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:38 crc kubenswrapper[4821]: I1125 10:33:38.317482 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:38Z","lastTransitionTime":"2025-11-25T10:33:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:38 crc kubenswrapper[4821]: I1125 10:33:38.419667 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:38 crc kubenswrapper[4821]: I1125 10:33:38.419698 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:38 crc kubenswrapper[4821]: I1125 10:33:38.419707 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:38 crc kubenswrapper[4821]: I1125 10:33:38.419719 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:38 crc kubenswrapper[4821]: I1125 10:33:38.419728 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:38Z","lastTransitionTime":"2025-11-25T10:33:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:38 crc kubenswrapper[4821]: I1125 10:33:38.522442 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:38 crc kubenswrapper[4821]: I1125 10:33:38.522526 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:38 crc kubenswrapper[4821]: I1125 10:33:38.522544 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:38 crc kubenswrapper[4821]: I1125 10:33:38.522569 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:38 crc kubenswrapper[4821]: I1125 10:33:38.522587 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:38Z","lastTransitionTime":"2025-11-25T10:33:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:38 crc kubenswrapper[4821]: I1125 10:33:38.625597 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:38 crc kubenswrapper[4821]: I1125 10:33:38.625657 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:38 crc kubenswrapper[4821]: I1125 10:33:38.625668 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:38 crc kubenswrapper[4821]: I1125 10:33:38.625684 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:38 crc kubenswrapper[4821]: I1125 10:33:38.625694 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:38Z","lastTransitionTime":"2025-11-25T10:33:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:38 crc kubenswrapper[4821]: I1125 10:33:38.728338 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:38 crc kubenswrapper[4821]: I1125 10:33:38.728382 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:38 crc kubenswrapper[4821]: I1125 10:33:38.728394 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:38 crc kubenswrapper[4821]: I1125 10:33:38.728410 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:38 crc kubenswrapper[4821]: I1125 10:33:38.728421 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:38Z","lastTransitionTime":"2025-11-25T10:33:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:38 crc kubenswrapper[4821]: I1125 10:33:38.831033 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:38 crc kubenswrapper[4821]: I1125 10:33:38.831072 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:38 crc kubenswrapper[4821]: I1125 10:33:38.831088 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:38 crc kubenswrapper[4821]: I1125 10:33:38.831103 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:38 crc kubenswrapper[4821]: I1125 10:33:38.831114 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:38Z","lastTransitionTime":"2025-11-25T10:33:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:38 crc kubenswrapper[4821]: I1125 10:33:38.933368 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:38 crc kubenswrapper[4821]: I1125 10:33:38.933405 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:38 crc kubenswrapper[4821]: I1125 10:33:38.933438 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:38 crc kubenswrapper[4821]: I1125 10:33:38.933452 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:38 crc kubenswrapper[4821]: I1125 10:33:38.933462 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:38Z","lastTransitionTime":"2025-11-25T10:33:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:39 crc kubenswrapper[4821]: I1125 10:33:39.035593 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:39 crc kubenswrapper[4821]: I1125 10:33:39.035641 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:39 crc kubenswrapper[4821]: I1125 10:33:39.035655 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:39 crc kubenswrapper[4821]: I1125 10:33:39.035671 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:39 crc kubenswrapper[4821]: I1125 10:33:39.035685 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:39Z","lastTransitionTime":"2025-11-25T10:33:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:39 crc kubenswrapper[4821]: I1125 10:33:39.113622 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:33:39 crc kubenswrapper[4821]: I1125 10:33:39.113622 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:33:39 crc kubenswrapper[4821]: I1125 10:33:39.113755 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:33:39 crc kubenswrapper[4821]: E1125 10:33:39.113902 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:33:39 crc kubenswrapper[4821]: E1125 10:33:39.114141 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:33:39 crc kubenswrapper[4821]: E1125 10:33:39.114432 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:33:39 crc kubenswrapper[4821]: I1125 10:33:39.137298 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:39 crc kubenswrapper[4821]: I1125 10:33:39.137334 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:39 crc kubenswrapper[4821]: I1125 10:33:39.137342 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:39 crc kubenswrapper[4821]: I1125 10:33:39.137355 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:39 crc kubenswrapper[4821]: I1125 10:33:39.137363 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:39Z","lastTransitionTime":"2025-11-25T10:33:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:39 crc kubenswrapper[4821]: I1125 10:33:39.239935 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:39 crc kubenswrapper[4821]: I1125 10:33:39.239979 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:39 crc kubenswrapper[4821]: I1125 10:33:39.239990 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:39 crc kubenswrapper[4821]: I1125 10:33:39.240005 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:39 crc kubenswrapper[4821]: I1125 10:33:39.240017 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:39Z","lastTransitionTime":"2025-11-25T10:33:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:39 crc kubenswrapper[4821]: I1125 10:33:39.342048 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:39 crc kubenswrapper[4821]: I1125 10:33:39.342095 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:39 crc kubenswrapper[4821]: I1125 10:33:39.342106 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:39 crc kubenswrapper[4821]: I1125 10:33:39.342121 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:39 crc kubenswrapper[4821]: I1125 10:33:39.342131 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:39Z","lastTransitionTime":"2025-11-25T10:33:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:39 crc kubenswrapper[4821]: I1125 10:33:39.445061 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:39 crc kubenswrapper[4821]: I1125 10:33:39.445096 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:39 crc kubenswrapper[4821]: I1125 10:33:39.445106 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:39 crc kubenswrapper[4821]: I1125 10:33:39.445132 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:39 crc kubenswrapper[4821]: I1125 10:33:39.445143 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:39Z","lastTransitionTime":"2025-11-25T10:33:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:39 crc kubenswrapper[4821]: I1125 10:33:39.546987 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:39 crc kubenswrapper[4821]: I1125 10:33:39.547032 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:39 crc kubenswrapper[4821]: I1125 10:33:39.547044 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:39 crc kubenswrapper[4821]: I1125 10:33:39.547077 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:39 crc kubenswrapper[4821]: I1125 10:33:39.547090 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:39Z","lastTransitionTime":"2025-11-25T10:33:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:39 crc kubenswrapper[4821]: I1125 10:33:39.649936 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:39 crc kubenswrapper[4821]: I1125 10:33:39.649972 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:39 crc kubenswrapper[4821]: I1125 10:33:39.649981 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:39 crc kubenswrapper[4821]: I1125 10:33:39.649994 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:39 crc kubenswrapper[4821]: I1125 10:33:39.650004 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:39Z","lastTransitionTime":"2025-11-25T10:33:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:39 crc kubenswrapper[4821]: I1125 10:33:39.752194 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:39 crc kubenswrapper[4821]: I1125 10:33:39.752226 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:39 crc kubenswrapper[4821]: I1125 10:33:39.752236 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:39 crc kubenswrapper[4821]: I1125 10:33:39.752254 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:39 crc kubenswrapper[4821]: I1125 10:33:39.752271 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:39Z","lastTransitionTime":"2025-11-25T10:33:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:39 crc kubenswrapper[4821]: I1125 10:33:39.854624 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:39 crc kubenswrapper[4821]: I1125 10:33:39.854660 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:39 crc kubenswrapper[4821]: I1125 10:33:39.854669 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:39 crc kubenswrapper[4821]: I1125 10:33:39.854682 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:39 crc kubenswrapper[4821]: I1125 10:33:39.854690 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:39Z","lastTransitionTime":"2025-11-25T10:33:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:39 crc kubenswrapper[4821]: I1125 10:33:39.956634 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:39 crc kubenswrapper[4821]: I1125 10:33:39.956681 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:39 crc kubenswrapper[4821]: I1125 10:33:39.956693 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:39 crc kubenswrapper[4821]: I1125 10:33:39.956709 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:39 crc kubenswrapper[4821]: I1125 10:33:39.956721 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:39Z","lastTransitionTime":"2025-11-25T10:33:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:40 crc kubenswrapper[4821]: I1125 10:33:40.060422 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:40 crc kubenswrapper[4821]: I1125 10:33:40.060486 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:40 crc kubenswrapper[4821]: I1125 10:33:40.060498 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:40 crc kubenswrapper[4821]: I1125 10:33:40.060522 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:40 crc kubenswrapper[4821]: I1125 10:33:40.060541 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:40Z","lastTransitionTime":"2025-11-25T10:33:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:40 crc kubenswrapper[4821]: I1125 10:33:40.114362 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4msn" Nov 25 10:33:40 crc kubenswrapper[4821]: E1125 10:33:40.114732 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4msn" podUID="1f12cef3-7e40-4f61-836e-23cd0e578e71" Nov 25 10:33:40 crc kubenswrapper[4821]: I1125 10:33:40.131343 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62fcb632-f12a-46e2-bb7d-a5311f291ee7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28d43531927e75c3745f100940f7f6bf79fd21026c1309194df8cd7a39361b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69633332d03d2551ff4abb253d358de398eab18486b92db630ce06c0447fb9cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ccb78d9a51ac83ee208c2451685a2512de9deaf742a663bb8c892aa59b76162\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27826e155a9352b1e1866e83fdab683dbcd749cc04b6a4da2db25ecabe3d9a52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://991252e5f56f128827a5054b8e228284bc0266166582cbaedaadc53ca2f5b401\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:40Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:40 crc kubenswrapper[4821]: I1125 10:33:40.144399 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63b471fd-5e16-4761-8b07-44a135d5f5f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a68a57fa7fba8f351f264e427a50bed23f57775644e9c268aa7305efb6916794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67df7c609d5a5d658be4c3e873ec3052bd070ad9469f8ff44c29a9571e3e5d95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://269ced131bc0777dc6752070470c95b4bedfe4b4240940eabd769500ddef0404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e48c0e3b58359c6c0a493d58f7d2aec6011747791a3c0d7788102e1fec18865\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e48c0e3b58359c6c0a493d58f7d2aec6011747791a3c0d7788102e1fec18865\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:40Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:40 crc kubenswrapper[4821]: I1125 10:33:40.158453 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:40Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:40 crc kubenswrapper[4821]: I1125 10:33:40.163138 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:40 crc kubenswrapper[4821]: I1125 10:33:40.163210 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:40 crc kubenswrapper[4821]: I1125 10:33:40.163220 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:40 crc kubenswrapper[4821]: I1125 10:33:40.163235 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:40 crc kubenswrapper[4821]: I1125 10:33:40.163245 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:40Z","lastTransitionTime":"2025-11-25T10:33:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:40 crc kubenswrapper[4821]: I1125 10:33:40.174622 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f948b87-ac86-4de6-ad64-c2ef947f84d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4482f73a45bc82bd55770c3fd788becacec94893fed1c305fed1c81c753d43aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ba68bd093be94d0d40f3851638db5d26bdfc52d175b1ac78afb91dda7affcfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2krbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:40Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:40 crc kubenswrapper[4821]: I1125 10:33:40.189822 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qppnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"357491b8-cd9c-486e-9898-5ba33678513d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc88d700ca3a449ef1cb32d6124b16d0024d6981ced12e276083f86a1b322ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd9c047c5ce1d896f803ff6cafa9bbf3a18a391b44539e1f1312e08c89e51cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd9c047c5ce1d896f803ff6cafa9bbf3a18a391b44539e1f1312e08c89e51cb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://565a6db63fb69abbba81ffdf6085797900858d263cd6affcb0b5f195474c2516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://565a6db63fb69abbba81ffdf6085797900858d263cd6affcb0b5f195474c2516\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://544885302dffff8336a9aa3c4afa80d69088ffaa2402648f4a329e0a798a8c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://544885302dffff8336a9aa3c4afa80d69088ffaa2402648f4a329e0a798a8c99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8bc193fb3f85b980c31af2e826cee90adb633ad2ee82ff1d46e182ab52e8fab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8bc193fb3f85b980c31af2e826cee90adb633ad2ee82ff1d46e182ab52e8fab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7713e2a35bac8df1e8987fbcd452459fa3541caad0c52ed57fda187bd7c83d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7713e2a35bac8df1e8987fbcd452459fa3541caad0c52ed57fda187bd7c83d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34adfca89026fc20007a73a8cbd4e7c7b3137a581c3476867af17fabe976f71f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34adfca89026fc20007a73a8cbd4e7c7b3137a581c3476867af17fabe976f71f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qppnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:40Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:40 crc kubenswrapper[4821]: I1125 10:33:40.201382 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tgdlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2336955b-a703-4db4-8d0f-d08088ed59b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb0362942bfac635ee6292501b1b851772c01d75dd658eef4b17669a716de72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vjfp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tgdlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:40Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:40 crc kubenswrapper[4821]: I1125 10:33:40.215730 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06f949936bac99cfbf89d720c9549c074351d9077bd963fa024d87004c6be23e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:40Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:40 crc kubenswrapper[4821]: I1125 10:33:40.228026 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384d8ba58264ab89796d3b53afb0c4c1be33cb4fba719a8beef4afc2be7d66a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:40Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:40 crc kubenswrapper[4821]: I1125 10:33:40.240155 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:40Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:40 crc kubenswrapper[4821]: I1125 10:33:40.252762 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17b8714be3bd121e06641873e003f08560b1f71e56d6ea936966df7d4896eb03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85731e6b6dbe1777bc7161a010f39f79523ec4b08d71d5d8a975573aa7fe4957\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:40Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:40 crc kubenswrapper[4821]: I1125 10:33:40.262581 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zlk96" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d55aa91-9cb9-4c95-bc27-6fec05684c10\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4521b5baee45ffc9c7661c781c0b3de458dff325ad525dde283d46af71c0cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vkd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zlk96\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:40Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:40 crc kubenswrapper[4821]: I1125 10:33:40.265822 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:40 crc kubenswrapper[4821]: I1125 10:33:40.265885 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:40 crc kubenswrapper[4821]: I1125 10:33:40.265900 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:40 crc kubenswrapper[4821]: I1125 10:33:40.265923 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:40 crc kubenswrapper[4821]: I1125 10:33:40.265968 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:40Z","lastTransitionTime":"2025-11-25T10:33:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:40 crc kubenswrapper[4821]: I1125 10:33:40.274099 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6hbqd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f71b3a2-63cb-443c-9f41-39e57030746c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65473b0e63c2e9c7c09a2bb8f31396daf1d87c2c83d20ed3508ce2fe7358dda3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxdmq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://595f9ea75461f18e763d84799e368becc07b72acdc5337d20d6995bf53c3d27a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zxdmq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6hbqd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:40Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:40 crc kubenswrapper[4821]: I1125 10:33:40.286213 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3e50766-ed07-4d22-95d0-3e0a9d822a17\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9525225d46e6f9225786ef8dd0ebe108d41efa26f6c94451c069aa5eb30da58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://733440f5c7493b5e1f4597ebb8f9d3d79507df0a1bc33393fb4534b5a2f2f077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://733440f5c7493b5e1f4597ebb8f9d3d79507df0a1bc33393fb4534b5a2f2f077\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:40Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:40 crc kubenswrapper[4821]: I1125 10:33:40.307992 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d9e58df-1755-4422-8d04-211f98c7c011\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea762f627aa69fb88a9bd82545fa677e9ccfa97f99a5e92c0aa490e619c0f9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a06b811e5ec52ce984a2403ba2c98332ba60466d08b3580ad9192fac6caef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0385f6ed9597974267f7c45547f6b6725b4ee83a2cc0508ff2d1c1ce8753ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b915b565f7fbb3e4e856e7f1035a519e82bfb3023660c121f6e3813de353bb92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18465f3a6916c36457a305ba5d568c9077f34f1d000532ed518f237d100008b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db2616cb96c8a040547477572b58b1f6349d02b376451088ecc95350ad27ed12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3e8e9cb61578376845f17a35c4bd4b88affa2be0ca49a1e83cafee7ff839751\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3fe1f522b7e5b892a5f4a37b8a3c4d00d929bfa2d5957ef26c68686ba99ab29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:40Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:40 crc kubenswrapper[4821]: I1125 10:33:40.321497 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25595ccd-762d-414a-9916-4fe38d478b07\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d4b2d8db4d13392d9789e161fd4b8f6eef983b79128ee584d72e4c06565bc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf780f791771ea43a8c8176d02fdfb5204406e023207299369ae46ceb342baaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0251519fd4262f2ea5c64959db49433afdafe36b7ed2404d660eeb7c365d2746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad5c205e0eeecfd43cbdf0eb1a334e3e8faa5e9a9e8dd202c397d5cbb36e081e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:40Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:40 crc kubenswrapper[4821]: I1125 10:33:40.334273 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:40Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:40 crc kubenswrapper[4821]: I1125 10:33:40.347885 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ldzbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b84e32ae-4897-4644-a656-6bd82012cc81\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://418dc8577fdc98324677dbccfcf9317b67cab4408432b30df2818b22f88d426e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ef433744761f26e8e1f6c0db123821a11923ffd774830744285cd4178d89533\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T10:33:19Z\\\",\\\"message\\\":\\\"2025-11-25T10:32:34+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_60a08f2b-f038-4c96-9f8b-1bd34a1965eb\\\\n2025-11-25T10:32:34+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_60a08f2b-f038-4c96-9f8b-1bd34a1965eb to /host/opt/cni/bin/\\\\n2025-11-25T10:32:34Z [verbose] multus-daemon started\\\\n2025-11-25T10:32:34Z [verbose] Readiness Indicator file check\\\\n2025-11-25T10:33:19Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:33:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g5pmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ldzbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:40Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:40 crc kubenswrapper[4821]: I1125 10:33:40.368646 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:40 crc kubenswrapper[4821]: I1125 10:33:40.368710 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:40 crc kubenswrapper[4821]: I1125 10:33:40.368730 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:40 crc kubenswrapper[4821]: I1125 10:33:40.368754 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:40 crc kubenswrapper[4821]: I1125 10:33:40.368766 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:40Z","lastTransitionTime":"2025-11-25T10:33:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:40 crc kubenswrapper[4821]: I1125 10:33:40.371412 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"954085fa-d03a-4e88-9dfd-566257e3558d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f00746930c4588e548f0e2b60d52bf448aff7401b676a7855cb0f4cce00d329\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a14a078ba7d2aed95992b36502008f315f0d3a634b64ed49da70398b2562d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9d7254196b3702e48083ae98e829b73c539adf69ed5b2bbdc62189d09ae1762\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d727430108977574365f0e9c04b44e03307bd7ff89cdbb1e48a6e916f504d6b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94c7949f077b1ca2b310bedb090aa72fc4bd7e8f2c8bf438096edfca8568f36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed2e7cad7bb24b5b56a1295098a6a7c230c8f9b9c29a88ef653ae8b6ac31667d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90d8792a79993c04a0eb774498395d955deff5c42c693892426685d157cf9cff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90d8792a79993c04a0eb774498395d955deff5c42c693892426685d157cf9cff\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T10:33:26Z\\\",\\\"message\\\":\\\" success event on pod openshift-ovn-kubernetes/ovnkube-node-mzhnw\\\\nI1125 10:33:26.842402 6839 ovn.go:134] Ensuring zone local for Pod openshift-network-node-identity/network-node-identity-vrzqb in node crc\\\\nI1125 10:33:26.842411 6839 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-apiserver/kube-apiserver-crc\\\\nF1125 10:33:26.842413 6839 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:26Z is after 2025-08-24T17:21:41Z]\\\\nI1125 10:33:26.842420 6839 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI1125 10:33:26.842421 6839 loadbalancer.go:304] Deleted 0 stale LB\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T10:33:26Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-mzhnw_openshift-ovn-kubernetes(954085fa-d03a-4e88-9dfd-566257e3558d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dbf499f4ae6161752abaa7a466a027c6f874016be90a534e69f07e3dd3f3d9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrb87\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-mzhnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:40Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:40 crc kubenswrapper[4821]: I1125 10:33:40.383340 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-d4msn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f12cef3-7e40-4f61-836e-23cd0e578e71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s7lkv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s7lkv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:44Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-d4msn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:40Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:40 crc kubenswrapper[4821]: I1125 10:33:40.472616 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:40 crc kubenswrapper[4821]: I1125 10:33:40.472676 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:40 crc kubenswrapper[4821]: I1125 10:33:40.472689 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:40 crc kubenswrapper[4821]: I1125 10:33:40.472711 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:40 crc kubenswrapper[4821]: I1125 10:33:40.472726 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:40Z","lastTransitionTime":"2025-11-25T10:33:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:40 crc kubenswrapper[4821]: I1125 10:33:40.575768 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:40 crc kubenswrapper[4821]: I1125 10:33:40.575820 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:40 crc kubenswrapper[4821]: I1125 10:33:40.575836 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:40 crc kubenswrapper[4821]: I1125 10:33:40.575856 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:40 crc kubenswrapper[4821]: I1125 10:33:40.575868 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:40Z","lastTransitionTime":"2025-11-25T10:33:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:40 crc kubenswrapper[4821]: I1125 10:33:40.679060 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:40 crc kubenswrapper[4821]: I1125 10:33:40.679101 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:40 crc kubenswrapper[4821]: I1125 10:33:40.679110 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:40 crc kubenswrapper[4821]: I1125 10:33:40.679126 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:40 crc kubenswrapper[4821]: I1125 10:33:40.679141 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:40Z","lastTransitionTime":"2025-11-25T10:33:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:40 crc kubenswrapper[4821]: I1125 10:33:40.781650 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:40 crc kubenswrapper[4821]: I1125 10:33:40.781678 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:40 crc kubenswrapper[4821]: I1125 10:33:40.781686 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:40 crc kubenswrapper[4821]: I1125 10:33:40.781700 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:40 crc kubenswrapper[4821]: I1125 10:33:40.781709 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:40Z","lastTransitionTime":"2025-11-25T10:33:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:40 crc kubenswrapper[4821]: I1125 10:33:40.883686 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:40 crc kubenswrapper[4821]: I1125 10:33:40.883720 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:40 crc kubenswrapper[4821]: I1125 10:33:40.883731 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:40 crc kubenswrapper[4821]: I1125 10:33:40.883747 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:40 crc kubenswrapper[4821]: I1125 10:33:40.883757 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:40Z","lastTransitionTime":"2025-11-25T10:33:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:40 crc kubenswrapper[4821]: I1125 10:33:40.985479 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:40 crc kubenswrapper[4821]: I1125 10:33:40.985508 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:40 crc kubenswrapper[4821]: I1125 10:33:40.985518 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:40 crc kubenswrapper[4821]: I1125 10:33:40.985534 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:40 crc kubenswrapper[4821]: I1125 10:33:40.985544 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:40Z","lastTransitionTime":"2025-11-25T10:33:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:41 crc kubenswrapper[4821]: I1125 10:33:41.087326 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:41 crc kubenswrapper[4821]: I1125 10:33:41.087358 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:41 crc kubenswrapper[4821]: I1125 10:33:41.087368 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:41 crc kubenswrapper[4821]: I1125 10:33:41.087384 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:41 crc kubenswrapper[4821]: I1125 10:33:41.087394 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:41Z","lastTransitionTime":"2025-11-25T10:33:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:41 crc kubenswrapper[4821]: I1125 10:33:41.114266 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:33:41 crc kubenswrapper[4821]: E1125 10:33:41.114411 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:33:41 crc kubenswrapper[4821]: I1125 10:33:41.114464 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:33:41 crc kubenswrapper[4821]: I1125 10:33:41.114516 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:33:41 crc kubenswrapper[4821]: E1125 10:33:41.114592 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:33:41 crc kubenswrapper[4821]: E1125 10:33:41.114971 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:33:41 crc kubenswrapper[4821]: I1125 10:33:41.115363 4821 scope.go:117] "RemoveContainer" containerID="90d8792a79993c04a0eb774498395d955deff5c42c693892426685d157cf9cff" Nov 25 10:33:41 crc kubenswrapper[4821]: E1125 10:33:41.115570 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-mzhnw_openshift-ovn-kubernetes(954085fa-d03a-4e88-9dfd-566257e3558d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" podUID="954085fa-d03a-4e88-9dfd-566257e3558d" Nov 25 10:33:41 crc kubenswrapper[4821]: I1125 10:33:41.189518 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:41 crc kubenswrapper[4821]: I1125 10:33:41.189565 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:41 crc kubenswrapper[4821]: I1125 10:33:41.189574 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:41 crc kubenswrapper[4821]: I1125 10:33:41.189587 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:41 crc kubenswrapper[4821]: I1125 10:33:41.189595 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:41Z","lastTransitionTime":"2025-11-25T10:33:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:41 crc kubenswrapper[4821]: I1125 10:33:41.291496 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:41 crc kubenswrapper[4821]: I1125 10:33:41.291532 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:41 crc kubenswrapper[4821]: I1125 10:33:41.291540 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:41 crc kubenswrapper[4821]: I1125 10:33:41.291557 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:41 crc kubenswrapper[4821]: I1125 10:33:41.291566 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:41Z","lastTransitionTime":"2025-11-25T10:33:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:41 crc kubenswrapper[4821]: I1125 10:33:41.393142 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:41 crc kubenswrapper[4821]: I1125 10:33:41.393193 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:41 crc kubenswrapper[4821]: I1125 10:33:41.393240 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:41 crc kubenswrapper[4821]: I1125 10:33:41.393255 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:41 crc kubenswrapper[4821]: I1125 10:33:41.393264 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:41Z","lastTransitionTime":"2025-11-25T10:33:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:41 crc kubenswrapper[4821]: I1125 10:33:41.496239 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:41 crc kubenswrapper[4821]: I1125 10:33:41.496281 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:41 crc kubenswrapper[4821]: I1125 10:33:41.496289 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:41 crc kubenswrapper[4821]: I1125 10:33:41.496303 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:41 crc kubenswrapper[4821]: I1125 10:33:41.496314 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:41Z","lastTransitionTime":"2025-11-25T10:33:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:41 crc kubenswrapper[4821]: I1125 10:33:41.599195 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:41 crc kubenswrapper[4821]: I1125 10:33:41.599232 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:41 crc kubenswrapper[4821]: I1125 10:33:41.599244 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:41 crc kubenswrapper[4821]: I1125 10:33:41.599261 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:41 crc kubenswrapper[4821]: I1125 10:33:41.599270 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:41Z","lastTransitionTime":"2025-11-25T10:33:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:41 crc kubenswrapper[4821]: I1125 10:33:41.701608 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:41 crc kubenswrapper[4821]: I1125 10:33:41.701668 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:41 crc kubenswrapper[4821]: I1125 10:33:41.701686 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:41 crc kubenswrapper[4821]: I1125 10:33:41.701708 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:41 crc kubenswrapper[4821]: I1125 10:33:41.701724 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:41Z","lastTransitionTime":"2025-11-25T10:33:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:41 crc kubenswrapper[4821]: I1125 10:33:41.803627 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:41 crc kubenswrapper[4821]: I1125 10:33:41.803672 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:41 crc kubenswrapper[4821]: I1125 10:33:41.803696 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:41 crc kubenswrapper[4821]: I1125 10:33:41.803717 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:41 crc kubenswrapper[4821]: I1125 10:33:41.803728 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:41Z","lastTransitionTime":"2025-11-25T10:33:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:41 crc kubenswrapper[4821]: I1125 10:33:41.905932 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:41 crc kubenswrapper[4821]: I1125 10:33:41.905968 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:41 crc kubenswrapper[4821]: I1125 10:33:41.905984 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:41 crc kubenswrapper[4821]: I1125 10:33:41.905999 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:41 crc kubenswrapper[4821]: I1125 10:33:41.906008 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:41Z","lastTransitionTime":"2025-11-25T10:33:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:42 crc kubenswrapper[4821]: I1125 10:33:42.008286 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:42 crc kubenswrapper[4821]: I1125 10:33:42.008334 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:42 crc kubenswrapper[4821]: I1125 10:33:42.008348 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:42 crc kubenswrapper[4821]: I1125 10:33:42.008367 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:42 crc kubenswrapper[4821]: I1125 10:33:42.008379 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:42Z","lastTransitionTime":"2025-11-25T10:33:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:42 crc kubenswrapper[4821]: I1125 10:33:42.111292 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:42 crc kubenswrapper[4821]: I1125 10:33:42.111363 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:42 crc kubenswrapper[4821]: I1125 10:33:42.111380 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:42 crc kubenswrapper[4821]: I1125 10:33:42.111409 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:42 crc kubenswrapper[4821]: I1125 10:33:42.111427 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:42Z","lastTransitionTime":"2025-11-25T10:33:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:42 crc kubenswrapper[4821]: I1125 10:33:42.113717 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4msn" Nov 25 10:33:42 crc kubenswrapper[4821]: E1125 10:33:42.113892 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4msn" podUID="1f12cef3-7e40-4f61-836e-23cd0e578e71" Nov 25 10:33:42 crc kubenswrapper[4821]: I1125 10:33:42.213822 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:42 crc kubenswrapper[4821]: I1125 10:33:42.213865 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:42 crc kubenswrapper[4821]: I1125 10:33:42.213876 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:42 crc kubenswrapper[4821]: I1125 10:33:42.213892 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:42 crc kubenswrapper[4821]: I1125 10:33:42.213903 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:42Z","lastTransitionTime":"2025-11-25T10:33:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:42 crc kubenswrapper[4821]: I1125 10:33:42.315900 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:42 crc kubenswrapper[4821]: I1125 10:33:42.316011 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:42 crc kubenswrapper[4821]: I1125 10:33:42.316026 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:42 crc kubenswrapper[4821]: I1125 10:33:42.316050 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:42 crc kubenswrapper[4821]: I1125 10:33:42.316070 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:42Z","lastTransitionTime":"2025-11-25T10:33:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:42 crc kubenswrapper[4821]: I1125 10:33:42.418108 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:42 crc kubenswrapper[4821]: I1125 10:33:42.418155 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:42 crc kubenswrapper[4821]: I1125 10:33:42.418189 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:42 crc kubenswrapper[4821]: I1125 10:33:42.418209 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:42 crc kubenswrapper[4821]: I1125 10:33:42.418223 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:42Z","lastTransitionTime":"2025-11-25T10:33:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:42 crc kubenswrapper[4821]: I1125 10:33:42.519899 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:42 crc kubenswrapper[4821]: I1125 10:33:42.519959 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:42 crc kubenswrapper[4821]: I1125 10:33:42.519976 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:42 crc kubenswrapper[4821]: I1125 10:33:42.519997 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:42 crc kubenswrapper[4821]: I1125 10:33:42.520011 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:42Z","lastTransitionTime":"2025-11-25T10:33:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:42 crc kubenswrapper[4821]: I1125 10:33:42.622634 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:42 crc kubenswrapper[4821]: I1125 10:33:42.622680 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:42 crc kubenswrapper[4821]: I1125 10:33:42.622695 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:42 crc kubenswrapper[4821]: I1125 10:33:42.622712 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:42 crc kubenswrapper[4821]: I1125 10:33:42.622722 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:42Z","lastTransitionTime":"2025-11-25T10:33:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:42 crc kubenswrapper[4821]: I1125 10:33:42.725113 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:42 crc kubenswrapper[4821]: I1125 10:33:42.725192 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:42 crc kubenswrapper[4821]: I1125 10:33:42.725209 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:42 crc kubenswrapper[4821]: I1125 10:33:42.725228 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:42 crc kubenswrapper[4821]: I1125 10:33:42.725240 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:42Z","lastTransitionTime":"2025-11-25T10:33:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:42 crc kubenswrapper[4821]: I1125 10:33:42.827260 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:42 crc kubenswrapper[4821]: I1125 10:33:42.827310 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:42 crc kubenswrapper[4821]: I1125 10:33:42.827324 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:42 crc kubenswrapper[4821]: I1125 10:33:42.827340 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:42 crc kubenswrapper[4821]: I1125 10:33:42.827352 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:42Z","lastTransitionTime":"2025-11-25T10:33:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:42 crc kubenswrapper[4821]: I1125 10:33:42.929701 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:42 crc kubenswrapper[4821]: I1125 10:33:42.929742 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:42 crc kubenswrapper[4821]: I1125 10:33:42.929752 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:42 crc kubenswrapper[4821]: I1125 10:33:42.929767 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:42 crc kubenswrapper[4821]: I1125 10:33:42.929777 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:42Z","lastTransitionTime":"2025-11-25T10:33:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:43 crc kubenswrapper[4821]: I1125 10:33:43.057547 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:43 crc kubenswrapper[4821]: I1125 10:33:43.057583 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:43 crc kubenswrapper[4821]: I1125 10:33:43.057592 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:43 crc kubenswrapper[4821]: I1125 10:33:43.057605 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:43 crc kubenswrapper[4821]: I1125 10:33:43.057613 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:43Z","lastTransitionTime":"2025-11-25T10:33:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:43 crc kubenswrapper[4821]: I1125 10:33:43.113373 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:33:43 crc kubenswrapper[4821]: E1125 10:33:43.113490 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:33:43 crc kubenswrapper[4821]: I1125 10:33:43.113675 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:33:43 crc kubenswrapper[4821]: E1125 10:33:43.113736 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:33:43 crc kubenswrapper[4821]: I1125 10:33:43.113841 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:33:43 crc kubenswrapper[4821]: E1125 10:33:43.113888 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:33:43 crc kubenswrapper[4821]: I1125 10:33:43.159305 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:43 crc kubenswrapper[4821]: I1125 10:33:43.159334 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:43 crc kubenswrapper[4821]: I1125 10:33:43.159344 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:43 crc kubenswrapper[4821]: I1125 10:33:43.159359 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:43 crc kubenswrapper[4821]: I1125 10:33:43.159370 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:43Z","lastTransitionTime":"2025-11-25T10:33:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:43 crc kubenswrapper[4821]: I1125 10:33:43.261776 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:43 crc kubenswrapper[4821]: I1125 10:33:43.261810 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:43 crc kubenswrapper[4821]: I1125 10:33:43.261819 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:43 crc kubenswrapper[4821]: I1125 10:33:43.261833 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:43 crc kubenswrapper[4821]: I1125 10:33:43.261842 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:43Z","lastTransitionTime":"2025-11-25T10:33:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:43 crc kubenswrapper[4821]: I1125 10:33:43.364102 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:43 crc kubenswrapper[4821]: I1125 10:33:43.364138 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:43 crc kubenswrapper[4821]: I1125 10:33:43.364147 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:43 crc kubenswrapper[4821]: I1125 10:33:43.364193 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:43 crc kubenswrapper[4821]: I1125 10:33:43.364223 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:43Z","lastTransitionTime":"2025-11-25T10:33:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:43 crc kubenswrapper[4821]: I1125 10:33:43.467140 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:43 crc kubenswrapper[4821]: I1125 10:33:43.467201 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:43 crc kubenswrapper[4821]: I1125 10:33:43.467213 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:43 crc kubenswrapper[4821]: I1125 10:33:43.467229 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:43 crc kubenswrapper[4821]: I1125 10:33:43.467240 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:43Z","lastTransitionTime":"2025-11-25T10:33:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:43 crc kubenswrapper[4821]: I1125 10:33:43.569539 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:43 crc kubenswrapper[4821]: I1125 10:33:43.569580 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:43 crc kubenswrapper[4821]: I1125 10:33:43.569588 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:43 crc kubenswrapper[4821]: I1125 10:33:43.569602 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:43 crc kubenswrapper[4821]: I1125 10:33:43.569611 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:43Z","lastTransitionTime":"2025-11-25T10:33:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:43 crc kubenswrapper[4821]: I1125 10:33:43.671561 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:43 crc kubenswrapper[4821]: I1125 10:33:43.671602 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:43 crc kubenswrapper[4821]: I1125 10:33:43.671613 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:43 crc kubenswrapper[4821]: I1125 10:33:43.671638 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:43 crc kubenswrapper[4821]: I1125 10:33:43.671652 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:43Z","lastTransitionTime":"2025-11-25T10:33:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:43 crc kubenswrapper[4821]: I1125 10:33:43.774307 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:43 crc kubenswrapper[4821]: I1125 10:33:43.774352 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:43 crc kubenswrapper[4821]: I1125 10:33:43.774361 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:43 crc kubenswrapper[4821]: I1125 10:33:43.774374 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:43 crc kubenswrapper[4821]: I1125 10:33:43.774382 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:43Z","lastTransitionTime":"2025-11-25T10:33:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:43 crc kubenswrapper[4821]: I1125 10:33:43.876840 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:43 crc kubenswrapper[4821]: I1125 10:33:43.876880 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:43 crc kubenswrapper[4821]: I1125 10:33:43.876891 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:43 crc kubenswrapper[4821]: I1125 10:33:43.876907 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:43 crc kubenswrapper[4821]: I1125 10:33:43.876917 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:43Z","lastTransitionTime":"2025-11-25T10:33:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:43 crc kubenswrapper[4821]: I1125 10:33:43.979589 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:43 crc kubenswrapper[4821]: I1125 10:33:43.979630 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:43 crc kubenswrapper[4821]: I1125 10:33:43.979642 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:43 crc kubenswrapper[4821]: I1125 10:33:43.979657 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:43 crc kubenswrapper[4821]: I1125 10:33:43.979667 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:43Z","lastTransitionTime":"2025-11-25T10:33:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:44 crc kubenswrapper[4821]: I1125 10:33:44.082485 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:44 crc kubenswrapper[4821]: I1125 10:33:44.082532 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:44 crc kubenswrapper[4821]: I1125 10:33:44.082544 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:44 crc kubenswrapper[4821]: I1125 10:33:44.082562 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:44 crc kubenswrapper[4821]: I1125 10:33:44.082574 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:44Z","lastTransitionTime":"2025-11-25T10:33:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:44 crc kubenswrapper[4821]: I1125 10:33:44.113488 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4msn" Nov 25 10:33:44 crc kubenswrapper[4821]: E1125 10:33:44.113645 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4msn" podUID="1f12cef3-7e40-4f61-836e-23cd0e578e71" Nov 25 10:33:44 crc kubenswrapper[4821]: I1125 10:33:44.184734 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:44 crc kubenswrapper[4821]: I1125 10:33:44.184780 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:44 crc kubenswrapper[4821]: I1125 10:33:44.184791 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:44 crc kubenswrapper[4821]: I1125 10:33:44.184808 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:44 crc kubenswrapper[4821]: I1125 10:33:44.184822 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:44Z","lastTransitionTime":"2025-11-25T10:33:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:44 crc kubenswrapper[4821]: I1125 10:33:44.287420 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:44 crc kubenswrapper[4821]: I1125 10:33:44.287452 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:44 crc kubenswrapper[4821]: I1125 10:33:44.287470 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:44 crc kubenswrapper[4821]: I1125 10:33:44.287488 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:44 crc kubenswrapper[4821]: I1125 10:33:44.287499 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:44Z","lastTransitionTime":"2025-11-25T10:33:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:44 crc kubenswrapper[4821]: I1125 10:33:44.389426 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:44 crc kubenswrapper[4821]: I1125 10:33:44.389498 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:44 crc kubenswrapper[4821]: I1125 10:33:44.389511 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:44 crc kubenswrapper[4821]: I1125 10:33:44.389528 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:44 crc kubenswrapper[4821]: I1125 10:33:44.389539 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:44Z","lastTransitionTime":"2025-11-25T10:33:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:44 crc kubenswrapper[4821]: I1125 10:33:44.491575 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:44 crc kubenswrapper[4821]: I1125 10:33:44.491630 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:44 crc kubenswrapper[4821]: I1125 10:33:44.491647 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:44 crc kubenswrapper[4821]: I1125 10:33:44.491667 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:44 crc kubenswrapper[4821]: I1125 10:33:44.491683 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:44Z","lastTransitionTime":"2025-11-25T10:33:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:44 crc kubenswrapper[4821]: I1125 10:33:44.593978 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:44 crc kubenswrapper[4821]: I1125 10:33:44.594018 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:44 crc kubenswrapper[4821]: I1125 10:33:44.594026 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:44 crc kubenswrapper[4821]: I1125 10:33:44.594041 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:44 crc kubenswrapper[4821]: I1125 10:33:44.594067 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:44Z","lastTransitionTime":"2025-11-25T10:33:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:44 crc kubenswrapper[4821]: I1125 10:33:44.696079 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:44 crc kubenswrapper[4821]: I1125 10:33:44.696122 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:44 crc kubenswrapper[4821]: I1125 10:33:44.696131 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:44 crc kubenswrapper[4821]: I1125 10:33:44.696148 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:44 crc kubenswrapper[4821]: I1125 10:33:44.696175 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:44Z","lastTransitionTime":"2025-11-25T10:33:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:44 crc kubenswrapper[4821]: I1125 10:33:44.798695 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:44 crc kubenswrapper[4821]: I1125 10:33:44.798742 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:44 crc kubenswrapper[4821]: I1125 10:33:44.798753 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:44 crc kubenswrapper[4821]: I1125 10:33:44.798769 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:44 crc kubenswrapper[4821]: I1125 10:33:44.798781 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:44Z","lastTransitionTime":"2025-11-25T10:33:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:44 crc kubenswrapper[4821]: I1125 10:33:44.901253 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:44 crc kubenswrapper[4821]: I1125 10:33:44.901300 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:44 crc kubenswrapper[4821]: I1125 10:33:44.901309 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:44 crc kubenswrapper[4821]: I1125 10:33:44.901329 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:44 crc kubenswrapper[4821]: I1125 10:33:44.901338 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:44Z","lastTransitionTime":"2025-11-25T10:33:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:45 crc kubenswrapper[4821]: I1125 10:33:45.003401 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:45 crc kubenswrapper[4821]: I1125 10:33:45.003444 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:45 crc kubenswrapper[4821]: I1125 10:33:45.003455 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:45 crc kubenswrapper[4821]: I1125 10:33:45.003470 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:45 crc kubenswrapper[4821]: I1125 10:33:45.003480 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:45Z","lastTransitionTime":"2025-11-25T10:33:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:45 crc kubenswrapper[4821]: I1125 10:33:45.105911 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:45 crc kubenswrapper[4821]: I1125 10:33:45.105953 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:45 crc kubenswrapper[4821]: I1125 10:33:45.105964 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:45 crc kubenswrapper[4821]: I1125 10:33:45.105981 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:45 crc kubenswrapper[4821]: I1125 10:33:45.105993 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:45Z","lastTransitionTime":"2025-11-25T10:33:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:45 crc kubenswrapper[4821]: I1125 10:33:45.114152 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:33:45 crc kubenswrapper[4821]: I1125 10:33:45.114204 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:33:45 crc kubenswrapper[4821]: E1125 10:33:45.114381 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:33:45 crc kubenswrapper[4821]: I1125 10:33:45.114429 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:33:45 crc kubenswrapper[4821]: E1125 10:33:45.115034 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:33:45 crc kubenswrapper[4821]: E1125 10:33:45.114646 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:33:45 crc kubenswrapper[4821]: I1125 10:33:45.208313 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:45 crc kubenswrapper[4821]: I1125 10:33:45.208348 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:45 crc kubenswrapper[4821]: I1125 10:33:45.208357 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:45 crc kubenswrapper[4821]: I1125 10:33:45.208373 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:45 crc kubenswrapper[4821]: I1125 10:33:45.208384 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:45Z","lastTransitionTime":"2025-11-25T10:33:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:45 crc kubenswrapper[4821]: I1125 10:33:45.310515 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:45 crc kubenswrapper[4821]: I1125 10:33:45.310561 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:45 crc kubenswrapper[4821]: I1125 10:33:45.310577 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:45 crc kubenswrapper[4821]: I1125 10:33:45.310592 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:45 crc kubenswrapper[4821]: I1125 10:33:45.310604 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:45Z","lastTransitionTime":"2025-11-25T10:33:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:45 crc kubenswrapper[4821]: I1125 10:33:45.412847 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:45 crc kubenswrapper[4821]: I1125 10:33:45.412894 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:45 crc kubenswrapper[4821]: I1125 10:33:45.412902 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:45 crc kubenswrapper[4821]: I1125 10:33:45.412916 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:45 crc kubenswrapper[4821]: I1125 10:33:45.412925 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:45Z","lastTransitionTime":"2025-11-25T10:33:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:45 crc kubenswrapper[4821]: I1125 10:33:45.514409 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:45 crc kubenswrapper[4821]: I1125 10:33:45.514447 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:45 crc kubenswrapper[4821]: I1125 10:33:45.514458 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:45 crc kubenswrapper[4821]: I1125 10:33:45.514474 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:45 crc kubenswrapper[4821]: I1125 10:33:45.514486 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:45Z","lastTransitionTime":"2025-11-25T10:33:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:45 crc kubenswrapper[4821]: I1125 10:33:45.616310 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:45 crc kubenswrapper[4821]: I1125 10:33:45.616355 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:45 crc kubenswrapper[4821]: I1125 10:33:45.616371 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:45 crc kubenswrapper[4821]: I1125 10:33:45.616387 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:45 crc kubenswrapper[4821]: I1125 10:33:45.616396 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:45Z","lastTransitionTime":"2025-11-25T10:33:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:45 crc kubenswrapper[4821]: I1125 10:33:45.718583 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:45 crc kubenswrapper[4821]: I1125 10:33:45.718626 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:45 crc kubenswrapper[4821]: I1125 10:33:45.718636 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:45 crc kubenswrapper[4821]: I1125 10:33:45.718651 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:45 crc kubenswrapper[4821]: I1125 10:33:45.718661 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:45Z","lastTransitionTime":"2025-11-25T10:33:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:45 crc kubenswrapper[4821]: I1125 10:33:45.821378 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:45 crc kubenswrapper[4821]: I1125 10:33:45.821417 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:45 crc kubenswrapper[4821]: I1125 10:33:45.821426 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:45 crc kubenswrapper[4821]: I1125 10:33:45.821441 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:45 crc kubenswrapper[4821]: I1125 10:33:45.821454 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:45Z","lastTransitionTime":"2025-11-25T10:33:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:45 crc kubenswrapper[4821]: I1125 10:33:45.923554 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:45 crc kubenswrapper[4821]: I1125 10:33:45.923588 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:45 crc kubenswrapper[4821]: I1125 10:33:45.923597 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:45 crc kubenswrapper[4821]: I1125 10:33:45.923609 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:45 crc kubenswrapper[4821]: I1125 10:33:45.923618 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:45Z","lastTransitionTime":"2025-11-25T10:33:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:46 crc kubenswrapper[4821]: I1125 10:33:46.026463 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:46 crc kubenswrapper[4821]: I1125 10:33:46.026510 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:46 crc kubenswrapper[4821]: I1125 10:33:46.026522 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:46 crc kubenswrapper[4821]: I1125 10:33:46.026540 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:46 crc kubenswrapper[4821]: I1125 10:33:46.026549 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:46Z","lastTransitionTime":"2025-11-25T10:33:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:46 crc kubenswrapper[4821]: I1125 10:33:46.113719 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4msn" Nov 25 10:33:46 crc kubenswrapper[4821]: E1125 10:33:46.113860 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4msn" podUID="1f12cef3-7e40-4f61-836e-23cd0e578e71" Nov 25 10:33:46 crc kubenswrapper[4821]: I1125 10:33:46.128623 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:46 crc kubenswrapper[4821]: I1125 10:33:46.128664 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:46 crc kubenswrapper[4821]: I1125 10:33:46.128673 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:46 crc kubenswrapper[4821]: I1125 10:33:46.128686 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:46 crc kubenswrapper[4821]: I1125 10:33:46.128697 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:46Z","lastTransitionTime":"2025-11-25T10:33:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:46 crc kubenswrapper[4821]: I1125 10:33:46.231252 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:46 crc kubenswrapper[4821]: I1125 10:33:46.231291 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:46 crc kubenswrapper[4821]: I1125 10:33:46.231301 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:46 crc kubenswrapper[4821]: I1125 10:33:46.231341 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:46 crc kubenswrapper[4821]: I1125 10:33:46.231353 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:46Z","lastTransitionTime":"2025-11-25T10:33:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:46 crc kubenswrapper[4821]: I1125 10:33:46.333962 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:46 crc kubenswrapper[4821]: I1125 10:33:46.334003 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:46 crc kubenswrapper[4821]: I1125 10:33:46.334015 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:46 crc kubenswrapper[4821]: I1125 10:33:46.334032 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:46 crc kubenswrapper[4821]: I1125 10:33:46.334043 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:46Z","lastTransitionTime":"2025-11-25T10:33:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:46 crc kubenswrapper[4821]: I1125 10:33:46.436030 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:46 crc kubenswrapper[4821]: I1125 10:33:46.436085 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:46 crc kubenswrapper[4821]: I1125 10:33:46.436103 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:46 crc kubenswrapper[4821]: I1125 10:33:46.436124 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:46 crc kubenswrapper[4821]: I1125 10:33:46.436138 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:46Z","lastTransitionTime":"2025-11-25T10:33:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:46 crc kubenswrapper[4821]: I1125 10:33:46.538888 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:46 crc kubenswrapper[4821]: I1125 10:33:46.538923 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:46 crc kubenswrapper[4821]: I1125 10:33:46.538932 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:46 crc kubenswrapper[4821]: I1125 10:33:46.538947 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:46 crc kubenswrapper[4821]: I1125 10:33:46.538957 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:46Z","lastTransitionTime":"2025-11-25T10:33:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:46 crc kubenswrapper[4821]: I1125 10:33:46.641685 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:46 crc kubenswrapper[4821]: I1125 10:33:46.641732 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:46 crc kubenswrapper[4821]: I1125 10:33:46.641742 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:46 crc kubenswrapper[4821]: I1125 10:33:46.641759 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:46 crc kubenswrapper[4821]: I1125 10:33:46.641773 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:46Z","lastTransitionTime":"2025-11-25T10:33:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:46 crc kubenswrapper[4821]: I1125 10:33:46.743778 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:46 crc kubenswrapper[4821]: I1125 10:33:46.743806 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:46 crc kubenswrapper[4821]: I1125 10:33:46.743815 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:46 crc kubenswrapper[4821]: I1125 10:33:46.743829 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:46 crc kubenswrapper[4821]: I1125 10:33:46.743838 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:46Z","lastTransitionTime":"2025-11-25T10:33:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:46 crc kubenswrapper[4821]: I1125 10:33:46.846512 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:46 crc kubenswrapper[4821]: I1125 10:33:46.846564 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:46 crc kubenswrapper[4821]: I1125 10:33:46.846579 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:46 crc kubenswrapper[4821]: I1125 10:33:46.846600 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:46 crc kubenswrapper[4821]: I1125 10:33:46.846612 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:46Z","lastTransitionTime":"2025-11-25T10:33:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:46 crc kubenswrapper[4821]: I1125 10:33:46.949148 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:46 crc kubenswrapper[4821]: I1125 10:33:46.949211 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:46 crc kubenswrapper[4821]: I1125 10:33:46.949222 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:46 crc kubenswrapper[4821]: I1125 10:33:46.949239 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:46 crc kubenswrapper[4821]: I1125 10:33:46.949250 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:46Z","lastTransitionTime":"2025-11-25T10:33:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.051970 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.052030 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.052041 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.052057 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.052069 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:47Z","lastTransitionTime":"2025-11-25T10:33:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.113825 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.113865 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.113827 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:33:47 crc kubenswrapper[4821]: E1125 10:33:47.113960 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:33:47 crc kubenswrapper[4821]: E1125 10:33:47.114034 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:33:47 crc kubenswrapper[4821]: E1125 10:33:47.114091 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.154275 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.154324 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.154333 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.154347 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.154357 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:47Z","lastTransitionTime":"2025-11-25T10:33:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.256980 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.257021 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.257030 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.257044 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.257053 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:47Z","lastTransitionTime":"2025-11-25T10:33:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.358937 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.358969 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.358977 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.358990 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.358999 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:47Z","lastTransitionTime":"2025-11-25T10:33:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.460699 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.460741 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.460751 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.460766 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.460778 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:47Z","lastTransitionTime":"2025-11-25T10:33:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.511715 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.511762 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.511776 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.511792 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.511802 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:47Z","lastTransitionTime":"2025-11-25T10:33:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:47 crc kubenswrapper[4821]: E1125 10:33:47.523374 4821 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a5b2b070-39b9-412c-90f1-c7ed2e9e0950\\\",\\\"systemUUID\\\":\\\"ea86e80f-41b7-4190-ad7e-d8c549d535de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:47Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.526415 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.526447 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.526455 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.526468 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.526477 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:47Z","lastTransitionTime":"2025-11-25T10:33:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:47 crc kubenswrapper[4821]: E1125 10:33:47.537323 4821 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a5b2b070-39b9-412c-90f1-c7ed2e9e0950\\\",\\\"systemUUID\\\":\\\"ea86e80f-41b7-4190-ad7e-d8c549d535de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:47Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.540735 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.540772 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.540780 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.540796 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.540806 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:47Z","lastTransitionTime":"2025-11-25T10:33:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:47 crc kubenswrapper[4821]: E1125 10:33:47.553972 4821 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a5b2b070-39b9-412c-90f1-c7ed2e9e0950\\\",\\\"systemUUID\\\":\\\"ea86e80f-41b7-4190-ad7e-d8c549d535de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:47Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.557193 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.557237 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.557247 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.557263 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.557272 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:47Z","lastTransitionTime":"2025-11-25T10:33:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:47 crc kubenswrapper[4821]: E1125 10:33:47.567995 4821 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a5b2b070-39b9-412c-90f1-c7ed2e9e0950\\\",\\\"systemUUID\\\":\\\"ea86e80f-41b7-4190-ad7e-d8c549d535de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:47Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.571087 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.571142 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.571152 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.571181 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.571200 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:47Z","lastTransitionTime":"2025-11-25T10:33:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:47 crc kubenswrapper[4821]: E1125 10:33:47.581452 4821 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T10:33:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T10:33:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a5b2b070-39b9-412c-90f1-c7ed2e9e0950\\\",\\\"systemUUID\\\":\\\"ea86e80f-41b7-4190-ad7e-d8c549d535de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:47Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:47 crc kubenswrapper[4821]: E1125 10:33:47.581590 4821 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.582813 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.582840 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.582850 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.582867 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.582878 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:47Z","lastTransitionTime":"2025-11-25T10:33:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.685660 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.685697 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.685706 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.685722 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.685734 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:47Z","lastTransitionTime":"2025-11-25T10:33:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.788240 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.788274 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.788285 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.788300 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.788311 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:47Z","lastTransitionTime":"2025-11-25T10:33:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.890853 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.890895 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.890909 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.890924 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.890935 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:47Z","lastTransitionTime":"2025-11-25T10:33:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.992667 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.992702 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.992712 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.992727 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:47 crc kubenswrapper[4821]: I1125 10:33:47.992737 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:47Z","lastTransitionTime":"2025-11-25T10:33:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:48 crc kubenswrapper[4821]: I1125 10:33:48.095042 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:48 crc kubenswrapper[4821]: I1125 10:33:48.095069 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:48 crc kubenswrapper[4821]: I1125 10:33:48.095077 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:48 crc kubenswrapper[4821]: I1125 10:33:48.095089 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:48 crc kubenswrapper[4821]: I1125 10:33:48.095100 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:48Z","lastTransitionTime":"2025-11-25T10:33:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:48 crc kubenswrapper[4821]: I1125 10:33:48.114505 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4msn" Nov 25 10:33:48 crc kubenswrapper[4821]: E1125 10:33:48.114655 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4msn" podUID="1f12cef3-7e40-4f61-836e-23cd0e578e71" Nov 25 10:33:48 crc kubenswrapper[4821]: I1125 10:33:48.197311 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:48 crc kubenswrapper[4821]: I1125 10:33:48.197341 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:48 crc kubenswrapper[4821]: I1125 10:33:48.197349 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:48 crc kubenswrapper[4821]: I1125 10:33:48.197367 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:48 crc kubenswrapper[4821]: I1125 10:33:48.197375 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:48Z","lastTransitionTime":"2025-11-25T10:33:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:48 crc kubenswrapper[4821]: I1125 10:33:48.299677 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:48 crc kubenswrapper[4821]: I1125 10:33:48.299735 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:48 crc kubenswrapper[4821]: I1125 10:33:48.299744 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:48 crc kubenswrapper[4821]: I1125 10:33:48.299757 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:48 crc kubenswrapper[4821]: I1125 10:33:48.299766 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:48Z","lastTransitionTime":"2025-11-25T10:33:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:48 crc kubenswrapper[4821]: I1125 10:33:48.402099 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:48 crc kubenswrapper[4821]: I1125 10:33:48.402131 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:48 crc kubenswrapper[4821]: I1125 10:33:48.402139 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:48 crc kubenswrapper[4821]: I1125 10:33:48.402151 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:48 crc kubenswrapper[4821]: I1125 10:33:48.402181 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:48Z","lastTransitionTime":"2025-11-25T10:33:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:48 crc kubenswrapper[4821]: I1125 10:33:48.427915 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1f12cef3-7e40-4f61-836e-23cd0e578e71-metrics-certs\") pod \"network-metrics-daemon-d4msn\" (UID: \"1f12cef3-7e40-4f61-836e-23cd0e578e71\") " pod="openshift-multus/network-metrics-daemon-d4msn" Nov 25 10:33:48 crc kubenswrapper[4821]: E1125 10:33:48.428030 4821 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 10:33:48 crc kubenswrapper[4821]: E1125 10:33:48.428091 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1f12cef3-7e40-4f61-836e-23cd0e578e71-metrics-certs podName:1f12cef3-7e40-4f61-836e-23cd0e578e71 nodeName:}" failed. No retries permitted until 2025-11-25 10:34:52.428074789 +0000 UTC m=+162.964394636 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/1f12cef3-7e40-4f61-836e-23cd0e578e71-metrics-certs") pod "network-metrics-daemon-d4msn" (UID: "1f12cef3-7e40-4f61-836e-23cd0e578e71") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 10:33:48 crc kubenswrapper[4821]: I1125 10:33:48.504389 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:48 crc kubenswrapper[4821]: I1125 10:33:48.504485 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:48 crc kubenswrapper[4821]: I1125 10:33:48.504507 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:48 crc kubenswrapper[4821]: I1125 10:33:48.504538 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:48 crc kubenswrapper[4821]: I1125 10:33:48.504561 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:48Z","lastTransitionTime":"2025-11-25T10:33:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:48 crc kubenswrapper[4821]: I1125 10:33:48.606717 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:48 crc kubenswrapper[4821]: I1125 10:33:48.606765 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:48 crc kubenswrapper[4821]: I1125 10:33:48.606773 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:48 crc kubenswrapper[4821]: I1125 10:33:48.606788 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:48 crc kubenswrapper[4821]: I1125 10:33:48.606799 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:48Z","lastTransitionTime":"2025-11-25T10:33:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:48 crc kubenswrapper[4821]: I1125 10:33:48.709088 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:48 crc kubenswrapper[4821]: I1125 10:33:48.709139 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:48 crc kubenswrapper[4821]: I1125 10:33:48.709150 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:48 crc kubenswrapper[4821]: I1125 10:33:48.709188 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:48 crc kubenswrapper[4821]: I1125 10:33:48.709200 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:48Z","lastTransitionTime":"2025-11-25T10:33:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:48 crc kubenswrapper[4821]: I1125 10:33:48.812056 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:48 crc kubenswrapper[4821]: I1125 10:33:48.812110 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:48 crc kubenswrapper[4821]: I1125 10:33:48.812119 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:48 crc kubenswrapper[4821]: I1125 10:33:48.812134 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:48 crc kubenswrapper[4821]: I1125 10:33:48.812143 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:48Z","lastTransitionTime":"2025-11-25T10:33:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:48 crc kubenswrapper[4821]: I1125 10:33:48.913990 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:48 crc kubenswrapper[4821]: I1125 10:33:48.914032 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:48 crc kubenswrapper[4821]: I1125 10:33:48.914044 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:48 crc kubenswrapper[4821]: I1125 10:33:48.914064 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:48 crc kubenswrapper[4821]: I1125 10:33:48.914078 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:48Z","lastTransitionTime":"2025-11-25T10:33:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:49 crc kubenswrapper[4821]: I1125 10:33:49.016341 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:49 crc kubenswrapper[4821]: I1125 10:33:49.016383 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:49 crc kubenswrapper[4821]: I1125 10:33:49.016392 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:49 crc kubenswrapper[4821]: I1125 10:33:49.016411 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:49 crc kubenswrapper[4821]: I1125 10:33:49.016420 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:49Z","lastTransitionTime":"2025-11-25T10:33:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:49 crc kubenswrapper[4821]: I1125 10:33:49.113785 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:33:49 crc kubenswrapper[4821]: I1125 10:33:49.113857 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:33:49 crc kubenswrapper[4821]: I1125 10:33:49.113801 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:33:49 crc kubenswrapper[4821]: E1125 10:33:49.113966 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:33:49 crc kubenswrapper[4821]: E1125 10:33:49.114075 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:33:49 crc kubenswrapper[4821]: E1125 10:33:49.114150 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:33:49 crc kubenswrapper[4821]: I1125 10:33:49.118531 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:49 crc kubenswrapper[4821]: I1125 10:33:49.118565 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:49 crc kubenswrapper[4821]: I1125 10:33:49.118575 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:49 crc kubenswrapper[4821]: I1125 10:33:49.118591 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:49 crc kubenswrapper[4821]: I1125 10:33:49.118644 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:49Z","lastTransitionTime":"2025-11-25T10:33:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:49 crc kubenswrapper[4821]: I1125 10:33:49.221416 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:49 crc kubenswrapper[4821]: I1125 10:33:49.221467 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:49 crc kubenswrapper[4821]: I1125 10:33:49.221476 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:49 crc kubenswrapper[4821]: I1125 10:33:49.221491 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:49 crc kubenswrapper[4821]: I1125 10:33:49.221503 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:49Z","lastTransitionTime":"2025-11-25T10:33:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:49 crc kubenswrapper[4821]: I1125 10:33:49.323838 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:49 crc kubenswrapper[4821]: I1125 10:33:49.323874 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:49 crc kubenswrapper[4821]: I1125 10:33:49.323942 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:49 crc kubenswrapper[4821]: I1125 10:33:49.323955 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:49 crc kubenswrapper[4821]: I1125 10:33:49.323965 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:49Z","lastTransitionTime":"2025-11-25T10:33:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:49 crc kubenswrapper[4821]: I1125 10:33:49.425791 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:49 crc kubenswrapper[4821]: I1125 10:33:49.425843 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:49 crc kubenswrapper[4821]: I1125 10:33:49.425853 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:49 crc kubenswrapper[4821]: I1125 10:33:49.425868 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:49 crc kubenswrapper[4821]: I1125 10:33:49.425877 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:49Z","lastTransitionTime":"2025-11-25T10:33:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:49 crc kubenswrapper[4821]: I1125 10:33:49.527538 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:49 crc kubenswrapper[4821]: I1125 10:33:49.527597 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:49 crc kubenswrapper[4821]: I1125 10:33:49.527615 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:49 crc kubenswrapper[4821]: I1125 10:33:49.527639 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:49 crc kubenswrapper[4821]: I1125 10:33:49.527658 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:49Z","lastTransitionTime":"2025-11-25T10:33:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:49 crc kubenswrapper[4821]: I1125 10:33:49.630006 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:49 crc kubenswrapper[4821]: I1125 10:33:49.630040 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:49 crc kubenswrapper[4821]: I1125 10:33:49.630048 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:49 crc kubenswrapper[4821]: I1125 10:33:49.630062 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:49 crc kubenswrapper[4821]: I1125 10:33:49.630070 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:49Z","lastTransitionTime":"2025-11-25T10:33:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:49 crc kubenswrapper[4821]: I1125 10:33:49.732286 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:49 crc kubenswrapper[4821]: I1125 10:33:49.732354 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:49 crc kubenswrapper[4821]: I1125 10:33:49.732365 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:49 crc kubenswrapper[4821]: I1125 10:33:49.732381 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:49 crc kubenswrapper[4821]: I1125 10:33:49.732392 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:49Z","lastTransitionTime":"2025-11-25T10:33:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:49 crc kubenswrapper[4821]: I1125 10:33:49.834686 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:49 crc kubenswrapper[4821]: I1125 10:33:49.834723 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:49 crc kubenswrapper[4821]: I1125 10:33:49.834733 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:49 crc kubenswrapper[4821]: I1125 10:33:49.834747 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:49 crc kubenswrapper[4821]: I1125 10:33:49.834755 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:49Z","lastTransitionTime":"2025-11-25T10:33:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:49 crc kubenswrapper[4821]: I1125 10:33:49.937482 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:49 crc kubenswrapper[4821]: I1125 10:33:49.937548 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:49 crc kubenswrapper[4821]: I1125 10:33:49.937565 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:49 crc kubenswrapper[4821]: I1125 10:33:49.937590 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:49 crc kubenswrapper[4821]: I1125 10:33:49.937612 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:49Z","lastTransitionTime":"2025-11-25T10:33:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:50 crc kubenswrapper[4821]: I1125 10:33:50.040997 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:50 crc kubenswrapper[4821]: I1125 10:33:50.041058 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:50 crc kubenswrapper[4821]: I1125 10:33:50.041075 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:50 crc kubenswrapper[4821]: I1125 10:33:50.041102 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:50 crc kubenswrapper[4821]: I1125 10:33:50.041119 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:50Z","lastTransitionTime":"2025-11-25T10:33:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:50 crc kubenswrapper[4821]: I1125 10:33:50.114351 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4msn" Nov 25 10:33:50 crc kubenswrapper[4821]: E1125 10:33:50.114781 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4msn" podUID="1f12cef3-7e40-4f61-836e-23cd0e578e71" Nov 25 10:33:50 crc kubenswrapper[4821]: I1125 10:33:50.137716 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:50Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:50 crc kubenswrapper[4821]: I1125 10:33:50.143996 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:50 crc kubenswrapper[4821]: I1125 10:33:50.144086 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:50 crc kubenswrapper[4821]: I1125 10:33:50.144116 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:50 crc kubenswrapper[4821]: I1125 10:33:50.144154 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:50 crc kubenswrapper[4821]: I1125 10:33:50.144228 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:50Z","lastTransitionTime":"2025-11-25T10:33:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:50 crc kubenswrapper[4821]: I1125 10:33:50.153502 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f948b87-ac86-4de6-ad64-c2ef947f84d4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4482f73a45bc82bd55770c3fd788becacec94893fed1c305fed1c81c753d43aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ba68bd093be94d0d40f3851638db5d26bdfc52d175b1ac78afb91dda7affcfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qq48b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2krbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:50Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:50 crc kubenswrapper[4821]: I1125 10:33:50.175288 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qppnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"357491b8-cd9c-486e-9898-5ba33678513d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc88d700ca3a449ef1cb32d6124b16d0024d6981ced12e276083f86a1b322ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd9c047c5ce1d896f803ff6cafa9bbf3a18a391b44539e1f1312e08c89e51cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd9c047c5ce1d896f803ff6cafa9bbf3a18a391b44539e1f1312e08c89e51cb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://565a6db63fb69abbba81ffdf6085797900858d263cd6affcb0b5f195474c2516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://565a6db63fb69abbba81ffdf6085797900858d263cd6affcb0b5f195474c2516\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://544885302dffff8336a9aa3c4afa80d69088ffaa2402648f4a329e0a798a8c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://544885302dffff8336a9aa3c4afa80d69088ffaa2402648f4a329e0a798a8c99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8bc193fb3f85b980c31af2e826cee90adb633ad2ee82ff1d46e182ab52e8fab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8bc193fb3f85b980c31af2e826cee90adb633ad2ee82ff1d46e182ab52e8fab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7713e2a35bac8df1e8987fbcd452459fa3541caad0c52ed57fda187bd7c83d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7713e2a35bac8df1e8987fbcd452459fa3541caad0c52ed57fda187bd7c83d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34adfca89026fc20007a73a8cbd4e7c7b3137a581c3476867af17fabe976f71f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34adfca89026fc20007a73a8cbd4e7c7b3137a581c3476867af17fabe976f71f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hxd4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:30Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qppnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:50Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:50 crc kubenswrapper[4821]: I1125 10:33:50.192863 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tgdlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2336955b-a703-4db4-8d0f-d08088ed59b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb0362942bfac635ee6292501b1b851772c01d75dd658eef4b17669a716de72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vjfp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:32Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tgdlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:50Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:50 crc kubenswrapper[4821]: I1125 10:33:50.210605 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62fcb632-f12a-46e2-bb7d-a5311f291ee7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a28d43531927e75c3745f100940f7f6bf79fd21026c1309194df8cd7a39361b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69633332d03d2551ff4abb253d358de398eab18486b92db630ce06c0447fb9cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ccb78d9a51ac83ee208c2451685a2512de9deaf742a663bb8c892aa59b76162\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27826e155a9352b1e1866e83fdab683dbcd749cc04b6a4da2db25ecabe3d9a52\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://991252e5f56f128827a5054b8e228284bc0266166582cbaedaadc53ca2f5b401\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d505cb2c9a8efbeea321aae8b505febf1777170ad16e92617df2959bb511b7b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:50Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:50 crc kubenswrapper[4821]: I1125 10:33:50.223541 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63b471fd-5e16-4761-8b07-44a135d5f5f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a68a57fa7fba8f351f264e427a50bed23f57775644e9c268aa7305efb6916794\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67df7c609d5a5d658be4c3e873ec3052bd070ad9469f8ff44c29a9571e3e5d95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://269ced131bc0777dc6752070470c95b4bedfe4b4240940eabd769500ddef0404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e48c0e3b58359c6c0a493d58f7d2aec6011747791a3c0d7788102e1fec18865\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e48c0e3b58359c6c0a493d58f7d2aec6011747791a3c0d7788102e1fec18865\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T10:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T10:32:11Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T10:32:10Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:50Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:50 crc kubenswrapper[4821]: I1125 10:33:50.236264 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06f949936bac99cfbf89d720c9549c074351d9077bd963fa024d87004c6be23e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:50Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:50 crc kubenswrapper[4821]: I1125 10:33:50.246409 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:50 crc kubenswrapper[4821]: I1125 10:33:50.246446 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:50 crc kubenswrapper[4821]: I1125 10:33:50.246457 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:50 crc kubenswrapper[4821]: I1125 10:33:50.246475 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:50 crc kubenswrapper[4821]: I1125 10:33:50.246488 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:50Z","lastTransitionTime":"2025-11-25T10:33:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:50 crc kubenswrapper[4821]: I1125 10:33:50.247814 4821 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T10:32:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384d8ba58264ab89796d3b53afb0c4c1be33cb4fba719a8beef4afc2be7d66a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T10:32:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T10:33:50Z is after 2025-08-24T17:21:41Z" Nov 25 10:33:50 crc kubenswrapper[4821]: I1125 10:33:50.268291 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-zlk96" podStartSLOduration=80.268273848 podStartE2EDuration="1m20.268273848s" podCreationTimestamp="2025-11-25 10:32:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:33:50.268153305 +0000 UTC m=+100.804473162" watchObservedRunningTime="2025-11-25 10:33:50.268273848 +0000 UTC m=+100.804593695" Nov 25 10:33:50 crc kubenswrapper[4821]: I1125 10:33:50.282458 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6hbqd" podStartSLOduration=80.28242011 podStartE2EDuration="1m20.28242011s" podCreationTimestamp="2025-11-25 10:32:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:33:50.280016996 +0000 UTC m=+100.816336873" watchObservedRunningTime="2025-11-25 10:33:50.28242011 +0000 UTC m=+100.818739957" Nov 25 10:33:50 crc kubenswrapper[4821]: I1125 10:33:50.326022 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=78.326002365 podStartE2EDuration="1m18.326002365s" podCreationTimestamp="2025-11-25 10:32:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:33:50.32582526 +0000 UTC m=+100.862145107" watchObservedRunningTime="2025-11-25 10:33:50.326002365 +0000 UTC m=+100.862322222" Nov 25 10:33:50 crc kubenswrapper[4821]: I1125 10:33:50.348565 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-ldzbz" podStartSLOduration=80.348548549 podStartE2EDuration="1m20.348548549s" podCreationTimestamp="2025-11-25 10:32:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:33:50.348180609 +0000 UTC m=+100.884500456" watchObservedRunningTime="2025-11-25 10:33:50.348548549 +0000 UTC m=+100.884868396" Nov 25 10:33:50 crc kubenswrapper[4821]: I1125 10:33:50.349649 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:50 crc kubenswrapper[4821]: I1125 10:33:50.349796 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:50 crc kubenswrapper[4821]: I1125 10:33:50.349886 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:50 crc kubenswrapper[4821]: I1125 10:33:50.349991 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:50 crc kubenswrapper[4821]: I1125 10:33:50.350074 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:50Z","lastTransitionTime":"2025-11-25T10:33:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:50 crc kubenswrapper[4821]: I1125 10:33:50.416721 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=82.416700342 podStartE2EDuration="1m22.416700342s" podCreationTimestamp="2025-11-25 10:32:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:33:50.414690736 +0000 UTC m=+100.951010593" watchObservedRunningTime="2025-11-25 10:33:50.416700342 +0000 UTC m=+100.953020189" Nov 25 10:33:50 crc kubenswrapper[4821]: I1125 10:33:50.417647 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=27.417640104 podStartE2EDuration="27.417640104s" podCreationTimestamp="2025-11-25 10:33:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:33:50.393676637 +0000 UTC m=+100.929996484" watchObservedRunningTime="2025-11-25 10:33:50.417640104 +0000 UTC m=+100.953959951" Nov 25 10:33:50 crc kubenswrapper[4821]: I1125 10:33:50.452051 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:50 crc kubenswrapper[4821]: I1125 10:33:50.452086 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:50 crc kubenswrapper[4821]: I1125 10:33:50.452095 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:50 crc kubenswrapper[4821]: I1125 10:33:50.452109 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:50 crc kubenswrapper[4821]: I1125 10:33:50.452120 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:50Z","lastTransitionTime":"2025-11-25T10:33:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:50 crc kubenswrapper[4821]: I1125 10:33:50.553941 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:50 crc kubenswrapper[4821]: I1125 10:33:50.553984 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:50 crc kubenswrapper[4821]: I1125 10:33:50.553996 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:50 crc kubenswrapper[4821]: I1125 10:33:50.554016 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:50 crc kubenswrapper[4821]: I1125 10:33:50.554028 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:50Z","lastTransitionTime":"2025-11-25T10:33:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:50 crc kubenswrapper[4821]: I1125 10:33:50.655994 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:50 crc kubenswrapper[4821]: I1125 10:33:50.656040 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:50 crc kubenswrapper[4821]: I1125 10:33:50.656053 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:50 crc kubenswrapper[4821]: I1125 10:33:50.656071 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:50 crc kubenswrapper[4821]: I1125 10:33:50.656083 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:50Z","lastTransitionTime":"2025-11-25T10:33:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:50 crc kubenswrapper[4821]: I1125 10:33:50.758991 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:50 crc kubenswrapper[4821]: I1125 10:33:50.759051 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:50 crc kubenswrapper[4821]: I1125 10:33:50.759078 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:50 crc kubenswrapper[4821]: I1125 10:33:50.759100 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:50 crc kubenswrapper[4821]: I1125 10:33:50.759115 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:50Z","lastTransitionTime":"2025-11-25T10:33:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:50 crc kubenswrapper[4821]: I1125 10:33:50.861671 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:50 crc kubenswrapper[4821]: I1125 10:33:50.861716 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:50 crc kubenswrapper[4821]: I1125 10:33:50.861728 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:50 crc kubenswrapper[4821]: I1125 10:33:50.861744 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:50 crc kubenswrapper[4821]: I1125 10:33:50.861754 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:50Z","lastTransitionTime":"2025-11-25T10:33:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:50 crc kubenswrapper[4821]: I1125 10:33:50.964613 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:50 crc kubenswrapper[4821]: I1125 10:33:50.964663 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:50 crc kubenswrapper[4821]: I1125 10:33:50.964682 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:50 crc kubenswrapper[4821]: I1125 10:33:50.964712 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:50 crc kubenswrapper[4821]: I1125 10:33:50.964736 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:50Z","lastTransitionTime":"2025-11-25T10:33:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:51 crc kubenswrapper[4821]: I1125 10:33:51.067381 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:51 crc kubenswrapper[4821]: I1125 10:33:51.067444 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:51 crc kubenswrapper[4821]: I1125 10:33:51.067461 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:51 crc kubenswrapper[4821]: I1125 10:33:51.067486 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:51 crc kubenswrapper[4821]: I1125 10:33:51.067503 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:51Z","lastTransitionTime":"2025-11-25T10:33:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:51 crc kubenswrapper[4821]: I1125 10:33:51.114220 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:33:51 crc kubenswrapper[4821]: I1125 10:33:51.114264 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:33:51 crc kubenswrapper[4821]: I1125 10:33:51.114451 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:33:51 crc kubenswrapper[4821]: E1125 10:33:51.114453 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:33:51 crc kubenswrapper[4821]: E1125 10:33:51.114602 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:33:51 crc kubenswrapper[4821]: E1125 10:33:51.114785 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:33:51 crc kubenswrapper[4821]: I1125 10:33:51.170918 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:51 crc kubenswrapper[4821]: I1125 10:33:51.170956 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:51 crc kubenswrapper[4821]: I1125 10:33:51.170965 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:51 crc kubenswrapper[4821]: I1125 10:33:51.170979 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:51 crc kubenswrapper[4821]: I1125 10:33:51.170988 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:51Z","lastTransitionTime":"2025-11-25T10:33:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:51 crc kubenswrapper[4821]: I1125 10:33:51.273487 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:51 crc kubenswrapper[4821]: I1125 10:33:51.273533 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:51 crc kubenswrapper[4821]: I1125 10:33:51.273545 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:51 crc kubenswrapper[4821]: I1125 10:33:51.273562 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:51 crc kubenswrapper[4821]: I1125 10:33:51.273608 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:51Z","lastTransitionTime":"2025-11-25T10:33:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:51 crc kubenswrapper[4821]: I1125 10:33:51.375663 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:51 crc kubenswrapper[4821]: I1125 10:33:51.375712 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:51 crc kubenswrapper[4821]: I1125 10:33:51.375726 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:51 crc kubenswrapper[4821]: I1125 10:33:51.375742 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:51 crc kubenswrapper[4821]: I1125 10:33:51.375754 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:51Z","lastTransitionTime":"2025-11-25T10:33:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:51 crc kubenswrapper[4821]: I1125 10:33:51.478318 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:51 crc kubenswrapper[4821]: I1125 10:33:51.478395 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:51 crc kubenswrapper[4821]: I1125 10:33:51.478405 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:51 crc kubenswrapper[4821]: I1125 10:33:51.478418 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:51 crc kubenswrapper[4821]: I1125 10:33:51.478429 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:51Z","lastTransitionTime":"2025-11-25T10:33:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:51 crc kubenswrapper[4821]: I1125 10:33:51.580178 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:51 crc kubenswrapper[4821]: I1125 10:33:51.580260 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:51 crc kubenswrapper[4821]: I1125 10:33:51.580271 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:51 crc kubenswrapper[4821]: I1125 10:33:51.580284 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:51 crc kubenswrapper[4821]: I1125 10:33:51.580293 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:51Z","lastTransitionTime":"2025-11-25T10:33:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:51 crc kubenswrapper[4821]: I1125 10:33:51.682540 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:51 crc kubenswrapper[4821]: I1125 10:33:51.682598 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:51 crc kubenswrapper[4821]: I1125 10:33:51.682613 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:51 crc kubenswrapper[4821]: I1125 10:33:51.682632 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:51 crc kubenswrapper[4821]: I1125 10:33:51.682643 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:51Z","lastTransitionTime":"2025-11-25T10:33:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:51 crc kubenswrapper[4821]: I1125 10:33:51.785782 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:51 crc kubenswrapper[4821]: I1125 10:33:51.785848 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:51 crc kubenswrapper[4821]: I1125 10:33:51.785867 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:51 crc kubenswrapper[4821]: I1125 10:33:51.785894 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:51 crc kubenswrapper[4821]: I1125 10:33:51.785907 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:51Z","lastTransitionTime":"2025-11-25T10:33:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:51 crc kubenswrapper[4821]: I1125 10:33:51.888088 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:51 crc kubenswrapper[4821]: I1125 10:33:51.888129 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:51 crc kubenswrapper[4821]: I1125 10:33:51.888138 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:51 crc kubenswrapper[4821]: I1125 10:33:51.888150 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:51 crc kubenswrapper[4821]: I1125 10:33:51.888191 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:51Z","lastTransitionTime":"2025-11-25T10:33:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:51 crc kubenswrapper[4821]: I1125 10:33:51.990726 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:51 crc kubenswrapper[4821]: I1125 10:33:51.990778 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:51 crc kubenswrapper[4821]: I1125 10:33:51.990788 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:51 crc kubenswrapper[4821]: I1125 10:33:51.990809 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:51 crc kubenswrapper[4821]: I1125 10:33:51.990821 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:51Z","lastTransitionTime":"2025-11-25T10:33:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:52 crc kubenswrapper[4821]: I1125 10:33:52.093452 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:52 crc kubenswrapper[4821]: I1125 10:33:52.093505 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:52 crc kubenswrapper[4821]: I1125 10:33:52.093515 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:52 crc kubenswrapper[4821]: I1125 10:33:52.093532 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:52 crc kubenswrapper[4821]: I1125 10:33:52.093544 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:52Z","lastTransitionTime":"2025-11-25T10:33:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:52 crc kubenswrapper[4821]: I1125 10:33:52.113997 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4msn" Nov 25 10:33:52 crc kubenswrapper[4821]: E1125 10:33:52.114231 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4msn" podUID="1f12cef3-7e40-4f61-836e-23cd0e578e71" Nov 25 10:33:52 crc kubenswrapper[4821]: I1125 10:33:52.195629 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:52 crc kubenswrapper[4821]: I1125 10:33:52.195661 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:52 crc kubenswrapper[4821]: I1125 10:33:52.195671 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:52 crc kubenswrapper[4821]: I1125 10:33:52.195687 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:52 crc kubenswrapper[4821]: I1125 10:33:52.195699 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:52Z","lastTransitionTime":"2025-11-25T10:33:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:52 crc kubenswrapper[4821]: I1125 10:33:52.298444 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:52 crc kubenswrapper[4821]: I1125 10:33:52.298490 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:52 crc kubenswrapper[4821]: I1125 10:33:52.298504 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:52 crc kubenswrapper[4821]: I1125 10:33:52.298519 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:52 crc kubenswrapper[4821]: I1125 10:33:52.298529 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:52Z","lastTransitionTime":"2025-11-25T10:33:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:52 crc kubenswrapper[4821]: I1125 10:33:52.400929 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:52 crc kubenswrapper[4821]: I1125 10:33:52.400980 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:52 crc kubenswrapper[4821]: I1125 10:33:52.400997 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:52 crc kubenswrapper[4821]: I1125 10:33:52.401016 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:52 crc kubenswrapper[4821]: I1125 10:33:52.401037 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:52Z","lastTransitionTime":"2025-11-25T10:33:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:52 crc kubenswrapper[4821]: I1125 10:33:52.503492 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:52 crc kubenswrapper[4821]: I1125 10:33:52.503531 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:52 crc kubenswrapper[4821]: I1125 10:33:52.503542 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:52 crc kubenswrapper[4821]: I1125 10:33:52.503557 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:52 crc kubenswrapper[4821]: I1125 10:33:52.503569 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:52Z","lastTransitionTime":"2025-11-25T10:33:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:52 crc kubenswrapper[4821]: I1125 10:33:52.606087 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:52 crc kubenswrapper[4821]: I1125 10:33:52.606128 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:52 crc kubenswrapper[4821]: I1125 10:33:52.606142 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:52 crc kubenswrapper[4821]: I1125 10:33:52.606176 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:52 crc kubenswrapper[4821]: I1125 10:33:52.606188 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:52Z","lastTransitionTime":"2025-11-25T10:33:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:52 crc kubenswrapper[4821]: I1125 10:33:52.708875 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:52 crc kubenswrapper[4821]: I1125 10:33:52.708958 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:52 crc kubenswrapper[4821]: I1125 10:33:52.708976 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:52 crc kubenswrapper[4821]: I1125 10:33:52.709002 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:52 crc kubenswrapper[4821]: I1125 10:33:52.709023 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:52Z","lastTransitionTime":"2025-11-25T10:33:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:52 crc kubenswrapper[4821]: I1125 10:33:52.811178 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:52 crc kubenswrapper[4821]: I1125 10:33:52.811224 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:52 crc kubenswrapper[4821]: I1125 10:33:52.811284 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:52 crc kubenswrapper[4821]: I1125 10:33:52.811302 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:52 crc kubenswrapper[4821]: I1125 10:33:52.811318 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:52Z","lastTransitionTime":"2025-11-25T10:33:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:52 crc kubenswrapper[4821]: I1125 10:33:52.913177 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:52 crc kubenswrapper[4821]: I1125 10:33:52.913221 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:52 crc kubenswrapper[4821]: I1125 10:33:52.913230 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:52 crc kubenswrapper[4821]: I1125 10:33:52.913245 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:52 crc kubenswrapper[4821]: I1125 10:33:52.913257 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:52Z","lastTransitionTime":"2025-11-25T10:33:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:53 crc kubenswrapper[4821]: I1125 10:33:53.015448 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:53 crc kubenswrapper[4821]: I1125 10:33:53.015485 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:53 crc kubenswrapper[4821]: I1125 10:33:53.015493 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:53 crc kubenswrapper[4821]: I1125 10:33:53.015506 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:53 crc kubenswrapper[4821]: I1125 10:33:53.015516 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:53Z","lastTransitionTime":"2025-11-25T10:33:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:53 crc kubenswrapper[4821]: I1125 10:33:53.113349 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:33:53 crc kubenswrapper[4821]: E1125 10:33:53.113454 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:33:53 crc kubenswrapper[4821]: I1125 10:33:53.113354 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:33:53 crc kubenswrapper[4821]: E1125 10:33:53.113553 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:33:53 crc kubenswrapper[4821]: I1125 10:33:53.113364 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:33:53 crc kubenswrapper[4821]: E1125 10:33:53.113635 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:33:53 crc kubenswrapper[4821]: I1125 10:33:53.118054 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:53 crc kubenswrapper[4821]: I1125 10:33:53.118079 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:53 crc kubenswrapper[4821]: I1125 10:33:53.118089 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:53 crc kubenswrapper[4821]: I1125 10:33:53.118103 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:53 crc kubenswrapper[4821]: I1125 10:33:53.118113 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:53Z","lastTransitionTime":"2025-11-25T10:33:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:53 crc kubenswrapper[4821]: I1125 10:33:53.219766 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:53 crc kubenswrapper[4821]: I1125 10:33:53.219803 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:53 crc kubenswrapper[4821]: I1125 10:33:53.219812 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:53 crc kubenswrapper[4821]: I1125 10:33:53.219834 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:53 crc kubenswrapper[4821]: I1125 10:33:53.219844 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:53Z","lastTransitionTime":"2025-11-25T10:33:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:53 crc kubenswrapper[4821]: I1125 10:33:53.321963 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:53 crc kubenswrapper[4821]: I1125 10:33:53.322028 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:53 crc kubenswrapper[4821]: I1125 10:33:53.322043 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:53 crc kubenswrapper[4821]: I1125 10:33:53.322105 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:53 crc kubenswrapper[4821]: I1125 10:33:53.322122 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:53Z","lastTransitionTime":"2025-11-25T10:33:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:53 crc kubenswrapper[4821]: I1125 10:33:53.425235 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:53 crc kubenswrapper[4821]: I1125 10:33:53.425317 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:53 crc kubenswrapper[4821]: I1125 10:33:53.425343 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:53 crc kubenswrapper[4821]: I1125 10:33:53.425373 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:53 crc kubenswrapper[4821]: I1125 10:33:53.425394 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:53Z","lastTransitionTime":"2025-11-25T10:33:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:53 crc kubenswrapper[4821]: I1125 10:33:53.528881 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:53 crc kubenswrapper[4821]: I1125 10:33:53.528934 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:53 crc kubenswrapper[4821]: I1125 10:33:53.528953 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:53 crc kubenswrapper[4821]: I1125 10:33:53.528976 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:53 crc kubenswrapper[4821]: I1125 10:33:53.528992 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:53Z","lastTransitionTime":"2025-11-25T10:33:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:53 crc kubenswrapper[4821]: I1125 10:33:53.632839 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:53 crc kubenswrapper[4821]: I1125 10:33:53.632904 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:53 crc kubenswrapper[4821]: I1125 10:33:53.632929 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:53 crc kubenswrapper[4821]: I1125 10:33:53.632960 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:53 crc kubenswrapper[4821]: I1125 10:33:53.632981 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:53Z","lastTransitionTime":"2025-11-25T10:33:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:53 crc kubenswrapper[4821]: I1125 10:33:53.735799 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:53 crc kubenswrapper[4821]: I1125 10:33:53.735851 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:53 crc kubenswrapper[4821]: I1125 10:33:53.735866 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:53 crc kubenswrapper[4821]: I1125 10:33:53.735886 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:53 crc kubenswrapper[4821]: I1125 10:33:53.735900 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:53Z","lastTransitionTime":"2025-11-25T10:33:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:53 crc kubenswrapper[4821]: I1125 10:33:53.843479 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:53 crc kubenswrapper[4821]: I1125 10:33:53.843522 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:53 crc kubenswrapper[4821]: I1125 10:33:53.843532 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:53 crc kubenswrapper[4821]: I1125 10:33:53.843588 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:53 crc kubenswrapper[4821]: I1125 10:33:53.843599 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:53Z","lastTransitionTime":"2025-11-25T10:33:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:53 crc kubenswrapper[4821]: I1125 10:33:53.946002 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:53 crc kubenswrapper[4821]: I1125 10:33:53.946049 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:53 crc kubenswrapper[4821]: I1125 10:33:53.946058 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:53 crc kubenswrapper[4821]: I1125 10:33:53.946073 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:53 crc kubenswrapper[4821]: I1125 10:33:53.946084 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:53Z","lastTransitionTime":"2025-11-25T10:33:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:54 crc kubenswrapper[4821]: I1125 10:33:54.049276 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:54 crc kubenswrapper[4821]: I1125 10:33:54.049348 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:54 crc kubenswrapper[4821]: I1125 10:33:54.049366 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:54 crc kubenswrapper[4821]: I1125 10:33:54.049390 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:54 crc kubenswrapper[4821]: I1125 10:33:54.049411 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:54Z","lastTransitionTime":"2025-11-25T10:33:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:54 crc kubenswrapper[4821]: I1125 10:33:54.114023 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4msn" Nov 25 10:33:54 crc kubenswrapper[4821]: E1125 10:33:54.114820 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4msn" podUID="1f12cef3-7e40-4f61-836e-23cd0e578e71" Nov 25 10:33:54 crc kubenswrapper[4821]: I1125 10:33:54.115407 4821 scope.go:117] "RemoveContainer" containerID="90d8792a79993c04a0eb774498395d955deff5c42c693892426685d157cf9cff" Nov 25 10:33:54 crc kubenswrapper[4821]: E1125 10:33:54.115753 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-mzhnw_openshift-ovn-kubernetes(954085fa-d03a-4e88-9dfd-566257e3558d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" podUID="954085fa-d03a-4e88-9dfd-566257e3558d" Nov 25 10:33:54 crc kubenswrapper[4821]: I1125 10:33:54.151923 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:54 crc kubenswrapper[4821]: I1125 10:33:54.151974 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:54 crc kubenswrapper[4821]: I1125 10:33:54.151988 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:54 crc kubenswrapper[4821]: I1125 10:33:54.152006 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:54 crc kubenswrapper[4821]: I1125 10:33:54.152019 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:54Z","lastTransitionTime":"2025-11-25T10:33:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:54 crc kubenswrapper[4821]: I1125 10:33:54.255144 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:54 crc kubenswrapper[4821]: I1125 10:33:54.255214 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:54 crc kubenswrapper[4821]: I1125 10:33:54.255263 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:54 crc kubenswrapper[4821]: I1125 10:33:54.255285 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:54 crc kubenswrapper[4821]: I1125 10:33:54.255296 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:54Z","lastTransitionTime":"2025-11-25T10:33:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:54 crc kubenswrapper[4821]: I1125 10:33:54.357284 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:54 crc kubenswrapper[4821]: I1125 10:33:54.357327 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:54 crc kubenswrapper[4821]: I1125 10:33:54.357338 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:54 crc kubenswrapper[4821]: I1125 10:33:54.357352 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:54 crc kubenswrapper[4821]: I1125 10:33:54.357362 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:54Z","lastTransitionTime":"2025-11-25T10:33:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:54 crc kubenswrapper[4821]: I1125 10:33:54.459985 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:54 crc kubenswrapper[4821]: I1125 10:33:54.460035 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:54 crc kubenswrapper[4821]: I1125 10:33:54.460046 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:54 crc kubenswrapper[4821]: I1125 10:33:54.460062 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:54 crc kubenswrapper[4821]: I1125 10:33:54.460073 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:54Z","lastTransitionTime":"2025-11-25T10:33:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:54 crc kubenswrapper[4821]: I1125 10:33:54.562394 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:54 crc kubenswrapper[4821]: I1125 10:33:54.562430 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:54 crc kubenswrapper[4821]: I1125 10:33:54.562439 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:54 crc kubenswrapper[4821]: I1125 10:33:54.562453 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:54 crc kubenswrapper[4821]: I1125 10:33:54.562463 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:54Z","lastTransitionTime":"2025-11-25T10:33:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:54 crc kubenswrapper[4821]: I1125 10:33:54.665239 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:54 crc kubenswrapper[4821]: I1125 10:33:54.665292 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:54 crc kubenswrapper[4821]: I1125 10:33:54.665302 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:54 crc kubenswrapper[4821]: I1125 10:33:54.665316 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:54 crc kubenswrapper[4821]: I1125 10:33:54.665326 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:54Z","lastTransitionTime":"2025-11-25T10:33:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:54 crc kubenswrapper[4821]: I1125 10:33:54.768038 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:54 crc kubenswrapper[4821]: I1125 10:33:54.768080 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:54 crc kubenswrapper[4821]: I1125 10:33:54.768092 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:54 crc kubenswrapper[4821]: I1125 10:33:54.768111 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:54 crc kubenswrapper[4821]: I1125 10:33:54.768123 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:54Z","lastTransitionTime":"2025-11-25T10:33:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:54 crc kubenswrapper[4821]: I1125 10:33:54.871316 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:54 crc kubenswrapper[4821]: I1125 10:33:54.871370 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:54 crc kubenswrapper[4821]: I1125 10:33:54.871382 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:54 crc kubenswrapper[4821]: I1125 10:33:54.871398 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:54 crc kubenswrapper[4821]: I1125 10:33:54.871806 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:54Z","lastTransitionTime":"2025-11-25T10:33:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:54 crc kubenswrapper[4821]: I1125 10:33:54.974876 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:54 crc kubenswrapper[4821]: I1125 10:33:54.974918 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:54 crc kubenswrapper[4821]: I1125 10:33:54.974927 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:54 crc kubenswrapper[4821]: I1125 10:33:54.974941 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:54 crc kubenswrapper[4821]: I1125 10:33:54.974950 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:54Z","lastTransitionTime":"2025-11-25T10:33:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:55 crc kubenswrapper[4821]: I1125 10:33:55.078326 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:55 crc kubenswrapper[4821]: I1125 10:33:55.078378 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:55 crc kubenswrapper[4821]: I1125 10:33:55.078390 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:55 crc kubenswrapper[4821]: I1125 10:33:55.078407 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:55 crc kubenswrapper[4821]: I1125 10:33:55.078418 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:55Z","lastTransitionTime":"2025-11-25T10:33:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:55 crc kubenswrapper[4821]: I1125 10:33:55.113922 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:33:55 crc kubenswrapper[4821]: E1125 10:33:55.114080 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:33:55 crc kubenswrapper[4821]: I1125 10:33:55.114122 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:33:55 crc kubenswrapper[4821]: I1125 10:33:55.114239 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:33:55 crc kubenswrapper[4821]: E1125 10:33:55.114321 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:33:55 crc kubenswrapper[4821]: E1125 10:33:55.114489 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:33:55 crc kubenswrapper[4821]: I1125 10:33:55.182294 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:55 crc kubenswrapper[4821]: I1125 10:33:55.182359 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:55 crc kubenswrapper[4821]: I1125 10:33:55.182376 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:55 crc kubenswrapper[4821]: I1125 10:33:55.182404 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:55 crc kubenswrapper[4821]: I1125 10:33:55.182424 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:55Z","lastTransitionTime":"2025-11-25T10:33:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:55 crc kubenswrapper[4821]: I1125 10:33:55.284717 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:55 crc kubenswrapper[4821]: I1125 10:33:55.284751 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:55 crc kubenswrapper[4821]: I1125 10:33:55.284759 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:55 crc kubenswrapper[4821]: I1125 10:33:55.284771 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:55 crc kubenswrapper[4821]: I1125 10:33:55.284780 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:55Z","lastTransitionTime":"2025-11-25T10:33:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:55 crc kubenswrapper[4821]: I1125 10:33:55.387534 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:55 crc kubenswrapper[4821]: I1125 10:33:55.387635 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:55 crc kubenswrapper[4821]: I1125 10:33:55.387650 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:55 crc kubenswrapper[4821]: I1125 10:33:55.387673 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:55 crc kubenswrapper[4821]: I1125 10:33:55.387688 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:55Z","lastTransitionTime":"2025-11-25T10:33:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:55 crc kubenswrapper[4821]: I1125 10:33:55.490427 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:55 crc kubenswrapper[4821]: I1125 10:33:55.490482 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:55 crc kubenswrapper[4821]: I1125 10:33:55.490495 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:55 crc kubenswrapper[4821]: I1125 10:33:55.490516 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:55 crc kubenswrapper[4821]: I1125 10:33:55.490532 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:55Z","lastTransitionTime":"2025-11-25T10:33:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:55 crc kubenswrapper[4821]: I1125 10:33:55.593033 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:55 crc kubenswrapper[4821]: I1125 10:33:55.593125 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:55 crc kubenswrapper[4821]: I1125 10:33:55.593206 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:55 crc kubenswrapper[4821]: I1125 10:33:55.593225 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:55 crc kubenswrapper[4821]: I1125 10:33:55.593235 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:55Z","lastTransitionTime":"2025-11-25T10:33:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:55 crc kubenswrapper[4821]: I1125 10:33:55.695948 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:55 crc kubenswrapper[4821]: I1125 10:33:55.695990 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:55 crc kubenswrapper[4821]: I1125 10:33:55.696000 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:55 crc kubenswrapper[4821]: I1125 10:33:55.696015 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:55 crc kubenswrapper[4821]: I1125 10:33:55.696026 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:55Z","lastTransitionTime":"2025-11-25T10:33:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:55 crc kubenswrapper[4821]: I1125 10:33:55.798355 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:55 crc kubenswrapper[4821]: I1125 10:33:55.798384 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:55 crc kubenswrapper[4821]: I1125 10:33:55.798394 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:55 crc kubenswrapper[4821]: I1125 10:33:55.798408 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:55 crc kubenswrapper[4821]: I1125 10:33:55.798416 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:55Z","lastTransitionTime":"2025-11-25T10:33:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:55 crc kubenswrapper[4821]: I1125 10:33:55.900999 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:55 crc kubenswrapper[4821]: I1125 10:33:55.901045 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:55 crc kubenswrapper[4821]: I1125 10:33:55.901056 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:55 crc kubenswrapper[4821]: I1125 10:33:55.901073 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:55 crc kubenswrapper[4821]: I1125 10:33:55.901088 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:55Z","lastTransitionTime":"2025-11-25T10:33:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:56 crc kubenswrapper[4821]: I1125 10:33:56.002988 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:56 crc kubenswrapper[4821]: I1125 10:33:56.003030 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:56 crc kubenswrapper[4821]: I1125 10:33:56.003045 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:56 crc kubenswrapper[4821]: I1125 10:33:56.003066 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:56 crc kubenswrapper[4821]: I1125 10:33:56.003077 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:56Z","lastTransitionTime":"2025-11-25T10:33:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:56 crc kubenswrapper[4821]: I1125 10:33:56.105175 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:56 crc kubenswrapper[4821]: I1125 10:33:56.105282 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:56 crc kubenswrapper[4821]: I1125 10:33:56.105294 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:56 crc kubenswrapper[4821]: I1125 10:33:56.105311 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:56 crc kubenswrapper[4821]: I1125 10:33:56.105319 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:56Z","lastTransitionTime":"2025-11-25T10:33:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:56 crc kubenswrapper[4821]: I1125 10:33:56.113443 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4msn" Nov 25 10:33:56 crc kubenswrapper[4821]: E1125 10:33:56.113533 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4msn" podUID="1f12cef3-7e40-4f61-836e-23cd0e578e71" Nov 25 10:33:56 crc kubenswrapper[4821]: I1125 10:33:56.208431 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:56 crc kubenswrapper[4821]: I1125 10:33:56.208500 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:56 crc kubenswrapper[4821]: I1125 10:33:56.208515 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:56 crc kubenswrapper[4821]: I1125 10:33:56.208536 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:56 crc kubenswrapper[4821]: I1125 10:33:56.208552 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:56Z","lastTransitionTime":"2025-11-25T10:33:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:56 crc kubenswrapper[4821]: I1125 10:33:56.310904 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:56 crc kubenswrapper[4821]: I1125 10:33:56.310956 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:56 crc kubenswrapper[4821]: I1125 10:33:56.310971 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:56 crc kubenswrapper[4821]: I1125 10:33:56.310991 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:56 crc kubenswrapper[4821]: I1125 10:33:56.311004 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:56Z","lastTransitionTime":"2025-11-25T10:33:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:56 crc kubenswrapper[4821]: I1125 10:33:56.414212 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:56 crc kubenswrapper[4821]: I1125 10:33:56.414309 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:56 crc kubenswrapper[4821]: I1125 10:33:56.414322 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:56 crc kubenswrapper[4821]: I1125 10:33:56.414341 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:56 crc kubenswrapper[4821]: I1125 10:33:56.414352 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:56Z","lastTransitionTime":"2025-11-25T10:33:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:56 crc kubenswrapper[4821]: I1125 10:33:56.516821 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:56 crc kubenswrapper[4821]: I1125 10:33:56.516862 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:56 crc kubenswrapper[4821]: I1125 10:33:56.516870 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:56 crc kubenswrapper[4821]: I1125 10:33:56.516886 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:56 crc kubenswrapper[4821]: I1125 10:33:56.516895 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:56Z","lastTransitionTime":"2025-11-25T10:33:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:56 crc kubenswrapper[4821]: I1125 10:33:56.619436 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:56 crc kubenswrapper[4821]: I1125 10:33:56.619491 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:56 crc kubenswrapper[4821]: I1125 10:33:56.619502 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:56 crc kubenswrapper[4821]: I1125 10:33:56.619517 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:56 crc kubenswrapper[4821]: I1125 10:33:56.619528 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:56Z","lastTransitionTime":"2025-11-25T10:33:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:56 crc kubenswrapper[4821]: I1125 10:33:56.721631 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:56 crc kubenswrapper[4821]: I1125 10:33:56.721697 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:56 crc kubenswrapper[4821]: I1125 10:33:56.721707 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:56 crc kubenswrapper[4821]: I1125 10:33:56.721743 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:56 crc kubenswrapper[4821]: I1125 10:33:56.721757 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:56Z","lastTransitionTime":"2025-11-25T10:33:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:56 crc kubenswrapper[4821]: I1125 10:33:56.824142 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:56 crc kubenswrapper[4821]: I1125 10:33:56.824215 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:56 crc kubenswrapper[4821]: I1125 10:33:56.824227 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:56 crc kubenswrapper[4821]: I1125 10:33:56.824244 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:56 crc kubenswrapper[4821]: I1125 10:33:56.824258 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:56Z","lastTransitionTime":"2025-11-25T10:33:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:56 crc kubenswrapper[4821]: I1125 10:33:56.926239 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:56 crc kubenswrapper[4821]: I1125 10:33:56.926285 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:56 crc kubenswrapper[4821]: I1125 10:33:56.926297 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:56 crc kubenswrapper[4821]: I1125 10:33:56.926317 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:56 crc kubenswrapper[4821]: I1125 10:33:56.926331 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:56Z","lastTransitionTime":"2025-11-25T10:33:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:57 crc kubenswrapper[4821]: I1125 10:33:57.028747 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:57 crc kubenswrapper[4821]: I1125 10:33:57.028795 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:57 crc kubenswrapper[4821]: I1125 10:33:57.028803 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:57 crc kubenswrapper[4821]: I1125 10:33:57.028818 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:57 crc kubenswrapper[4821]: I1125 10:33:57.028828 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:57Z","lastTransitionTime":"2025-11-25T10:33:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:57 crc kubenswrapper[4821]: I1125 10:33:57.113766 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:33:57 crc kubenswrapper[4821]: I1125 10:33:57.113871 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:33:57 crc kubenswrapper[4821]: E1125 10:33:57.113913 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:33:57 crc kubenswrapper[4821]: I1125 10:33:57.113933 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:33:57 crc kubenswrapper[4821]: E1125 10:33:57.114069 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:33:57 crc kubenswrapper[4821]: E1125 10:33:57.114143 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:33:57 crc kubenswrapper[4821]: I1125 10:33:57.131409 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:57 crc kubenswrapper[4821]: I1125 10:33:57.131451 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:57 crc kubenswrapper[4821]: I1125 10:33:57.131459 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:57 crc kubenswrapper[4821]: I1125 10:33:57.131476 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:57 crc kubenswrapper[4821]: I1125 10:33:57.131487 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:57Z","lastTransitionTime":"2025-11-25T10:33:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:57 crc kubenswrapper[4821]: I1125 10:33:57.234258 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:57 crc kubenswrapper[4821]: I1125 10:33:57.234298 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:57 crc kubenswrapper[4821]: I1125 10:33:57.234311 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:57 crc kubenswrapper[4821]: I1125 10:33:57.234329 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:57 crc kubenswrapper[4821]: I1125 10:33:57.234342 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:57Z","lastTransitionTime":"2025-11-25T10:33:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:57 crc kubenswrapper[4821]: I1125 10:33:57.337111 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:57 crc kubenswrapper[4821]: I1125 10:33:57.337154 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:57 crc kubenswrapper[4821]: I1125 10:33:57.337177 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:57 crc kubenswrapper[4821]: I1125 10:33:57.337193 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:57 crc kubenswrapper[4821]: I1125 10:33:57.337203 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:57Z","lastTransitionTime":"2025-11-25T10:33:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:57 crc kubenswrapper[4821]: I1125 10:33:57.440207 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:57 crc kubenswrapper[4821]: I1125 10:33:57.440249 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:57 crc kubenswrapper[4821]: I1125 10:33:57.440256 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:57 crc kubenswrapper[4821]: I1125 10:33:57.440270 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:57 crc kubenswrapper[4821]: I1125 10:33:57.440281 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:57Z","lastTransitionTime":"2025-11-25T10:33:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:57 crc kubenswrapper[4821]: I1125 10:33:57.541951 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:57 crc kubenswrapper[4821]: I1125 10:33:57.541989 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:57 crc kubenswrapper[4821]: I1125 10:33:57.541997 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:57 crc kubenswrapper[4821]: I1125 10:33:57.542010 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:57 crc kubenswrapper[4821]: I1125 10:33:57.542020 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:57Z","lastTransitionTime":"2025-11-25T10:33:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:57 crc kubenswrapper[4821]: I1125 10:33:57.643695 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:57 crc kubenswrapper[4821]: I1125 10:33:57.643726 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:57 crc kubenswrapper[4821]: I1125 10:33:57.643734 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:57 crc kubenswrapper[4821]: I1125 10:33:57.643746 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:57 crc kubenswrapper[4821]: I1125 10:33:57.643755 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:57Z","lastTransitionTime":"2025-11-25T10:33:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:57 crc kubenswrapper[4821]: I1125 10:33:57.745744 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:57 crc kubenswrapper[4821]: I1125 10:33:57.745809 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:57 crc kubenswrapper[4821]: I1125 10:33:57.745827 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:57 crc kubenswrapper[4821]: I1125 10:33:57.745854 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:57 crc kubenswrapper[4821]: I1125 10:33:57.745871 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:57Z","lastTransitionTime":"2025-11-25T10:33:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:57 crc kubenswrapper[4821]: I1125 10:33:57.755136 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 10:33:57 crc kubenswrapper[4821]: I1125 10:33:57.755274 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 10:33:57 crc kubenswrapper[4821]: I1125 10:33:57.755289 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 10:33:57 crc kubenswrapper[4821]: I1125 10:33:57.755307 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 10:33:57 crc kubenswrapper[4821]: I1125 10:33:57.755317 4821 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T10:33:57Z","lastTransitionTime":"2025-11-25T10:33:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 10:33:57 crc kubenswrapper[4821]: I1125 10:33:57.798382 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-crprl"] Nov 25 10:33:57 crc kubenswrapper[4821]: I1125 10:33:57.798745 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-crprl" Nov 25 10:33:57 crc kubenswrapper[4821]: I1125 10:33:57.801283 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Nov 25 10:33:57 crc kubenswrapper[4821]: I1125 10:33:57.801304 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Nov 25 10:33:57 crc kubenswrapper[4821]: I1125 10:33:57.801511 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Nov 25 10:33:57 crc kubenswrapper[4821]: I1125 10:33:57.802737 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Nov 25 10:33:57 crc kubenswrapper[4821]: I1125 10:33:57.844907 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podStartSLOduration=87.844887804 podStartE2EDuration="1m27.844887804s" podCreationTimestamp="2025-11-25 10:32:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:33:57.825158554 +0000 UTC m=+108.361478431" watchObservedRunningTime="2025-11-25 10:33:57.844887804 +0000 UTC m=+108.381207651" Nov 25 10:33:57 crc kubenswrapper[4821]: I1125 10:33:57.845057 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-qppnw" podStartSLOduration=87.845052958 podStartE2EDuration="1m27.845052958s" podCreationTimestamp="2025-11-25 10:32:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:33:57.844563666 +0000 UTC m=+108.380883513" watchObservedRunningTime="2025-11-25 10:33:57.845052958 +0000 UTC m=+108.381372805" Nov 25 10:33:57 crc kubenswrapper[4821]: I1125 10:33:57.867430 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-tgdlt" podStartSLOduration=87.867412448 podStartE2EDuration="1m27.867412448s" podCreationTimestamp="2025-11-25 10:32:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:33:57.857314517 +0000 UTC m=+108.393634354" watchObservedRunningTime="2025-11-25 10:33:57.867412448 +0000 UTC m=+108.403732295" Nov 25 10:33:57 crc kubenswrapper[4821]: I1125 10:33:57.883207 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=89.883154006 podStartE2EDuration="1m29.883154006s" podCreationTimestamp="2025-11-25 10:32:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:33:57.882429459 +0000 UTC m=+108.418749306" watchObservedRunningTime="2025-11-25 10:33:57.883154006 +0000 UTC m=+108.419473853" Nov 25 10:33:57 crc kubenswrapper[4821]: I1125 10:33:57.894409 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=60.894393573 podStartE2EDuration="1m0.894393573s" podCreationTimestamp="2025-11-25 10:32:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:33:57.893605324 +0000 UTC m=+108.429925191" watchObservedRunningTime="2025-11-25 10:33:57.894393573 +0000 UTC m=+108.430713410" Nov 25 10:33:57 crc kubenswrapper[4821]: I1125 10:33:57.959716 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/319b4d35-0b80-4efc-9500-1f01bd0e20ae-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-crprl\" (UID: \"319b4d35-0b80-4efc-9500-1f01bd0e20ae\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-crprl" Nov 25 10:33:57 crc kubenswrapper[4821]: I1125 10:33:57.959750 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/319b4d35-0b80-4efc-9500-1f01bd0e20ae-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-crprl\" (UID: \"319b4d35-0b80-4efc-9500-1f01bd0e20ae\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-crprl" Nov 25 10:33:57 crc kubenswrapper[4821]: I1125 10:33:57.959772 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/319b4d35-0b80-4efc-9500-1f01bd0e20ae-service-ca\") pod \"cluster-version-operator-5c965bbfc6-crprl\" (UID: \"319b4d35-0b80-4efc-9500-1f01bd0e20ae\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-crprl" Nov 25 10:33:57 crc kubenswrapper[4821]: I1125 10:33:57.959812 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/319b4d35-0b80-4efc-9500-1f01bd0e20ae-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-crprl\" (UID: \"319b4d35-0b80-4efc-9500-1f01bd0e20ae\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-crprl" Nov 25 10:33:57 crc kubenswrapper[4821]: I1125 10:33:57.959827 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/319b4d35-0b80-4efc-9500-1f01bd0e20ae-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-crprl\" (UID: \"319b4d35-0b80-4efc-9500-1f01bd0e20ae\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-crprl" Nov 25 10:33:58 crc kubenswrapper[4821]: I1125 10:33:58.061272 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/319b4d35-0b80-4efc-9500-1f01bd0e20ae-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-crprl\" (UID: \"319b4d35-0b80-4efc-9500-1f01bd0e20ae\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-crprl" Nov 25 10:33:58 crc kubenswrapper[4821]: I1125 10:33:58.061304 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/319b4d35-0b80-4efc-9500-1f01bd0e20ae-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-crprl\" (UID: \"319b4d35-0b80-4efc-9500-1f01bd0e20ae\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-crprl" Nov 25 10:33:58 crc kubenswrapper[4821]: I1125 10:33:58.061324 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/319b4d35-0b80-4efc-9500-1f01bd0e20ae-service-ca\") pod \"cluster-version-operator-5c965bbfc6-crprl\" (UID: \"319b4d35-0b80-4efc-9500-1f01bd0e20ae\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-crprl" Nov 25 10:33:58 crc kubenswrapper[4821]: I1125 10:33:58.061386 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/319b4d35-0b80-4efc-9500-1f01bd0e20ae-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-crprl\" (UID: \"319b4d35-0b80-4efc-9500-1f01bd0e20ae\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-crprl" Nov 25 10:33:58 crc kubenswrapper[4821]: I1125 10:33:58.061410 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/319b4d35-0b80-4efc-9500-1f01bd0e20ae-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-crprl\" (UID: \"319b4d35-0b80-4efc-9500-1f01bd0e20ae\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-crprl" Nov 25 10:33:58 crc kubenswrapper[4821]: I1125 10:33:58.061717 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/319b4d35-0b80-4efc-9500-1f01bd0e20ae-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-crprl\" (UID: \"319b4d35-0b80-4efc-9500-1f01bd0e20ae\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-crprl" Nov 25 10:33:58 crc kubenswrapper[4821]: I1125 10:33:58.062282 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/319b4d35-0b80-4efc-9500-1f01bd0e20ae-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-crprl\" (UID: \"319b4d35-0b80-4efc-9500-1f01bd0e20ae\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-crprl" Nov 25 10:33:58 crc kubenswrapper[4821]: I1125 10:33:58.062773 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/319b4d35-0b80-4efc-9500-1f01bd0e20ae-service-ca\") pod \"cluster-version-operator-5c965bbfc6-crprl\" (UID: \"319b4d35-0b80-4efc-9500-1f01bd0e20ae\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-crprl" Nov 25 10:33:58 crc kubenswrapper[4821]: I1125 10:33:58.066704 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/319b4d35-0b80-4efc-9500-1f01bd0e20ae-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-crprl\" (UID: \"319b4d35-0b80-4efc-9500-1f01bd0e20ae\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-crprl" Nov 25 10:33:58 crc kubenswrapper[4821]: I1125 10:33:58.077258 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/319b4d35-0b80-4efc-9500-1f01bd0e20ae-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-crprl\" (UID: \"319b4d35-0b80-4efc-9500-1f01bd0e20ae\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-crprl" Nov 25 10:33:58 crc kubenswrapper[4821]: I1125 10:33:58.113833 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4msn" Nov 25 10:33:58 crc kubenswrapper[4821]: E1125 10:33:58.114104 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4msn" podUID="1f12cef3-7e40-4f61-836e-23cd0e578e71" Nov 25 10:33:58 crc kubenswrapper[4821]: I1125 10:33:58.114749 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-crprl" Nov 25 10:33:58 crc kubenswrapper[4821]: I1125 10:33:58.555153 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-crprl" event={"ID":"319b4d35-0b80-4efc-9500-1f01bd0e20ae","Type":"ContainerStarted","Data":"8c44a1aaaf5057b255e1379e7f0264db84d0327b1098d165596f4a36ab5100e8"} Nov 25 10:33:58 crc kubenswrapper[4821]: I1125 10:33:58.555821 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-crprl" event={"ID":"319b4d35-0b80-4efc-9500-1f01bd0e20ae","Type":"ContainerStarted","Data":"cf3f4c5e8caecc0bc82a8a635eacc5a39e9052972eac97fe1a1f5b74bdb74e36"} Nov 25 10:33:59 crc kubenswrapper[4821]: I1125 10:33:59.113317 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:33:59 crc kubenswrapper[4821]: I1125 10:33:59.113358 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:33:59 crc kubenswrapper[4821]: E1125 10:33:59.113439 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:33:59 crc kubenswrapper[4821]: I1125 10:33:59.113518 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:33:59 crc kubenswrapper[4821]: E1125 10:33:59.113674 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:33:59 crc kubenswrapper[4821]: E1125 10:33:59.113859 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:34:00 crc kubenswrapper[4821]: I1125 10:34:00.113381 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4msn" Nov 25 10:34:00 crc kubenswrapper[4821]: E1125 10:34:00.114433 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4msn" podUID="1f12cef3-7e40-4f61-836e-23cd0e578e71" Nov 25 10:34:01 crc kubenswrapper[4821]: I1125 10:34:01.113971 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:34:01 crc kubenswrapper[4821]: I1125 10:34:01.113971 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:34:01 crc kubenswrapper[4821]: I1125 10:34:01.114004 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:34:01 crc kubenswrapper[4821]: E1125 10:34:01.114233 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:34:01 crc kubenswrapper[4821]: E1125 10:34:01.114342 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:34:01 crc kubenswrapper[4821]: E1125 10:34:01.114557 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:34:02 crc kubenswrapper[4821]: I1125 10:34:02.113734 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4msn" Nov 25 10:34:02 crc kubenswrapper[4821]: E1125 10:34:02.113840 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4msn" podUID="1f12cef3-7e40-4f61-836e-23cd0e578e71" Nov 25 10:34:03 crc kubenswrapper[4821]: I1125 10:34:03.113465 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:34:03 crc kubenswrapper[4821]: I1125 10:34:03.113656 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:34:03 crc kubenswrapper[4821]: E1125 10:34:03.114343 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:34:03 crc kubenswrapper[4821]: E1125 10:34:03.114346 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:34:03 crc kubenswrapper[4821]: I1125 10:34:03.113716 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:34:03 crc kubenswrapper[4821]: E1125 10:34:03.114701 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:34:04 crc kubenswrapper[4821]: I1125 10:34:04.113927 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4msn" Nov 25 10:34:04 crc kubenswrapper[4821]: E1125 10:34:04.114113 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4msn" podUID="1f12cef3-7e40-4f61-836e-23cd0e578e71" Nov 25 10:34:05 crc kubenswrapper[4821]: I1125 10:34:05.113651 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:34:05 crc kubenswrapper[4821]: I1125 10:34:05.113663 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:34:05 crc kubenswrapper[4821]: E1125 10:34:05.113809 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:34:05 crc kubenswrapper[4821]: E1125 10:34:05.113888 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:34:05 crc kubenswrapper[4821]: I1125 10:34:05.113676 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:34:05 crc kubenswrapper[4821]: E1125 10:34:05.114079 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:34:06 crc kubenswrapper[4821]: I1125 10:34:06.113729 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4msn" Nov 25 10:34:06 crc kubenswrapper[4821]: E1125 10:34:06.113864 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4msn" podUID="1f12cef3-7e40-4f61-836e-23cd0e578e71" Nov 25 10:34:06 crc kubenswrapper[4821]: I1125 10:34:06.577843 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-ldzbz_b84e32ae-4897-4644-a656-6bd82012cc81/kube-multus/1.log" Nov 25 10:34:06 crc kubenswrapper[4821]: I1125 10:34:06.578205 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-ldzbz_b84e32ae-4897-4644-a656-6bd82012cc81/kube-multus/0.log" Nov 25 10:34:06 crc kubenswrapper[4821]: I1125 10:34:06.578247 4821 generic.go:334] "Generic (PLEG): container finished" podID="b84e32ae-4897-4644-a656-6bd82012cc81" containerID="418dc8577fdc98324677dbccfcf9317b67cab4408432b30df2818b22f88d426e" exitCode=1 Nov 25 10:34:06 crc kubenswrapper[4821]: I1125 10:34:06.578278 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-ldzbz" event={"ID":"b84e32ae-4897-4644-a656-6bd82012cc81","Type":"ContainerDied","Data":"418dc8577fdc98324677dbccfcf9317b67cab4408432b30df2818b22f88d426e"} Nov 25 10:34:06 crc kubenswrapper[4821]: I1125 10:34:06.578310 4821 scope.go:117] "RemoveContainer" containerID="8ef433744761f26e8e1f6c0db123821a11923ffd774830744285cd4178d89533" Nov 25 10:34:06 crc kubenswrapper[4821]: I1125 10:34:06.578668 4821 scope.go:117] "RemoveContainer" containerID="418dc8577fdc98324677dbccfcf9317b67cab4408432b30df2818b22f88d426e" Nov 25 10:34:06 crc kubenswrapper[4821]: E1125 10:34:06.578886 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-ldzbz_openshift-multus(b84e32ae-4897-4644-a656-6bd82012cc81)\"" pod="openshift-multus/multus-ldzbz" podUID="b84e32ae-4897-4644-a656-6bd82012cc81" Nov 25 10:34:06 crc kubenswrapper[4821]: I1125 10:34:06.594244 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-crprl" podStartSLOduration=96.594223549 podStartE2EDuration="1m36.594223549s" podCreationTimestamp="2025-11-25 10:32:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:33:58.569715561 +0000 UTC m=+109.106035468" watchObservedRunningTime="2025-11-25 10:34:06.594223549 +0000 UTC m=+117.130543396" Nov 25 10:34:07 crc kubenswrapper[4821]: I1125 10:34:07.114058 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:34:07 crc kubenswrapper[4821]: I1125 10:34:07.114088 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:34:07 crc kubenswrapper[4821]: I1125 10:34:07.114103 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:34:07 crc kubenswrapper[4821]: E1125 10:34:07.114357 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:34:07 crc kubenswrapper[4821]: E1125 10:34:07.114437 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:34:07 crc kubenswrapper[4821]: E1125 10:34:07.114507 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:34:07 crc kubenswrapper[4821]: I1125 10:34:07.582109 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-ldzbz_b84e32ae-4897-4644-a656-6bd82012cc81/kube-multus/1.log" Nov 25 10:34:08 crc kubenswrapper[4821]: I1125 10:34:08.113950 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4msn" Nov 25 10:34:08 crc kubenswrapper[4821]: E1125 10:34:08.114252 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4msn" podUID="1f12cef3-7e40-4f61-836e-23cd0e578e71" Nov 25 10:34:09 crc kubenswrapper[4821]: I1125 10:34:09.113808 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:34:09 crc kubenswrapper[4821]: I1125 10:34:09.113846 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:34:09 crc kubenswrapper[4821]: E1125 10:34:09.113922 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:34:09 crc kubenswrapper[4821]: I1125 10:34:09.113824 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:34:09 crc kubenswrapper[4821]: E1125 10:34:09.113996 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:34:09 crc kubenswrapper[4821]: E1125 10:34:09.114150 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:34:09 crc kubenswrapper[4821]: I1125 10:34:09.115042 4821 scope.go:117] "RemoveContainer" containerID="90d8792a79993c04a0eb774498395d955deff5c42c693892426685d157cf9cff" Nov 25 10:34:09 crc kubenswrapper[4821]: I1125 10:34:09.588934 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mzhnw_954085fa-d03a-4e88-9dfd-566257e3558d/ovnkube-controller/3.log" Nov 25 10:34:09 crc kubenswrapper[4821]: I1125 10:34:09.591719 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" event={"ID":"954085fa-d03a-4e88-9dfd-566257e3558d","Type":"ContainerStarted","Data":"32de39e5841057630116280f4a9e5e5728b786db001c2c0db0d57e7475a3a16d"} Nov 25 10:34:09 crc kubenswrapper[4821]: I1125 10:34:09.592055 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" Nov 25 10:34:09 crc kubenswrapper[4821]: I1125 10:34:09.622491 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" podStartSLOduration=99.622473795 podStartE2EDuration="1m39.622473795s" podCreationTimestamp="2025-11-25 10:32:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:34:09.622371203 +0000 UTC m=+120.158691050" watchObservedRunningTime="2025-11-25 10:34:09.622473795 +0000 UTC m=+120.158793642" Nov 25 10:34:09 crc kubenswrapper[4821]: I1125 10:34:09.749656 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-d4msn"] Nov 25 10:34:09 crc kubenswrapper[4821]: I1125 10:34:09.749788 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4msn" Nov 25 10:34:09 crc kubenswrapper[4821]: E1125 10:34:09.749896 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4msn" podUID="1f12cef3-7e40-4f61-836e-23cd0e578e71" Nov 25 10:34:10 crc kubenswrapper[4821]: E1125 10:34:10.157222 4821 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Nov 25 10:34:10 crc kubenswrapper[4821]: E1125 10:34:10.207376 4821 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 25 10:34:11 crc kubenswrapper[4821]: I1125 10:34:11.114133 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:34:11 crc kubenswrapper[4821]: I1125 10:34:11.114245 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:34:11 crc kubenswrapper[4821]: E1125 10:34:11.114293 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:34:11 crc kubenswrapper[4821]: I1125 10:34:11.114324 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:34:11 crc kubenswrapper[4821]: E1125 10:34:11.114449 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:34:11 crc kubenswrapper[4821]: E1125 10:34:11.114662 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:34:12 crc kubenswrapper[4821]: I1125 10:34:12.114022 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4msn" Nov 25 10:34:12 crc kubenswrapper[4821]: E1125 10:34:12.114485 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4msn" podUID="1f12cef3-7e40-4f61-836e-23cd0e578e71" Nov 25 10:34:13 crc kubenswrapper[4821]: I1125 10:34:13.113534 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:34:13 crc kubenswrapper[4821]: I1125 10:34:13.113641 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:34:13 crc kubenswrapper[4821]: I1125 10:34:13.113534 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:34:13 crc kubenswrapper[4821]: E1125 10:34:13.113779 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:34:13 crc kubenswrapper[4821]: E1125 10:34:13.113681 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:34:13 crc kubenswrapper[4821]: E1125 10:34:13.113874 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:34:14 crc kubenswrapper[4821]: I1125 10:34:14.113449 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4msn" Nov 25 10:34:14 crc kubenswrapper[4821]: E1125 10:34:14.113600 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4msn" podUID="1f12cef3-7e40-4f61-836e-23cd0e578e71" Nov 25 10:34:15 crc kubenswrapper[4821]: I1125 10:34:15.114042 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:34:15 crc kubenswrapper[4821]: I1125 10:34:15.114099 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:34:15 crc kubenswrapper[4821]: E1125 10:34:15.114177 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:34:15 crc kubenswrapper[4821]: I1125 10:34:15.114207 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:34:15 crc kubenswrapper[4821]: E1125 10:34:15.114299 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:34:15 crc kubenswrapper[4821]: E1125 10:34:15.114421 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:34:15 crc kubenswrapper[4821]: E1125 10:34:15.208949 4821 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 25 10:34:16 crc kubenswrapper[4821]: I1125 10:34:16.114094 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4msn" Nov 25 10:34:16 crc kubenswrapper[4821]: E1125 10:34:16.114260 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4msn" podUID="1f12cef3-7e40-4f61-836e-23cd0e578e71" Nov 25 10:34:17 crc kubenswrapper[4821]: I1125 10:34:17.113311 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:34:17 crc kubenswrapper[4821]: I1125 10:34:17.113407 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:34:17 crc kubenswrapper[4821]: E1125 10:34:17.113450 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:34:17 crc kubenswrapper[4821]: I1125 10:34:17.113534 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:34:17 crc kubenswrapper[4821]: E1125 10:34:17.113591 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:34:17 crc kubenswrapper[4821]: E1125 10:34:17.114574 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:34:18 crc kubenswrapper[4821]: I1125 10:34:18.114044 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4msn" Nov 25 10:34:18 crc kubenswrapper[4821]: E1125 10:34:18.114207 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4msn" podUID="1f12cef3-7e40-4f61-836e-23cd0e578e71" Nov 25 10:34:19 crc kubenswrapper[4821]: I1125 10:34:19.114289 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:34:19 crc kubenswrapper[4821]: E1125 10:34:19.114442 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:34:19 crc kubenswrapper[4821]: I1125 10:34:19.114467 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:34:19 crc kubenswrapper[4821]: I1125 10:34:19.114478 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:34:19 crc kubenswrapper[4821]: E1125 10:34:19.114672 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:34:19 crc kubenswrapper[4821]: E1125 10:34:19.114805 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:34:19 crc kubenswrapper[4821]: I1125 10:34:19.114876 4821 scope.go:117] "RemoveContainer" containerID="418dc8577fdc98324677dbccfcf9317b67cab4408432b30df2818b22f88d426e" Nov 25 10:34:19 crc kubenswrapper[4821]: I1125 10:34:19.624789 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-ldzbz_b84e32ae-4897-4644-a656-6bd82012cc81/kube-multus/1.log" Nov 25 10:34:19 crc kubenswrapper[4821]: I1125 10:34:19.624843 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-ldzbz" event={"ID":"b84e32ae-4897-4644-a656-6bd82012cc81","Type":"ContainerStarted","Data":"86b9194a6cea3434f70eb5f6d91aa470cb4321b386c78735cf4ca3da718f1026"} Nov 25 10:34:20 crc kubenswrapper[4821]: I1125 10:34:20.113916 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4msn" Nov 25 10:34:20 crc kubenswrapper[4821]: E1125 10:34:20.115139 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4msn" podUID="1f12cef3-7e40-4f61-836e-23cd0e578e71" Nov 25 10:34:20 crc kubenswrapper[4821]: E1125 10:34:20.209915 4821 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 25 10:34:21 crc kubenswrapper[4821]: I1125 10:34:21.113822 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:34:21 crc kubenswrapper[4821]: I1125 10:34:21.113887 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:34:21 crc kubenswrapper[4821]: E1125 10:34:21.113956 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:34:21 crc kubenswrapper[4821]: I1125 10:34:21.113993 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:34:21 crc kubenswrapper[4821]: E1125 10:34:21.114222 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:34:21 crc kubenswrapper[4821]: E1125 10:34:21.114328 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:34:22 crc kubenswrapper[4821]: I1125 10:34:22.114030 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4msn" Nov 25 10:34:22 crc kubenswrapper[4821]: E1125 10:34:22.114627 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4msn" podUID="1f12cef3-7e40-4f61-836e-23cd0e578e71" Nov 25 10:34:23 crc kubenswrapper[4821]: I1125 10:34:23.113793 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:34:23 crc kubenswrapper[4821]: I1125 10:34:23.113873 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:34:23 crc kubenswrapper[4821]: I1125 10:34:23.113905 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:34:23 crc kubenswrapper[4821]: E1125 10:34:23.113915 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:34:23 crc kubenswrapper[4821]: E1125 10:34:23.113995 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:34:23 crc kubenswrapper[4821]: E1125 10:34:23.114064 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:34:24 crc kubenswrapper[4821]: I1125 10:34:24.113255 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4msn" Nov 25 10:34:24 crc kubenswrapper[4821]: E1125 10:34:24.113392 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-d4msn" podUID="1f12cef3-7e40-4f61-836e-23cd0e578e71" Nov 25 10:34:25 crc kubenswrapper[4821]: I1125 10:34:25.113463 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:34:25 crc kubenswrapper[4821]: I1125 10:34:25.113552 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:34:25 crc kubenswrapper[4821]: I1125 10:34:25.113551 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:34:25 crc kubenswrapper[4821]: E1125 10:34:25.113694 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 10:34:25 crc kubenswrapper[4821]: E1125 10:34:25.113865 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 10:34:25 crc kubenswrapper[4821]: E1125 10:34:25.114042 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 10:34:26 crc kubenswrapper[4821]: I1125 10:34:26.113926 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4msn" Nov 25 10:34:26 crc kubenswrapper[4821]: I1125 10:34:26.115644 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Nov 25 10:34:26 crc kubenswrapper[4821]: I1125 10:34:26.117594 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Nov 25 10:34:27 crc kubenswrapper[4821]: I1125 10:34:27.113630 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:34:27 crc kubenswrapper[4821]: I1125 10:34:27.113630 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:34:27 crc kubenswrapper[4821]: I1125 10:34:27.113636 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:34:27 crc kubenswrapper[4821]: I1125 10:34:27.115307 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 25 10:34:27 crc kubenswrapper[4821]: I1125 10:34:27.115338 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 25 10:34:27 crc kubenswrapper[4821]: I1125 10:34:27.115753 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 25 10:34:27 crc kubenswrapper[4821]: I1125 10:34:27.116783 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.519772 4821 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.555179 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-86cnx"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.555678 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-86cnx" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.558196 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-mbn4k"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.558622 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-mbn4k" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.559545 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-t5hm7"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.559901 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-t5hm7" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.562929 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-q8tq8"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.563444 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-q8tq8" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.567023 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.568253 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.568802 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.570509 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-hf49d"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.571662 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-hf49d" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.572026 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.573119 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.573297 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.573750 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.573996 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.574246 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.573756 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.574509 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.574647 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.574652 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.574866 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.576669 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.576874 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.577736 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-58htt"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.578308 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5m5gj"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.578505 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-58htt" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.579272 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5m5gj" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.590465 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.590696 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.591986 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-dv9nt"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.592472 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dv9nt" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.630119 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-tpbzb"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.630512 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-tpbzb" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.631871 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.633785 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.634355 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.638784 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.639463 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.640183 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.640331 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.640498 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.640620 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.640784 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.640844 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.640935 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.641021 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.641153 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-zj4s2"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.641498 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.641805 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-zj4s2" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.641872 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k622z"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.642001 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.642177 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.642480 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.642596 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.642607 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k622z" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.642701 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.642805 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.643018 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.643142 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.643544 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.643782 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.643886 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.644451 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.644904 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.646411 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-hjvt9"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.646895 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-hjvt9" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.647591 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.649456 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zvztq"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.650126 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zvztq" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.650683 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-jb5sz"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.651554 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-jb5sz" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.653448 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.653642 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.653748 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.655962 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-hddc7"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.656682 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4kjbr"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.657121 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4kjbr" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.657253 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-v88lg"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.665787 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-hddc7" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.666967 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-p2jmh"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.667458 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-p2jmh" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.668063 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-v88lg" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.676981 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/518db853-3bbe-421b-a2dd-6d1b2e965120-config\") pod \"controller-manager-879f6c89f-58htt\" (UID: \"518db853-3bbe-421b-a2dd-6d1b2e965120\") " pod="openshift-controller-manager/controller-manager-879f6c89f-58htt" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.677027 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba27af47-1944-4970-a015-e96c7cee4862-config\") pod \"route-controller-manager-6576b87f9c-dv9nt\" (UID: \"ba27af47-1944-4970-a015-e96c7cee4862\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dv9nt" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.677054 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/88feae6a-5031-48ed-93ec-34ef8d2eff40-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-mbn4k\" (UID: \"88feae6a-5031-48ed-93ec-34ef8d2eff40\") " pod="openshift-authentication/oauth-openshift-558db77b4-mbn4k" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.677077 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gzl9t\" (UniqueName: \"kubernetes.io/projected/ba27af47-1944-4970-a015-e96c7cee4862-kube-api-access-gzl9t\") pod \"route-controller-manager-6576b87f9c-dv9nt\" (UID: \"ba27af47-1944-4970-a015-e96c7cee4862\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dv9nt" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.677102 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wpnl\" (UniqueName: \"kubernetes.io/projected/59ec70c2-c208-4529-a763-a0d7ea7e1dfc-kube-api-access-9wpnl\") pod \"machine-approver-56656f9798-86cnx\" (UID: \"59ec70c2-c208-4529-a763-a0d7ea7e1dfc\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-86cnx" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.677127 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5a7c1a05-b29c-4ff3-a4d6-90fe59eb1a01-serving-cert\") pod \"openshift-config-operator-7777fb866f-q8tq8\" (UID: \"5a7c1a05-b29c-4ff3-a4d6-90fe59eb1a01\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-q8tq8" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.677144 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/88feae6a-5031-48ed-93ec-34ef8d2eff40-audit-policies\") pod \"oauth-openshift-558db77b4-mbn4k\" (UID: \"88feae6a-5031-48ed-93ec-34ef8d2eff40\") " pod="openshift-authentication/oauth-openshift-558db77b4-mbn4k" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.681680 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/88feae6a-5031-48ed-93ec-34ef8d2eff40-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-mbn4k\" (UID: \"88feae6a-5031-48ed-93ec-34ef8d2eff40\") " pod="openshift-authentication/oauth-openshift-558db77b4-mbn4k" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.681754 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/88725123-3da9-4127-9506-80c90c9accd5-oauth-serving-cert\") pod \"console-f9d7485db-hf49d\" (UID: \"88725123-3da9-4127-9506-80c90c9accd5\") " pod="openshift-console/console-f9d7485db-hf49d" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.681791 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-trdqk\" (UniqueName: \"kubernetes.io/projected/518db853-3bbe-421b-a2dd-6d1b2e965120-kube-api-access-trdqk\") pod \"controller-manager-879f6c89f-58htt\" (UID: \"518db853-3bbe-421b-a2dd-6d1b2e965120\") " pod="openshift-controller-manager/controller-manager-879f6c89f-58htt" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.681816 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/88feae6a-5031-48ed-93ec-34ef8d2eff40-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-mbn4k\" (UID: \"88feae6a-5031-48ed-93ec-34ef8d2eff40\") " pod="openshift-authentication/oauth-openshift-558db77b4-mbn4k" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.681851 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/88feae6a-5031-48ed-93ec-34ef8d2eff40-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-mbn4k\" (UID: \"88feae6a-5031-48ed-93ec-34ef8d2eff40\") " pod="openshift-authentication/oauth-openshift-558db77b4-mbn4k" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.681953 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b75a7773-a689-46c1-b98e-e7c1be0878e6-trusted-ca\") pod \"console-operator-58897d9998-t5hm7\" (UID: \"b75a7773-a689-46c1-b98e-e7c1be0878e6\") " pod="openshift-console-operator/console-operator-58897d9998-t5hm7" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.681985 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/88feae6a-5031-48ed-93ec-34ef8d2eff40-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-mbn4k\" (UID: \"88feae6a-5031-48ed-93ec-34ef8d2eff40\") " pod="openshift-authentication/oauth-openshift-558db77b4-mbn4k" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.682016 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59ec70c2-c208-4529-a763-a0d7ea7e1dfc-config\") pod \"machine-approver-56656f9798-86cnx\" (UID: \"59ec70c2-c208-4529-a763-a0d7ea7e1dfc\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-86cnx" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.682045 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/518db853-3bbe-421b-a2dd-6d1b2e965120-serving-cert\") pod \"controller-manager-879f6c89f-58htt\" (UID: \"518db853-3bbe-421b-a2dd-6d1b2e965120\") " pod="openshift-controller-manager/controller-manager-879f6c89f-58htt" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.682070 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/518db853-3bbe-421b-a2dd-6d1b2e965120-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-58htt\" (UID: \"518db853-3bbe-421b-a2dd-6d1b2e965120\") " pod="openshift-controller-manager/controller-manager-879f6c89f-58htt" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.682095 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/88feae6a-5031-48ed-93ec-34ef8d2eff40-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-mbn4k\" (UID: \"88feae6a-5031-48ed-93ec-34ef8d2eff40\") " pod="openshift-authentication/oauth-openshift-558db77b4-mbn4k" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.682132 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njkct\" (UniqueName: \"kubernetes.io/projected/cb0f66e4-2c0f-4a95-89db-1aef53ae5b3b-kube-api-access-njkct\") pod \"openshift-apiserver-operator-796bbdcf4f-5m5gj\" (UID: \"cb0f66e4-2c0f-4a95-89db-1aef53ae5b3b\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5m5gj" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.682177 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/88725123-3da9-4127-9506-80c90c9accd5-console-config\") pod \"console-f9d7485db-hf49d\" (UID: \"88725123-3da9-4127-9506-80c90c9accd5\") " pod="openshift-console/console-f9d7485db-hf49d" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.682206 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ktx4k\" (UniqueName: \"kubernetes.io/projected/5a7c1a05-b29c-4ff3-a4d6-90fe59eb1a01-kube-api-access-ktx4k\") pod \"openshift-config-operator-7777fb866f-q8tq8\" (UID: \"5a7c1a05-b29c-4ff3-a4d6-90fe59eb1a01\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-q8tq8" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.682261 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/88725123-3da9-4127-9506-80c90c9accd5-service-ca\") pod \"console-f9d7485db-hf49d\" (UID: \"88725123-3da9-4127-9506-80c90c9accd5\") " pod="openshift-console/console-f9d7485db-hf49d" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.682288 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/88feae6a-5031-48ed-93ec-34ef8d2eff40-audit-dir\") pod \"oauth-openshift-558db77b4-mbn4k\" (UID: \"88feae6a-5031-48ed-93ec-34ef8d2eff40\") " pod="openshift-authentication/oauth-openshift-558db77b4-mbn4k" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.682318 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/88feae6a-5031-48ed-93ec-34ef8d2eff40-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-mbn4k\" (UID: \"88feae6a-5031-48ed-93ec-34ef8d2eff40\") " pod="openshift-authentication/oauth-openshift-558db77b4-mbn4k" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.682352 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/88725123-3da9-4127-9506-80c90c9accd5-trusted-ca-bundle\") pod \"console-f9d7485db-hf49d\" (UID: \"88725123-3da9-4127-9506-80c90c9accd5\") " pod="openshift-console/console-f9d7485db-hf49d" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.682374 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/88feae6a-5031-48ed-93ec-34ef8d2eff40-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-mbn4k\" (UID: \"88feae6a-5031-48ed-93ec-34ef8d2eff40\") " pod="openshift-authentication/oauth-openshift-558db77b4-mbn4k" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.682408 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/5a7c1a05-b29c-4ff3-a4d6-90fe59eb1a01-available-featuregates\") pod \"openshift-config-operator-7777fb866f-q8tq8\" (UID: \"5a7c1a05-b29c-4ff3-a4d6-90fe59eb1a01\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-q8tq8" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.682458 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb0f66e4-2c0f-4a95-89db-1aef53ae5b3b-config\") pod \"openshift-apiserver-operator-796bbdcf4f-5m5gj\" (UID: \"cb0f66e4-2c0f-4a95-89db-1aef53ae5b3b\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5m5gj" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.682499 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/88feae6a-5031-48ed-93ec-34ef8d2eff40-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-mbn4k\" (UID: \"88feae6a-5031-48ed-93ec-34ef8d2eff40\") " pod="openshift-authentication/oauth-openshift-558db77b4-mbn4k" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.682537 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ba27af47-1944-4970-a015-e96c7cee4862-client-ca\") pod \"route-controller-manager-6576b87f9c-dv9nt\" (UID: \"ba27af47-1944-4970-a015-e96c7cee4862\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dv9nt" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.682563 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b75a7773-a689-46c1-b98e-e7c1be0878e6-config\") pod \"console-operator-58897d9998-t5hm7\" (UID: \"b75a7773-a689-46c1-b98e-e7c1be0878e6\") " pod="openshift-console-operator/console-operator-58897d9998-t5hm7" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.682624 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/88725123-3da9-4127-9506-80c90c9accd5-console-oauth-config\") pod \"console-f9d7485db-hf49d\" (UID: \"88725123-3da9-4127-9506-80c90c9accd5\") " pod="openshift-console/console-f9d7485db-hf49d" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.682653 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b75a7773-a689-46c1-b98e-e7c1be0878e6-serving-cert\") pod \"console-operator-58897d9998-t5hm7\" (UID: \"b75a7773-a689-46c1-b98e-e7c1be0878e6\") " pod="openshift-console-operator/console-operator-58897d9998-t5hm7" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.682675 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/88feae6a-5031-48ed-93ec-34ef8d2eff40-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-mbn4k\" (UID: \"88feae6a-5031-48ed-93ec-34ef8d2eff40\") " pod="openshift-authentication/oauth-openshift-558db77b4-mbn4k" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.682703 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7b5z\" (UniqueName: \"kubernetes.io/projected/88feae6a-5031-48ed-93ec-34ef8d2eff40-kube-api-access-q7b5z\") pod \"oauth-openshift-558db77b4-mbn4k\" (UID: \"88feae6a-5031-48ed-93ec-34ef8d2eff40\") " pod="openshift-authentication/oauth-openshift-558db77b4-mbn4k" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.682732 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/59ec70c2-c208-4529-a763-a0d7ea7e1dfc-machine-approver-tls\") pod \"machine-approver-56656f9798-86cnx\" (UID: \"59ec70c2-c208-4529-a763-a0d7ea7e1dfc\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-86cnx" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.682758 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5x4xr\" (UniqueName: \"kubernetes.io/projected/88725123-3da9-4127-9506-80c90c9accd5-kube-api-access-5x4xr\") pod \"console-f9d7485db-hf49d\" (UID: \"88725123-3da9-4127-9506-80c90c9accd5\") " pod="openshift-console/console-f9d7485db-hf49d" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.682780 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/88725123-3da9-4127-9506-80c90c9accd5-console-serving-cert\") pod \"console-f9d7485db-hf49d\" (UID: \"88725123-3da9-4127-9506-80c90c9accd5\") " pod="openshift-console/console-f9d7485db-hf49d" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.682814 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8g76m\" (UniqueName: \"kubernetes.io/projected/b75a7773-a689-46c1-b98e-e7c1be0878e6-kube-api-access-8g76m\") pod \"console-operator-58897d9998-t5hm7\" (UID: \"b75a7773-a689-46c1-b98e-e7c1be0878e6\") " pod="openshift-console-operator/console-operator-58897d9998-t5hm7" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.682843 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cb0f66e4-2c0f-4a95-89db-1aef53ae5b3b-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-5m5gj\" (UID: \"cb0f66e4-2c0f-4a95-89db-1aef53ae5b3b\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5m5gj" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.682873 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/518db853-3bbe-421b-a2dd-6d1b2e965120-client-ca\") pod \"controller-manager-879f6c89f-58htt\" (UID: \"518db853-3bbe-421b-a2dd-6d1b2e965120\") " pod="openshift-controller-manager/controller-manager-879f6c89f-58htt" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.682943 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/88feae6a-5031-48ed-93ec-34ef8d2eff40-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-mbn4k\" (UID: \"88feae6a-5031-48ed-93ec-34ef8d2eff40\") " pod="openshift-authentication/oauth-openshift-558db77b4-mbn4k" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.683004 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/59ec70c2-c208-4529-a763-a0d7ea7e1dfc-auth-proxy-config\") pod \"machine-approver-56656f9798-86cnx\" (UID: \"59ec70c2-c208-4529-a763-a0d7ea7e1dfc\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-86cnx" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.683254 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ba27af47-1944-4970-a015-e96c7cee4862-serving-cert\") pod \"route-controller-manager-6576b87f9c-dv9nt\" (UID: \"ba27af47-1944-4970-a015-e96c7cee4862\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dv9nt" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.691323 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.691689 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.691856 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.692384 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.692732 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.692808 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.692882 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.693004 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.693098 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.693338 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.693351 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.693438 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.693457 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.693510 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.693893 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.693928 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.695190 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-whcwc"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.695954 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-whcwc" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.696767 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wxp2r"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.701351 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.701647 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.701786 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.711370 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.718223 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.718939 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.718952 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wxp2r" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.719083 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.720541 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.720909 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.722459 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.722960 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.723061 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.723432 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.722965 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.723668 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.723679 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.723005 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.723843 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.723034 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.724085 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.724131 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.724196 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.724243 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.724281 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.724348 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.724411 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.724474 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.724518 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.724572 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.725273 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.725631 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-n4lbr"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.726270 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-n4lbr" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.730525 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6xxs8"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.730705 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.731155 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6xxs8" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.732848 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.736444 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-ng2nv"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.736872 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.737275 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ng2nv" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.738662 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.740297 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.740467 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tkgh9"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.741029 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tkgh9" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.741596 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.741915 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.742869 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-w5k28"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.743767 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w5k28" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.745402 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-kps6n"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.746278 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kps6n" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.749057 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.749306 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.751895 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.752352 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.760916 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.765063 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-p92sp"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.766279 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-p92sp" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.767097 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.767414 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-dxf4x"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.771263 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-xnlvh"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.771821 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-dxf4x" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.773917 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-prvdm"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.774713 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.775625 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d8x5t"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.775749 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-prvdm" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.777053 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-7792m"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.778288 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d8x5t" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.778858 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-f2tjh"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.779860 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-7792m" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.786628 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/88feae6a-5031-48ed-93ec-34ef8d2eff40-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-mbn4k\" (UID: \"88feae6a-5031-48ed-93ec-34ef8d2eff40\") " pod="openshift-authentication/oauth-openshift-558db77b4-mbn4k" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.786676 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tpsds\" (UniqueName: \"kubernetes.io/projected/513166e2-0ac9-4299-bf84-d994a46da192-kube-api-access-tpsds\") pod \"cluster-samples-operator-665b6dd947-k622z\" (UID: \"513166e2-0ac9-4299-bf84-d994a46da192\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k622z" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.786706 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/064de93d-7bb5-47b5-9585-bf32be9a05f6-config\") pod \"apiserver-76f77b778f-jb5sz\" (UID: \"064de93d-7bb5-47b5-9585-bf32be9a05f6\") " pod="openshift-apiserver/apiserver-76f77b778f-jb5sz" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.786745 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b75a7773-a689-46c1-b98e-e7c1be0878e6-trusted-ca\") pod \"console-operator-58897d9998-t5hm7\" (UID: \"b75a7773-a689-46c1-b98e-e7c1be0878e6\") " pod="openshift-console-operator/console-operator-58897d9998-t5hm7" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.786744 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.786766 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/88feae6a-5031-48ed-93ec-34ef8d2eff40-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-mbn4k\" (UID: \"88feae6a-5031-48ed-93ec-34ef8d2eff40\") " pod="openshift-authentication/oauth-openshift-558db77b4-mbn4k" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.786788 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59ec70c2-c208-4529-a763-a0d7ea7e1dfc-config\") pod \"machine-approver-56656f9798-86cnx\" (UID: \"59ec70c2-c208-4529-a763-a0d7ea7e1dfc\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-86cnx" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.786807 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/518db853-3bbe-421b-a2dd-6d1b2e965120-serving-cert\") pod \"controller-manager-879f6c89f-58htt\" (UID: \"518db853-3bbe-421b-a2dd-6d1b2e965120\") " pod="openshift-controller-manager/controller-manager-879f6c89f-58htt" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.786826 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/518db853-3bbe-421b-a2dd-6d1b2e965120-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-58htt\" (UID: \"518db853-3bbe-421b-a2dd-6d1b2e965120\") " pod="openshift-controller-manager/controller-manager-879f6c89f-58htt" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.786853 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njkct\" (UniqueName: \"kubernetes.io/projected/cb0f66e4-2c0f-4a95-89db-1aef53ae5b3b-kube-api-access-njkct\") pod \"openshift-apiserver-operator-796bbdcf4f-5m5gj\" (UID: \"cb0f66e4-2c0f-4a95-89db-1aef53ae5b3b\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5m5gj" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.786874 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/88feae6a-5031-48ed-93ec-34ef8d2eff40-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-mbn4k\" (UID: \"88feae6a-5031-48ed-93ec-34ef8d2eff40\") " pod="openshift-authentication/oauth-openshift-558db77b4-mbn4k" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.786895 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/064de93d-7bb5-47b5-9585-bf32be9a05f6-etcd-client\") pod \"apiserver-76f77b778f-jb5sz\" (UID: \"064de93d-7bb5-47b5-9585-bf32be9a05f6\") " pod="openshift-apiserver/apiserver-76f77b778f-jb5sz" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.787230 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fq9pc\" (UniqueName: \"kubernetes.io/projected/b0e34420-095d-42aa-89c4-5a80331e1c01-kube-api-access-fq9pc\") pod \"dns-operator-744455d44c-hddc7\" (UID: \"b0e34420-095d-42aa-89c4-5a80331e1c01\") " pod="openshift-dns-operator/dns-operator-744455d44c-hddc7" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.787271 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ktx4k\" (UniqueName: \"kubernetes.io/projected/5a7c1a05-b29c-4ff3-a4d6-90fe59eb1a01-kube-api-access-ktx4k\") pod \"openshift-config-operator-7777fb866f-q8tq8\" (UID: \"5a7c1a05-b29c-4ff3-a4d6-90fe59eb1a01\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-q8tq8" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.787299 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/88725123-3da9-4127-9506-80c90c9accd5-console-config\") pod \"console-f9d7485db-hf49d\" (UID: \"88725123-3da9-4127-9506-80c90c9accd5\") " pod="openshift-console/console-f9d7485db-hf49d" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.787328 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/064de93d-7bb5-47b5-9585-bf32be9a05f6-encryption-config\") pod \"apiserver-76f77b778f-jb5sz\" (UID: \"064de93d-7bb5-47b5-9585-bf32be9a05f6\") " pod="openshift-apiserver/apiserver-76f77b778f-jb5sz" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.787353 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/24708c13-efe2-435b-acde-b29b48871f16-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-zj4s2\" (UID: \"24708c13-efe2-435b-acde-b29b48871f16\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zj4s2" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.787380 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/88feae6a-5031-48ed-93ec-34ef8d2eff40-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-mbn4k\" (UID: \"88feae6a-5031-48ed-93ec-34ef8d2eff40\") " pod="openshift-authentication/oauth-openshift-558db77b4-mbn4k" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.787402 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/064de93d-7bb5-47b5-9585-bf32be9a05f6-node-pullsecrets\") pod \"apiserver-76f77b778f-jb5sz\" (UID: \"064de93d-7bb5-47b5-9585-bf32be9a05f6\") " pod="openshift-apiserver/apiserver-76f77b778f-jb5sz" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.787427 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6cq5w\" (UniqueName: \"kubernetes.io/projected/2c2ecf12-49bd-4baf-90df-68486f27be1b-kube-api-access-6cq5w\") pod \"catalog-operator-68c6474976-4kjbr\" (UID: \"2c2ecf12-49bd-4baf-90df-68486f27be1b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4kjbr" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.787451 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/88725123-3da9-4127-9506-80c90c9accd5-service-ca\") pod \"console-f9d7485db-hf49d\" (UID: \"88725123-3da9-4127-9506-80c90c9accd5\") " pod="openshift-console/console-f9d7485db-hf49d" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.787472 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/88feae6a-5031-48ed-93ec-34ef8d2eff40-audit-dir\") pod \"oauth-openshift-558db77b4-mbn4k\" (UID: \"88feae6a-5031-48ed-93ec-34ef8d2eff40\") " pod="openshift-authentication/oauth-openshift-558db77b4-mbn4k" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.787494 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7fe78aa4-571c-488b-96bd-23ae4b6333db-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-zvztq\" (UID: \"7fe78aa4-571c-488b-96bd-23ae4b6333db\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zvztq" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.787516 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/064de93d-7bb5-47b5-9585-bf32be9a05f6-audit\") pod \"apiserver-76f77b778f-jb5sz\" (UID: \"064de93d-7bb5-47b5-9585-bf32be9a05f6\") " pod="openshift-apiserver/apiserver-76f77b778f-jb5sz" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.787537 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/064de93d-7bb5-47b5-9585-bf32be9a05f6-image-import-ca\") pod \"apiserver-76f77b778f-jb5sz\" (UID: \"064de93d-7bb5-47b5-9585-bf32be9a05f6\") " pod="openshift-apiserver/apiserver-76f77b778f-jb5sz" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.787561 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/064de93d-7bb5-47b5-9585-bf32be9a05f6-serving-cert\") pod \"apiserver-76f77b778f-jb5sz\" (UID: \"064de93d-7bb5-47b5-9585-bf32be9a05f6\") " pod="openshift-apiserver/apiserver-76f77b778f-jb5sz" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.787583 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/064de93d-7bb5-47b5-9585-bf32be9a05f6-trusted-ca-bundle\") pod \"apiserver-76f77b778f-jb5sz\" (UID: \"064de93d-7bb5-47b5-9585-bf32be9a05f6\") " pod="openshift-apiserver/apiserver-76f77b778f-jb5sz" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.787605 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v86wz\" (UniqueName: \"kubernetes.io/projected/064de93d-7bb5-47b5-9585-bf32be9a05f6-kube-api-access-v86wz\") pod \"apiserver-76f77b778f-jb5sz\" (UID: \"064de93d-7bb5-47b5-9585-bf32be9a05f6\") " pod="openshift-apiserver/apiserver-76f77b778f-jb5sz" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.787628 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8234c8fc-a19c-45c6-8d25-e65668a244ea-config\") pod \"kube-controller-manager-operator-78b949d7b-wxp2r\" (UID: \"8234c8fc-a19c-45c6-8d25-e65668a244ea\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wxp2r" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.787652 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/88725123-3da9-4127-9506-80c90c9accd5-trusted-ca-bundle\") pod \"console-f9d7485db-hf49d\" (UID: \"88725123-3da9-4127-9506-80c90c9accd5\") " pod="openshift-console/console-f9d7485db-hf49d" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.787675 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/88feae6a-5031-48ed-93ec-34ef8d2eff40-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-mbn4k\" (UID: \"88feae6a-5031-48ed-93ec-34ef8d2eff40\") " pod="openshift-authentication/oauth-openshift-558db77b4-mbn4k" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.787700 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/155595fc-76b3-4561-bd3f-c540076d4cd6-profile-collector-cert\") pod \"olm-operator-6b444d44fb-v88lg\" (UID: \"155595fc-76b3-4561-bd3f-c540076d4cd6\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-v88lg" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.787720 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4qsq\" (UniqueName: \"kubernetes.io/projected/402cf28b-25d2-428a-ab7b-1d70345a921e-kube-api-access-r4qsq\") pod \"multus-admission-controller-857f4d67dd-p2jmh\" (UID: \"402cf28b-25d2-428a-ab7b-1d70345a921e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-p2jmh" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.787745 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/5a7c1a05-b29c-4ff3-a4d6-90fe59eb1a01-available-featuregates\") pod \"openshift-config-operator-7777fb866f-q8tq8\" (UID: \"5a7c1a05-b29c-4ff3-a4d6-90fe59eb1a01\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-q8tq8" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.787767 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9f6c1e2f-b7e7-40e7-92c7-8ca1b57eb9f1-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-tpbzb\" (UID: \"9f6c1e2f-b7e7-40e7-92c7-8ca1b57eb9f1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tpbzb" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.787791 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bdd88e7d-edf6-41c1-a5a3-47746bb3ab99-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-whcwc\" (UID: \"bdd88e7d-edf6-41c1-a5a3-47746bb3ab99\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-whcwc" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.787811 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/24708c13-efe2-435b-acde-b29b48871f16-config\") pod \"machine-api-operator-5694c8668f-zj4s2\" (UID: \"24708c13-efe2-435b-acde-b29b48871f16\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zj4s2" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.787829 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/402cf28b-25d2-428a-ab7b-1d70345a921e-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-p2jmh\" (UID: \"402cf28b-25d2-428a-ab7b-1d70345a921e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-p2jmh" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.787849 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-44zw6\" (UniqueName: \"kubernetes.io/projected/f954533e-e8a0-4899-be72-241760feae95-kube-api-access-44zw6\") pod \"downloads-7954f5f757-hjvt9\" (UID: \"f954533e-e8a0-4899-be72-241760feae95\") " pod="openshift-console/downloads-7954f5f757-hjvt9" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.787880 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb0f66e4-2c0f-4a95-89db-1aef53ae5b3b-config\") pod \"openshift-apiserver-operator-796bbdcf4f-5m5gj\" (UID: \"cb0f66e4-2c0f-4a95-89db-1aef53ae5b3b\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5m5gj" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.787901 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f6c1e2f-b7e7-40e7-92c7-8ca1b57eb9f1-config\") pod \"authentication-operator-69f744f599-tpbzb\" (UID: \"9f6c1e2f-b7e7-40e7-92c7-8ca1b57eb9f1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tpbzb" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.787922 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9f6c1e2f-b7e7-40e7-92c7-8ca1b57eb9f1-service-ca-bundle\") pod \"authentication-operator-69f744f599-tpbzb\" (UID: \"9f6c1e2f-b7e7-40e7-92c7-8ca1b57eb9f1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tpbzb" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.787940 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4878872d-8035-442a-a355-7c966f37e8c1-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-n4lbr\" (UID: \"4878872d-8035-442a-a355-7c966f37e8c1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-n4lbr" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.787974 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/88feae6a-5031-48ed-93ec-34ef8d2eff40-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-mbn4k\" (UID: \"88feae6a-5031-48ed-93ec-34ef8d2eff40\") " pod="openshift-authentication/oauth-openshift-558db77b4-mbn4k" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.787992 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59ec70c2-c208-4529-a763-a0d7ea7e1dfc-config\") pod \"machine-approver-56656f9798-86cnx\" (UID: \"59ec70c2-c208-4529-a763-a0d7ea7e1dfc\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-86cnx" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.787998 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kbrfs\" (UniqueName: \"kubernetes.io/projected/9f6c1e2f-b7e7-40e7-92c7-8ca1b57eb9f1-kube-api-access-kbrfs\") pod \"authentication-operator-69f744f599-tpbzb\" (UID: \"9f6c1e2f-b7e7-40e7-92c7-8ca1b57eb9f1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tpbzb" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.788077 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8234c8fc-a19c-45c6-8d25-e65668a244ea-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-wxp2r\" (UID: \"8234c8fc-a19c-45c6-8d25-e65668a244ea\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wxp2r" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.788112 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b0e34420-095d-42aa-89c4-5a80331e1c01-metrics-tls\") pod \"dns-operator-744455d44c-hddc7\" (UID: \"b0e34420-095d-42aa-89c4-5a80331e1c01\") " pod="openshift-dns-operator/dns-operator-744455d44c-hddc7" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.788141 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e82d98b-7c99-496d-910d-e9bfae2ca6ec-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-6xxs8\" (UID: \"0e82d98b-7c99-496d-910d-e9bfae2ca6ec\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6xxs8" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.791243 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rc8sv\" (UniqueName: \"kubernetes.io/projected/0e82d98b-7c99-496d-910d-e9bfae2ca6ec-kube-api-access-rc8sv\") pod \"kube-storage-version-migrator-operator-b67b599dd-6xxs8\" (UID: \"0e82d98b-7c99-496d-910d-e9bfae2ca6ec\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6xxs8" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.791288 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ba27af47-1944-4970-a015-e96c7cee4862-client-ca\") pod \"route-controller-manager-6576b87f9c-dv9nt\" (UID: \"ba27af47-1944-4970-a015-e96c7cee4862\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dv9nt" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.791309 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b75a7773-a689-46c1-b98e-e7c1be0878e6-config\") pod \"console-operator-58897d9998-t5hm7\" (UID: \"b75a7773-a689-46c1-b98e-e7c1be0878e6\") " pod="openshift-console-operator/console-operator-58897d9998-t5hm7" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.791326 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/064de93d-7bb5-47b5-9585-bf32be9a05f6-etcd-serving-ca\") pod \"apiserver-76f77b778f-jb5sz\" (UID: \"064de93d-7bb5-47b5-9585-bf32be9a05f6\") " pod="openshift-apiserver/apiserver-76f77b778f-jb5sz" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.791374 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4878872d-8035-442a-a355-7c966f37e8c1-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-n4lbr\" (UID: \"4878872d-8035-442a-a355-7c966f37e8c1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-n4lbr" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.791400 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/88725123-3da9-4127-9506-80c90c9accd5-console-oauth-config\") pod \"console-f9d7485db-hf49d\" (UID: \"88725123-3da9-4127-9506-80c90c9accd5\") " pod="openshift-console/console-f9d7485db-hf49d" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.791417 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/064de93d-7bb5-47b5-9585-bf32be9a05f6-audit-dir\") pod \"apiserver-76f77b778f-jb5sz\" (UID: \"064de93d-7bb5-47b5-9585-bf32be9a05f6\") " pod="openshift-apiserver/apiserver-76f77b778f-jb5sz" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.791436 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/2c2ecf12-49bd-4baf-90df-68486f27be1b-srv-cert\") pod \"catalog-operator-68c6474976-4kjbr\" (UID: \"2c2ecf12-49bd-4baf-90df-68486f27be1b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4kjbr" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.791455 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7fe78aa4-571c-488b-96bd-23ae4b6333db-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-zvztq\" (UID: \"7fe78aa4-571c-488b-96bd-23ae4b6333db\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zvztq" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.791470 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bdd88e7d-edf6-41c1-a5a3-47746bb3ab99-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-whcwc\" (UID: \"bdd88e7d-edf6-41c1-a5a3-47746bb3ab99\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-whcwc" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.791485 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/155595fc-76b3-4561-bd3f-c540076d4cd6-srv-cert\") pod \"olm-operator-6b444d44fb-v88lg\" (UID: \"155595fc-76b3-4561-bd3f-c540076d4cd6\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-v88lg" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.791500 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4878872d-8035-442a-a355-7c966f37e8c1-config\") pod \"kube-apiserver-operator-766d6c64bb-n4lbr\" (UID: \"4878872d-8035-442a-a355-7c966f37e8c1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-n4lbr" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.791518 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b75a7773-a689-46c1-b98e-e7c1be0878e6-serving-cert\") pod \"console-operator-58897d9998-t5hm7\" (UID: \"b75a7773-a689-46c1-b98e-e7c1be0878e6\") " pod="openshift-console-operator/console-operator-58897d9998-t5hm7" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.791534 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/88feae6a-5031-48ed-93ec-34ef8d2eff40-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-mbn4k\" (UID: \"88feae6a-5031-48ed-93ec-34ef8d2eff40\") " pod="openshift-authentication/oauth-openshift-558db77b4-mbn4k" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.791551 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7b5z\" (UniqueName: \"kubernetes.io/projected/88feae6a-5031-48ed-93ec-34ef8d2eff40-kube-api-access-q7b5z\") pod \"oauth-openshift-558db77b4-mbn4k\" (UID: \"88feae6a-5031-48ed-93ec-34ef8d2eff40\") " pod="openshift-authentication/oauth-openshift-558db77b4-mbn4k" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.791572 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rf8tm\" (UniqueName: \"kubernetes.io/projected/155595fc-76b3-4561-bd3f-c540076d4cd6-kube-api-access-rf8tm\") pod \"olm-operator-6b444d44fb-v88lg\" (UID: \"155595fc-76b3-4561-bd3f-c540076d4cd6\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-v88lg" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.791591 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/59ec70c2-c208-4529-a763-a0d7ea7e1dfc-machine-approver-tls\") pod \"machine-approver-56656f9798-86cnx\" (UID: \"59ec70c2-c208-4529-a763-a0d7ea7e1dfc\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-86cnx" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.791607 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5x4xr\" (UniqueName: \"kubernetes.io/projected/88725123-3da9-4127-9506-80c90c9accd5-kube-api-access-5x4xr\") pod \"console-f9d7485db-hf49d\" (UID: \"88725123-3da9-4127-9506-80c90c9accd5\") " pod="openshift-console/console-f9d7485db-hf49d" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.791622 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tnlzh\" (UniqueName: \"kubernetes.io/projected/24708c13-efe2-435b-acde-b29b48871f16-kube-api-access-tnlzh\") pod \"machine-api-operator-5694c8668f-zj4s2\" (UID: \"24708c13-efe2-435b-acde-b29b48871f16\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zj4s2" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.791640 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/88725123-3da9-4127-9506-80c90c9accd5-console-serving-cert\") pod \"console-f9d7485db-hf49d\" (UID: \"88725123-3da9-4127-9506-80c90c9accd5\") " pod="openshift-console/console-f9d7485db-hf49d" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.791657 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bdd88e7d-edf6-41c1-a5a3-47746bb3ab99-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-whcwc\" (UID: \"bdd88e7d-edf6-41c1-a5a3-47746bb3ab99\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-whcwc" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.791679 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8g76m\" (UniqueName: \"kubernetes.io/projected/b75a7773-a689-46c1-b98e-e7c1be0878e6-kube-api-access-8g76m\") pod \"console-operator-58897d9998-t5hm7\" (UID: \"b75a7773-a689-46c1-b98e-e7c1be0878e6\") " pod="openshift-console-operator/console-operator-58897d9998-t5hm7" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.791700 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cb0f66e4-2c0f-4a95-89db-1aef53ae5b3b-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-5m5gj\" (UID: \"cb0f66e4-2c0f-4a95-89db-1aef53ae5b3b\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5m5gj" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.791722 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/59ec70c2-c208-4529-a763-a0d7ea7e1dfc-auth-proxy-config\") pod \"machine-approver-56656f9798-86cnx\" (UID: \"59ec70c2-c208-4529-a763-a0d7ea7e1dfc\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-86cnx" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.791737 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/518db853-3bbe-421b-a2dd-6d1b2e965120-client-ca\") pod \"controller-manager-879f6c89f-58htt\" (UID: \"518db853-3bbe-421b-a2dd-6d1b2e965120\") " pod="openshift-controller-manager/controller-manager-879f6c89f-58htt" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.791755 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/88feae6a-5031-48ed-93ec-34ef8d2eff40-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-mbn4k\" (UID: \"88feae6a-5031-48ed-93ec-34ef8d2eff40\") " pod="openshift-authentication/oauth-openshift-558db77b4-mbn4k" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.791771 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dt66w\" (UniqueName: \"kubernetes.io/projected/7fe78aa4-571c-488b-96bd-23ae4b6333db-kube-api-access-dt66w\") pod \"openshift-controller-manager-operator-756b6f6bc6-zvztq\" (UID: \"7fe78aa4-571c-488b-96bd-23ae4b6333db\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zvztq" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.791786 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ba27af47-1944-4970-a015-e96c7cee4862-serving-cert\") pod \"route-controller-manager-6576b87f9c-dv9nt\" (UID: \"ba27af47-1944-4970-a015-e96c7cee4862\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dv9nt" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.791804 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9f6c1e2f-b7e7-40e7-92c7-8ca1b57eb9f1-serving-cert\") pod \"authentication-operator-69f744f599-tpbzb\" (UID: \"9f6c1e2f-b7e7-40e7-92c7-8ca1b57eb9f1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tpbzb" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.791836 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/518db853-3bbe-421b-a2dd-6d1b2e965120-config\") pod \"controller-manager-879f6c89f-58htt\" (UID: \"518db853-3bbe-421b-a2dd-6d1b2e965120\") " pod="openshift-controller-manager/controller-manager-879f6c89f-58htt" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.791854 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba27af47-1944-4970-a015-e96c7cee4862-config\") pod \"route-controller-manager-6576b87f9c-dv9nt\" (UID: \"ba27af47-1944-4970-a015-e96c7cee4862\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dv9nt" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.791873 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/88feae6a-5031-48ed-93ec-34ef8d2eff40-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-mbn4k\" (UID: \"88feae6a-5031-48ed-93ec-34ef8d2eff40\") " pod="openshift-authentication/oauth-openshift-558db77b4-mbn4k" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.791891 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8234c8fc-a19c-45c6-8d25-e65668a244ea-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-wxp2r\" (UID: \"8234c8fc-a19c-45c6-8d25-e65668a244ea\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wxp2r" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.791907 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0e82d98b-7c99-496d-910d-e9bfae2ca6ec-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-6xxs8\" (UID: \"0e82d98b-7c99-496d-910d-e9bfae2ca6ec\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6xxs8" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.791927 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gzl9t\" (UniqueName: \"kubernetes.io/projected/ba27af47-1944-4970-a015-e96c7cee4862-kube-api-access-gzl9t\") pod \"route-controller-manager-6576b87f9c-dv9nt\" (UID: \"ba27af47-1944-4970-a015-e96c7cee4862\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dv9nt" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.791944 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5a7c1a05-b29c-4ff3-a4d6-90fe59eb1a01-serving-cert\") pod \"openshift-config-operator-7777fb866f-q8tq8\" (UID: \"5a7c1a05-b29c-4ff3-a4d6-90fe59eb1a01\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-q8tq8" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.791965 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/513166e2-0ac9-4299-bf84-d994a46da192-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-k622z\" (UID: \"513166e2-0ac9-4299-bf84-d994a46da192\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k622z" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.791985 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wpnl\" (UniqueName: \"kubernetes.io/projected/59ec70c2-c208-4529-a763-a0d7ea7e1dfc-kube-api-access-9wpnl\") pod \"machine-approver-56656f9798-86cnx\" (UID: \"59ec70c2-c208-4529-a763-a0d7ea7e1dfc\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-86cnx" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.792001 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/2c2ecf12-49bd-4baf-90df-68486f27be1b-profile-collector-cert\") pod \"catalog-operator-68c6474976-4kjbr\" (UID: \"2c2ecf12-49bd-4baf-90df-68486f27be1b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4kjbr" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.792018 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/88725123-3da9-4127-9506-80c90c9accd5-oauth-serving-cert\") pod \"console-f9d7485db-hf49d\" (UID: \"88725123-3da9-4127-9506-80c90c9accd5\") " pod="openshift-console/console-f9d7485db-hf49d" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.792036 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/88feae6a-5031-48ed-93ec-34ef8d2eff40-audit-policies\") pod \"oauth-openshift-558db77b4-mbn4k\" (UID: \"88feae6a-5031-48ed-93ec-34ef8d2eff40\") " pod="openshift-authentication/oauth-openshift-558db77b4-mbn4k" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.792052 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/88feae6a-5031-48ed-93ec-34ef8d2eff40-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-mbn4k\" (UID: \"88feae6a-5031-48ed-93ec-34ef8d2eff40\") " pod="openshift-authentication/oauth-openshift-558db77b4-mbn4k" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.792067 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/24708c13-efe2-435b-acde-b29b48871f16-images\") pod \"machine-api-operator-5694c8668f-zj4s2\" (UID: \"24708c13-efe2-435b-acde-b29b48871f16\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zj4s2" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.792085 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-trdqk\" (UniqueName: \"kubernetes.io/projected/518db853-3bbe-421b-a2dd-6d1b2e965120-kube-api-access-trdqk\") pod \"controller-manager-879f6c89f-58htt\" (UID: \"518db853-3bbe-421b-a2dd-6d1b2e965120\") " pod="openshift-controller-manager/controller-manager-879f6c89f-58htt" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.792101 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/88feae6a-5031-48ed-93ec-34ef8d2eff40-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-mbn4k\" (UID: \"88feae6a-5031-48ed-93ec-34ef8d2eff40\") " pod="openshift-authentication/oauth-openshift-558db77b4-mbn4k" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.794712 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/518db853-3bbe-421b-a2dd-6d1b2e965120-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-58htt\" (UID: \"518db853-3bbe-421b-a2dd-6d1b2e965120\") " pod="openshift-controller-manager/controller-manager-879f6c89f-58htt" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.795355 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/88feae6a-5031-48ed-93ec-34ef8d2eff40-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-mbn4k\" (UID: \"88feae6a-5031-48ed-93ec-34ef8d2eff40\") " pod="openshift-authentication/oauth-openshift-558db77b4-mbn4k" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.796031 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/88725123-3da9-4127-9506-80c90c9accd5-console-config\") pod \"console-f9d7485db-hf49d\" (UID: \"88725123-3da9-4127-9506-80c90c9accd5\") " pod="openshift-console/console-f9d7485db-hf49d" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.798648 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/88725123-3da9-4127-9506-80c90c9accd5-service-ca\") pod \"console-f9d7485db-hf49d\" (UID: \"88725123-3da9-4127-9506-80c90c9accd5\") " pod="openshift-console/console-f9d7485db-hf49d" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.800149 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba27af47-1944-4970-a015-e96c7cee4862-config\") pod \"route-controller-manager-6576b87f9c-dv9nt\" (UID: \"ba27af47-1944-4970-a015-e96c7cee4862\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dv9nt" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.801599 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/88725123-3da9-4127-9506-80c90c9accd5-trusted-ca-bundle\") pod \"console-f9d7485db-hf49d\" (UID: \"88725123-3da9-4127-9506-80c90c9accd5\") " pod="openshift-console/console-f9d7485db-hf49d" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.803263 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/88feae6a-5031-48ed-93ec-34ef8d2eff40-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-mbn4k\" (UID: \"88feae6a-5031-48ed-93ec-34ef8d2eff40\") " pod="openshift-authentication/oauth-openshift-558db77b4-mbn4k" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.803504 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/5a7c1a05-b29c-4ff3-a4d6-90fe59eb1a01-available-featuregates\") pod \"openshift-config-operator-7777fb866f-q8tq8\" (UID: \"5a7c1a05-b29c-4ff3-a4d6-90fe59eb1a01\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-q8tq8" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.803589 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.804457 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb0f66e4-2c0f-4a95-89db-1aef53ae5b3b-config\") pod \"openshift-apiserver-operator-796bbdcf4f-5m5gj\" (UID: \"cb0f66e4-2c0f-4a95-89db-1aef53ae5b3b\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5m5gj" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.804818 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/88725123-3da9-4127-9506-80c90c9accd5-oauth-serving-cert\") pod \"console-f9d7485db-hf49d\" (UID: \"88725123-3da9-4127-9506-80c90c9accd5\") " pod="openshift-console/console-f9d7485db-hf49d" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.805987 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/88feae6a-5031-48ed-93ec-34ef8d2eff40-audit-dir\") pod \"oauth-openshift-558db77b4-mbn4k\" (UID: \"88feae6a-5031-48ed-93ec-34ef8d2eff40\") " pod="openshift-authentication/oauth-openshift-558db77b4-mbn4k" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.806075 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/518db853-3bbe-421b-a2dd-6d1b2e965120-client-ca\") pod \"controller-manager-879f6c89f-58htt\" (UID: \"518db853-3bbe-421b-a2dd-6d1b2e965120\") " pod="openshift-controller-manager/controller-manager-879f6c89f-58htt" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.809172 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/59ec70c2-c208-4529-a763-a0d7ea7e1dfc-auth-proxy-config\") pod \"machine-approver-56656f9798-86cnx\" (UID: \"59ec70c2-c208-4529-a763-a0d7ea7e1dfc\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-86cnx" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.809710 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/88725123-3da9-4127-9506-80c90c9accd5-console-serving-cert\") pod \"console-f9d7485db-hf49d\" (UID: \"88725123-3da9-4127-9506-80c90c9accd5\") " pod="openshift-console/console-f9d7485db-hf49d" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.809896 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/88feae6a-5031-48ed-93ec-34ef8d2eff40-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-mbn4k\" (UID: \"88feae6a-5031-48ed-93ec-34ef8d2eff40\") " pod="openshift-authentication/oauth-openshift-558db77b4-mbn4k" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.810500 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/518db853-3bbe-421b-a2dd-6d1b2e965120-serving-cert\") pod \"controller-manager-879f6c89f-58htt\" (UID: \"518db853-3bbe-421b-a2dd-6d1b2e965120\") " pod="openshift-controller-manager/controller-manager-879f6c89f-58htt" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.811108 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/88725123-3da9-4127-9506-80c90c9accd5-console-oauth-config\") pod \"console-f9d7485db-hf49d\" (UID: \"88725123-3da9-4127-9506-80c90c9accd5\") " pod="openshift-console/console-f9d7485db-hf49d" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.811726 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/88feae6a-5031-48ed-93ec-34ef8d2eff40-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-mbn4k\" (UID: \"88feae6a-5031-48ed-93ec-34ef8d2eff40\") " pod="openshift-authentication/oauth-openshift-558db77b4-mbn4k" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.812655 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/88feae6a-5031-48ed-93ec-34ef8d2eff40-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-mbn4k\" (UID: \"88feae6a-5031-48ed-93ec-34ef8d2eff40\") " pod="openshift-authentication/oauth-openshift-558db77b4-mbn4k" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.813349 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ba27af47-1944-4970-a015-e96c7cee4862-client-ca\") pod \"route-controller-manager-6576b87f9c-dv9nt\" (UID: \"ba27af47-1944-4970-a015-e96c7cee4862\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dv9nt" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.814105 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5a7c1a05-b29c-4ff3-a4d6-90fe59eb1a01-serving-cert\") pod \"openshift-config-operator-7777fb866f-q8tq8\" (UID: \"5a7c1a05-b29c-4ff3-a4d6-90fe59eb1a01\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-q8tq8" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.814718 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/88feae6a-5031-48ed-93ec-34ef8d2eff40-audit-policies\") pod \"oauth-openshift-558db77b4-mbn4k\" (UID: \"88feae6a-5031-48ed-93ec-34ef8d2eff40\") " pod="openshift-authentication/oauth-openshift-558db77b4-mbn4k" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.815429 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/518db853-3bbe-421b-a2dd-6d1b2e965120-config\") pod \"controller-manager-879f6c89f-58htt\" (UID: \"518db853-3bbe-421b-a2dd-6d1b2e965120\") " pod="openshift-controller-manager/controller-manager-879f6c89f-58htt" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.815718 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b75a7773-a689-46c1-b98e-e7c1be0878e6-config\") pod \"console-operator-58897d9998-t5hm7\" (UID: \"b75a7773-a689-46c1-b98e-e7c1be0878e6\") " pod="openshift-console-operator/console-operator-58897d9998-t5hm7" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.816914 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/88feae6a-5031-48ed-93ec-34ef8d2eff40-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-mbn4k\" (UID: \"88feae6a-5031-48ed-93ec-34ef8d2eff40\") " pod="openshift-authentication/oauth-openshift-558db77b4-mbn4k" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.816947 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b75a7773-a689-46c1-b98e-e7c1be0878e6-trusted-ca\") pod \"console-operator-58897d9998-t5hm7\" (UID: \"b75a7773-a689-46c1-b98e-e7c1be0878e6\") " pod="openshift-console-operator/console-operator-58897d9998-t5hm7" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.818709 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/88feae6a-5031-48ed-93ec-34ef8d2eff40-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-mbn4k\" (UID: \"88feae6a-5031-48ed-93ec-34ef8d2eff40\") " pod="openshift-authentication/oauth-openshift-558db77b4-mbn4k" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.819533 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b75a7773-a689-46c1-b98e-e7c1be0878e6-serving-cert\") pod \"console-operator-58897d9998-t5hm7\" (UID: \"b75a7773-a689-46c1-b98e-e7c1be0878e6\") " pod="openshift-console-operator/console-operator-58897d9998-t5hm7" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.819686 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.819928 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/88feae6a-5031-48ed-93ec-34ef8d2eff40-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-mbn4k\" (UID: \"88feae6a-5031-48ed-93ec-34ef8d2eff40\") " pod="openshift-authentication/oauth-openshift-558db77b4-mbn4k" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.821985 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ba27af47-1944-4970-a015-e96c7cee4862-serving-cert\") pod \"route-controller-manager-6576b87f9c-dv9nt\" (UID: \"ba27af47-1944-4970-a015-e96c7cee4862\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dv9nt" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.822573 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cb0f66e4-2c0f-4a95-89db-1aef53ae5b3b-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-5m5gj\" (UID: \"cb0f66e4-2c0f-4a95-89db-1aef53ae5b3b\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5m5gj" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.823018 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/59ec70c2-c208-4529-a763-a0d7ea7e1dfc-machine-approver-tls\") pod \"machine-approver-56656f9798-86cnx\" (UID: \"59ec70c2-c208-4529-a763-a0d7ea7e1dfc\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-86cnx" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.824101 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/88feae6a-5031-48ed-93ec-34ef8d2eff40-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-mbn4k\" (UID: \"88feae6a-5031-48ed-93ec-34ef8d2eff40\") " pod="openshift-authentication/oauth-openshift-558db77b4-mbn4k" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.824603 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/88feae6a-5031-48ed-93ec-34ef8d2eff40-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-mbn4k\" (UID: \"88feae6a-5031-48ed-93ec-34ef8d2eff40\") " pod="openshift-authentication/oauth-openshift-558db77b4-mbn4k" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.825234 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/88feae6a-5031-48ed-93ec-34ef8d2eff40-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-mbn4k\" (UID: \"88feae6a-5031-48ed-93ec-34ef8d2eff40\") " pod="openshift-authentication/oauth-openshift-558db77b4-mbn4k" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.826378 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401110-dm86l"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.826917 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-898jb"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.827104 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-f2tjh" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.827458 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401110-dm86l" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.828189 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-t5hm7"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.828218 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-s7j8m"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.828768 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2stv6"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.829359 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-898jb" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.829414 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s7j8m" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.830496 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-v6pjk"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.830925 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-mbn4k"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.830956 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-hjvt9"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.830969 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-zj4s2"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.830982 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-dv9nt"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.830993 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5m5gj"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.831142 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-hddc7"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.831888 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2stv6" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.831971 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zvztq"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.832222 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-p2jmh"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.832279 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-hf49d"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.832295 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-xnlvh"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.832308 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-58htt"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.832321 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-rx6sh"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.831992 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-v6pjk" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.837699 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.839599 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-tsb45"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.840006 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-rx6sh" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.842792 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d8x5t"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.843026 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-q8tq8"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.843090 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-v88lg"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.843115 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wxp2r"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.843129 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-jb5sz"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.843141 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-kc2wp"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.842937 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-tsb45" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.843759 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-whcwc"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.843785 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tkgh9"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.843795 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6xxs8"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.843804 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k622z"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.843813 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-tpbzb"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.845336 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-kc2wp" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.847255 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-n4lbr"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.849616 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-w5k28"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.851351 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-prvdm"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.853183 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-898jb"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.855668 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-7792m"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.859336 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401110-dm86l"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.862379 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4kjbr"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.868970 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-kps6n"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.873588 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-ng2nv"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.873623 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-dxf4x"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.875487 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-f2tjh"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.876882 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2stv6"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.877778 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.878031 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-v6pjk"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.879231 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-rx6sh"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.880377 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-s7j8m"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.881721 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-tsb45"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.883003 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-cx6h6"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.883842 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-cx6h6" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.884027 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-cx6h6"] Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.893146 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/064de93d-7bb5-47b5-9585-bf32be9a05f6-etcd-serving-ca\") pod \"apiserver-76f77b778f-jb5sz\" (UID: \"064de93d-7bb5-47b5-9585-bf32be9a05f6\") " pod="openshift-apiserver/apiserver-76f77b778f-jb5sz" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.893198 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4878872d-8035-442a-a355-7c966f37e8c1-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-n4lbr\" (UID: \"4878872d-8035-442a-a355-7c966f37e8c1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-n4lbr" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.893219 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/d7e79465-8df8-41a7-ab16-5c730f9ceca2-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-f2tjh\" (UID: \"d7e79465-8df8-41a7-ab16-5c730f9ceca2\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-f2tjh" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.893239 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gfvzz\" (UniqueName: \"kubernetes.io/projected/4a74ce9e-b145-4e40-ad4e-67ecdf722195-kube-api-access-gfvzz\") pod \"control-plane-machine-set-operator-78cbb6b69f-2stv6\" (UID: \"4a74ce9e-b145-4e40-ad4e-67ecdf722195\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2stv6" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.893255 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/064de93d-7bb5-47b5-9585-bf32be9a05f6-audit-dir\") pod \"apiserver-76f77b778f-jb5sz\" (UID: \"064de93d-7bb5-47b5-9585-bf32be9a05f6\") " pod="openshift-apiserver/apiserver-76f77b778f-jb5sz" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.893270 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/2c2ecf12-49bd-4baf-90df-68486f27be1b-srv-cert\") pod \"catalog-operator-68c6474976-4kjbr\" (UID: \"2c2ecf12-49bd-4baf-90df-68486f27be1b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4kjbr" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.893286 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7fe78aa4-571c-488b-96bd-23ae4b6333db-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-zvztq\" (UID: \"7fe78aa4-571c-488b-96bd-23ae4b6333db\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zvztq" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.893299 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/155595fc-76b3-4561-bd3f-c540076d4cd6-srv-cert\") pod \"olm-operator-6b444d44fb-v88lg\" (UID: \"155595fc-76b3-4561-bd3f-c540076d4cd6\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-v88lg" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.893564 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/064de93d-7bb5-47b5-9585-bf32be9a05f6-audit-dir\") pod \"apiserver-76f77b778f-jb5sz\" (UID: \"064de93d-7bb5-47b5-9585-bf32be9a05f6\") " pod="openshift-apiserver/apiserver-76f77b778f-jb5sz" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.893568 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4878872d-8035-442a-a355-7c966f37e8c1-config\") pod \"kube-apiserver-operator-766d6c64bb-n4lbr\" (UID: \"4878872d-8035-442a-a355-7c966f37e8c1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-n4lbr" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.893743 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bdd88e7d-edf6-41c1-a5a3-47746bb3ab99-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-whcwc\" (UID: \"bdd88e7d-edf6-41c1-a5a3-47746bb3ab99\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-whcwc" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.893771 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tnlzh\" (UniqueName: \"kubernetes.io/projected/24708c13-efe2-435b-acde-b29b48871f16-kube-api-access-tnlzh\") pod \"machine-api-operator-5694c8668f-zj4s2\" (UID: \"24708c13-efe2-435b-acde-b29b48871f16\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zj4s2" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.893787 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rf8tm\" (UniqueName: \"kubernetes.io/projected/155595fc-76b3-4561-bd3f-c540076d4cd6-kube-api-access-rf8tm\") pod \"olm-operator-6b444d44fb-v88lg\" (UID: \"155595fc-76b3-4561-bd3f-c540076d4cd6\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-v88lg" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.893803 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bdd88e7d-edf6-41c1-a5a3-47746bb3ab99-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-whcwc\" (UID: \"bdd88e7d-edf6-41c1-a5a3-47746bb3ab99\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-whcwc" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.893822 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5ed0f2b3-fd6d-4892-8198-2b81d5ea1a7f-trusted-ca\") pod \"ingress-operator-5b745b69d9-ng2nv\" (UID: \"5ed0f2b3-fd6d-4892-8198-2b81d5ea1a7f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ng2nv" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.893845 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/590f742c-e0d3-4394-8ef3-08bde5d57d01-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-dxf4x\" (UID: \"590f742c-e0d3-4394-8ef3-08bde5d57d01\") " pod="openshift-marketplace/marketplace-operator-79b997595-dxf4x" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.893896 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dt66w\" (UniqueName: \"kubernetes.io/projected/7fe78aa4-571c-488b-96bd-23ae4b6333db-kube-api-access-dt66w\") pod \"openshift-controller-manager-operator-756b6f6bc6-zvztq\" (UID: \"7fe78aa4-571c-488b-96bd-23ae4b6333db\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zvztq" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.893912 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/4a74ce9e-b145-4e40-ad4e-67ecdf722195-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-2stv6\" (UID: \"4a74ce9e-b145-4e40-ad4e-67ecdf722195\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2stv6" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.893929 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9f6c1e2f-b7e7-40e7-92c7-8ca1b57eb9f1-serving-cert\") pod \"authentication-operator-69f744f599-tpbzb\" (UID: \"9f6c1e2f-b7e7-40e7-92c7-8ca1b57eb9f1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tpbzb" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.893951 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r2cg5\" (UniqueName: \"kubernetes.io/projected/2d3eed24-29f1-414a-a5d8-f8b021e6aff4-kube-api-access-r2cg5\") pod \"cluster-image-registry-operator-dc59b4c8b-tkgh9\" (UID: \"2d3eed24-29f1-414a-a5d8-f8b021e6aff4\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tkgh9" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.893997 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8234c8fc-a19c-45c6-8d25-e65668a244ea-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-wxp2r\" (UID: \"8234c8fc-a19c-45c6-8d25-e65668a244ea\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wxp2r" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.894022 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0e82d98b-7c99-496d-910d-e9bfae2ca6ec-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-6xxs8\" (UID: \"0e82d98b-7c99-496d-910d-e9bfae2ca6ec\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6xxs8" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.894206 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/513166e2-0ac9-4299-bf84-d994a46da192-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-k622z\" (UID: \"513166e2-0ac9-4299-bf84-d994a46da192\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k622z" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.894287 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xgpbf\" (UniqueName: \"kubernetes.io/projected/d7e79465-8df8-41a7-ab16-5c730f9ceca2-kube-api-access-xgpbf\") pod \"package-server-manager-789f6589d5-f2tjh\" (UID: \"d7e79465-8df8-41a7-ab16-5c730f9ceca2\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-f2tjh" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.894345 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/590f742c-e0d3-4394-8ef3-08bde5d57d01-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-dxf4x\" (UID: \"590f742c-e0d3-4394-8ef3-08bde5d57d01\") " pod="openshift-marketplace/marketplace-operator-79b997595-dxf4x" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.894367 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2nzbr\" (UniqueName: \"kubernetes.io/projected/499ccac5-6595-4d67-93ed-dd2d19d045bf-kube-api-access-2nzbr\") pod \"service-ca-9c57cc56f-7792m\" (UID: \"499ccac5-6595-4d67-93ed-dd2d19d045bf\") " pod="openshift-service-ca/service-ca-9c57cc56f-7792m" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.894387 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bwgwh\" (UniqueName: \"kubernetes.io/projected/5ed0f2b3-fd6d-4892-8198-2b81d5ea1a7f-kube-api-access-bwgwh\") pod \"ingress-operator-5b745b69d9-ng2nv\" (UID: \"5ed0f2b3-fd6d-4892-8198-2b81d5ea1a7f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ng2nv" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.894414 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/24708c13-efe2-435b-acde-b29b48871f16-images\") pod \"machine-api-operator-5694c8668f-zj4s2\" (UID: \"24708c13-efe2-435b-acde-b29b48871f16\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zj4s2" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.894436 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/2c2ecf12-49bd-4baf-90df-68486f27be1b-profile-collector-cert\") pod \"catalog-operator-68c6474976-4kjbr\" (UID: \"2c2ecf12-49bd-4baf-90df-68486f27be1b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4kjbr" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.894473 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tpsds\" (UniqueName: \"kubernetes.io/projected/513166e2-0ac9-4299-bf84-d994a46da192-kube-api-access-tpsds\") pod \"cluster-samples-operator-665b6dd947-k622z\" (UID: \"513166e2-0ac9-4299-bf84-d994a46da192\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k622z" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.894495 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/064de93d-7bb5-47b5-9585-bf32be9a05f6-config\") pod \"apiserver-76f77b778f-jb5sz\" (UID: \"064de93d-7bb5-47b5-9585-bf32be9a05f6\") " pod="openshift-apiserver/apiserver-76f77b778f-jb5sz" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.894523 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b18917c3-98cb-407b-97e5-5333aa5ca797-webhook-cert\") pod \"packageserver-d55dfcdfc-d8x5t\" (UID: \"b18917c3-98cb-407b-97e5-5333aa5ca797\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d8x5t" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.894560 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/2d3eed24-29f1-414a-a5d8-f8b021e6aff4-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-tkgh9\" (UID: \"2d3eed24-29f1-414a-a5d8-f8b021e6aff4\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tkgh9" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.894597 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/064de93d-7bb5-47b5-9585-bf32be9a05f6-etcd-client\") pod \"apiserver-76f77b778f-jb5sz\" (UID: \"064de93d-7bb5-47b5-9585-bf32be9a05f6\") " pod="openshift-apiserver/apiserver-76f77b778f-jb5sz" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.894600 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/064de93d-7bb5-47b5-9585-bf32be9a05f6-etcd-serving-ca\") pod \"apiserver-76f77b778f-jb5sz\" (UID: \"064de93d-7bb5-47b5-9585-bf32be9a05f6\") " pod="openshift-apiserver/apiserver-76f77b778f-jb5sz" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.894620 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fq9pc\" (UniqueName: \"kubernetes.io/projected/b0e34420-095d-42aa-89c4-5a80331e1c01-kube-api-access-fq9pc\") pod \"dns-operator-744455d44c-hddc7\" (UID: \"b0e34420-095d-42aa-89c4-5a80331e1c01\") " pod="openshift-dns-operator/dns-operator-744455d44c-hddc7" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.894650 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/064de93d-7bb5-47b5-9585-bf32be9a05f6-encryption-config\") pod \"apiserver-76f77b778f-jb5sz\" (UID: \"064de93d-7bb5-47b5-9585-bf32be9a05f6\") " pod="openshift-apiserver/apiserver-76f77b778f-jb5sz" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.894672 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/24708c13-efe2-435b-acde-b29b48871f16-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-zj4s2\" (UID: \"24708c13-efe2-435b-acde-b29b48871f16\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zj4s2" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.894693 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/064de93d-7bb5-47b5-9585-bf32be9a05f6-node-pullsecrets\") pod \"apiserver-76f77b778f-jb5sz\" (UID: \"064de93d-7bb5-47b5-9585-bf32be9a05f6\") " pod="openshift-apiserver/apiserver-76f77b778f-jb5sz" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.894714 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6cq5w\" (UniqueName: \"kubernetes.io/projected/2c2ecf12-49bd-4baf-90df-68486f27be1b-kube-api-access-6cq5w\") pod \"catalog-operator-68c6474976-4kjbr\" (UID: \"2c2ecf12-49bd-4baf-90df-68486f27be1b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4kjbr" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.894750 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7fe78aa4-571c-488b-96bd-23ae4b6333db-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-zvztq\" (UID: \"7fe78aa4-571c-488b-96bd-23ae4b6333db\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zvztq" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.894776 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h5s8b\" (UniqueName: \"kubernetes.io/projected/b18917c3-98cb-407b-97e5-5333aa5ca797-kube-api-access-h5s8b\") pod \"packageserver-d55dfcdfc-d8x5t\" (UID: \"b18917c3-98cb-407b-97e5-5333aa5ca797\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d8x5t" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.894800 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/064de93d-7bb5-47b5-9585-bf32be9a05f6-audit\") pod \"apiserver-76f77b778f-jb5sz\" (UID: \"064de93d-7bb5-47b5-9585-bf32be9a05f6\") " pod="openshift-apiserver/apiserver-76f77b778f-jb5sz" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.894822 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/064de93d-7bb5-47b5-9585-bf32be9a05f6-image-import-ca\") pod \"apiserver-76f77b778f-jb5sz\" (UID: \"064de93d-7bb5-47b5-9585-bf32be9a05f6\") " pod="openshift-apiserver/apiserver-76f77b778f-jb5sz" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.894846 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/064de93d-7bb5-47b5-9585-bf32be9a05f6-serving-cert\") pod \"apiserver-76f77b778f-jb5sz\" (UID: \"064de93d-7bb5-47b5-9585-bf32be9a05f6\") " pod="openshift-apiserver/apiserver-76f77b778f-jb5sz" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.894867 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5ed0f2b3-fd6d-4892-8198-2b81d5ea1a7f-bound-sa-token\") pod \"ingress-operator-5b745b69d9-ng2nv\" (UID: \"5ed0f2b3-fd6d-4892-8198-2b81d5ea1a7f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ng2nv" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.894894 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v86wz\" (UniqueName: \"kubernetes.io/projected/064de93d-7bb5-47b5-9585-bf32be9a05f6-kube-api-access-v86wz\") pod \"apiserver-76f77b778f-jb5sz\" (UID: \"064de93d-7bb5-47b5-9585-bf32be9a05f6\") " pod="openshift-apiserver/apiserver-76f77b778f-jb5sz" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.894919 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8234c8fc-a19c-45c6-8d25-e65668a244ea-config\") pod \"kube-controller-manager-operator-78b949d7b-wxp2r\" (UID: \"8234c8fc-a19c-45c6-8d25-e65668a244ea\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wxp2r" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.894944 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/064de93d-7bb5-47b5-9585-bf32be9a05f6-trusted-ca-bundle\") pod \"apiserver-76f77b778f-jb5sz\" (UID: \"064de93d-7bb5-47b5-9585-bf32be9a05f6\") " pod="openshift-apiserver/apiserver-76f77b778f-jb5sz" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.894965 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/155595fc-76b3-4561-bd3f-c540076d4cd6-profile-collector-cert\") pod \"olm-operator-6b444d44fb-v88lg\" (UID: \"155595fc-76b3-4561-bd3f-c540076d4cd6\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-v88lg" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.894988 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4qsq\" (UniqueName: \"kubernetes.io/projected/402cf28b-25d2-428a-ab7b-1d70345a921e-kube-api-access-r4qsq\") pod \"multus-admission-controller-857f4d67dd-p2jmh\" (UID: \"402cf28b-25d2-428a-ab7b-1d70345a921e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-p2jmh" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.895011 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2d3eed24-29f1-414a-a5d8-f8b021e6aff4-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-tkgh9\" (UID: \"2d3eed24-29f1-414a-a5d8-f8b021e6aff4\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tkgh9" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.895032 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2d3eed24-29f1-414a-a5d8-f8b021e6aff4-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-tkgh9\" (UID: \"2d3eed24-29f1-414a-a5d8-f8b021e6aff4\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tkgh9" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.895056 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9f6c1e2f-b7e7-40e7-92c7-8ca1b57eb9f1-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-tpbzb\" (UID: \"9f6c1e2f-b7e7-40e7-92c7-8ca1b57eb9f1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tpbzb" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.895082 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bdd88e7d-edf6-41c1-a5a3-47746bb3ab99-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-whcwc\" (UID: \"bdd88e7d-edf6-41c1-a5a3-47746bb3ab99\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-whcwc" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.895107 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/24708c13-efe2-435b-acde-b29b48871f16-config\") pod \"machine-api-operator-5694c8668f-zj4s2\" (UID: \"24708c13-efe2-435b-acde-b29b48871f16\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zj4s2" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.895123 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/24708c13-efe2-435b-acde-b29b48871f16-images\") pod \"machine-api-operator-5694c8668f-zj4s2\" (UID: \"24708c13-efe2-435b-acde-b29b48871f16\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zj4s2" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.895128 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/402cf28b-25d2-428a-ab7b-1d70345a921e-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-p2jmh\" (UID: \"402cf28b-25d2-428a-ab7b-1d70345a921e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-p2jmh" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.895198 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/b18917c3-98cb-407b-97e5-5333aa5ca797-tmpfs\") pod \"packageserver-d55dfcdfc-d8x5t\" (UID: \"b18917c3-98cb-407b-97e5-5333aa5ca797\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d8x5t" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.895215 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b18917c3-98cb-407b-97e5-5333aa5ca797-apiservice-cert\") pod \"packageserver-d55dfcdfc-d8x5t\" (UID: \"b18917c3-98cb-407b-97e5-5333aa5ca797\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d8x5t" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.894414 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7fe78aa4-571c-488b-96bd-23ae4b6333db-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-zvztq\" (UID: \"7fe78aa4-571c-488b-96bd-23ae4b6333db\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zvztq" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.895232 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9pxr7\" (UniqueName: \"kubernetes.io/projected/590f742c-e0d3-4394-8ef3-08bde5d57d01-kube-api-access-9pxr7\") pod \"marketplace-operator-79b997595-dxf4x\" (UID: \"590f742c-e0d3-4394-8ef3-08bde5d57d01\") " pod="openshift-marketplace/marketplace-operator-79b997595-dxf4x" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.895266 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f6c1e2f-b7e7-40e7-92c7-8ca1b57eb9f1-config\") pod \"authentication-operator-69f744f599-tpbzb\" (UID: \"9f6c1e2f-b7e7-40e7-92c7-8ca1b57eb9f1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tpbzb" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.895284 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9f6c1e2f-b7e7-40e7-92c7-8ca1b57eb9f1-service-ca-bundle\") pod \"authentication-operator-69f744f599-tpbzb\" (UID: \"9f6c1e2f-b7e7-40e7-92c7-8ca1b57eb9f1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tpbzb" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.895300 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4878872d-8035-442a-a355-7c966f37e8c1-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-n4lbr\" (UID: \"4878872d-8035-442a-a355-7c966f37e8c1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-n4lbr" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.895317 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-44zw6\" (UniqueName: \"kubernetes.io/projected/f954533e-e8a0-4899-be72-241760feae95-kube-api-access-44zw6\") pod \"downloads-7954f5f757-hjvt9\" (UID: \"f954533e-e8a0-4899-be72-241760feae95\") " pod="openshift-console/downloads-7954f5f757-hjvt9" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.895349 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kbrfs\" (UniqueName: \"kubernetes.io/projected/9f6c1e2f-b7e7-40e7-92c7-8ca1b57eb9f1-kube-api-access-kbrfs\") pod \"authentication-operator-69f744f599-tpbzb\" (UID: \"9f6c1e2f-b7e7-40e7-92c7-8ca1b57eb9f1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tpbzb" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.895372 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8234c8fc-a19c-45c6-8d25-e65668a244ea-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-wxp2r\" (UID: \"8234c8fc-a19c-45c6-8d25-e65668a244ea\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wxp2r" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.895391 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b0e34420-095d-42aa-89c4-5a80331e1c01-metrics-tls\") pod \"dns-operator-744455d44c-hddc7\" (UID: \"b0e34420-095d-42aa-89c4-5a80331e1c01\") " pod="openshift-dns-operator/dns-operator-744455d44c-hddc7" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.895409 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e82d98b-7c99-496d-910d-e9bfae2ca6ec-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-6xxs8\" (UID: \"0e82d98b-7c99-496d-910d-e9bfae2ca6ec\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6xxs8" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.895428 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/499ccac5-6595-4d67-93ed-dd2d19d045bf-signing-cabundle\") pod \"service-ca-9c57cc56f-7792m\" (UID: \"499ccac5-6595-4d67-93ed-dd2d19d045bf\") " pod="openshift-service-ca/service-ca-9c57cc56f-7792m" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.895445 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5ed0f2b3-fd6d-4892-8198-2b81d5ea1a7f-metrics-tls\") pod \"ingress-operator-5b745b69d9-ng2nv\" (UID: \"5ed0f2b3-fd6d-4892-8198-2b81d5ea1a7f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ng2nv" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.895466 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rc8sv\" (UniqueName: \"kubernetes.io/projected/0e82d98b-7c99-496d-910d-e9bfae2ca6ec-kube-api-access-rc8sv\") pod \"kube-storage-version-migrator-operator-b67b599dd-6xxs8\" (UID: \"0e82d98b-7c99-496d-910d-e9bfae2ca6ec\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6xxs8" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.895483 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/499ccac5-6595-4d67-93ed-dd2d19d045bf-signing-key\") pod \"service-ca-9c57cc56f-7792m\" (UID: \"499ccac5-6595-4d67-93ed-dd2d19d045bf\") " pod="openshift-service-ca/service-ca-9c57cc56f-7792m" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.895134 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/064de93d-7bb5-47b5-9585-bf32be9a05f6-config\") pod \"apiserver-76f77b778f-jb5sz\" (UID: \"064de93d-7bb5-47b5-9585-bf32be9a05f6\") " pod="openshift-apiserver/apiserver-76f77b778f-jb5sz" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.895567 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bdd88e7d-edf6-41c1-a5a3-47746bb3ab99-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-whcwc\" (UID: \"bdd88e7d-edf6-41c1-a5a3-47746bb3ab99\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-whcwc" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.895978 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/064de93d-7bb5-47b5-9585-bf32be9a05f6-node-pullsecrets\") pod \"apiserver-76f77b778f-jb5sz\" (UID: \"064de93d-7bb5-47b5-9585-bf32be9a05f6\") " pod="openshift-apiserver/apiserver-76f77b778f-jb5sz" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.896890 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9f6c1e2f-b7e7-40e7-92c7-8ca1b57eb9f1-service-ca-bundle\") pod \"authentication-operator-69f744f599-tpbzb\" (UID: \"9f6c1e2f-b7e7-40e7-92c7-8ca1b57eb9f1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tpbzb" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.897113 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8234c8fc-a19c-45c6-8d25-e65668a244ea-config\") pod \"kube-controller-manager-operator-78b949d7b-wxp2r\" (UID: \"8234c8fc-a19c-45c6-8d25-e65668a244ea\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wxp2r" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.897990 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f6c1e2f-b7e7-40e7-92c7-8ca1b57eb9f1-config\") pod \"authentication-operator-69f744f599-tpbzb\" (UID: \"9f6c1e2f-b7e7-40e7-92c7-8ca1b57eb9f1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tpbzb" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.898377 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9f6c1e2f-b7e7-40e7-92c7-8ca1b57eb9f1-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-tpbzb\" (UID: \"9f6c1e2f-b7e7-40e7-92c7-8ca1b57eb9f1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tpbzb" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.898644 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/155595fc-76b3-4561-bd3f-c540076d4cd6-srv-cert\") pod \"olm-operator-6b444d44fb-v88lg\" (UID: \"155595fc-76b3-4561-bd3f-c540076d4cd6\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-v88lg" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.898799 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9f6c1e2f-b7e7-40e7-92c7-8ca1b57eb9f1-serving-cert\") pod \"authentication-operator-69f744f599-tpbzb\" (UID: \"9f6c1e2f-b7e7-40e7-92c7-8ca1b57eb9f1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tpbzb" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.899053 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/24708c13-efe2-435b-acde-b29b48871f16-config\") pod \"machine-api-operator-5694c8668f-zj4s2\" (UID: \"24708c13-efe2-435b-acde-b29b48871f16\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zj4s2" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.899088 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bdd88e7d-edf6-41c1-a5a3-47746bb3ab99-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-whcwc\" (UID: \"bdd88e7d-edf6-41c1-a5a3-47746bb3ab99\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-whcwc" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.899340 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/064de93d-7bb5-47b5-9585-bf32be9a05f6-image-import-ca\") pod \"apiserver-76f77b778f-jb5sz\" (UID: \"064de93d-7bb5-47b5-9585-bf32be9a05f6\") " pod="openshift-apiserver/apiserver-76f77b778f-jb5sz" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.899468 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/064de93d-7bb5-47b5-9585-bf32be9a05f6-encryption-config\") pod \"apiserver-76f77b778f-jb5sz\" (UID: \"064de93d-7bb5-47b5-9585-bf32be9a05f6\") " pod="openshift-apiserver/apiserver-76f77b778f-jb5sz" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.899623 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/064de93d-7bb5-47b5-9585-bf32be9a05f6-etcd-client\") pod \"apiserver-76f77b778f-jb5sz\" (UID: \"064de93d-7bb5-47b5-9585-bf32be9a05f6\") " pod="openshift-apiserver/apiserver-76f77b778f-jb5sz" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.899825 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/064de93d-7bb5-47b5-9585-bf32be9a05f6-audit\") pod \"apiserver-76f77b778f-jb5sz\" (UID: \"064de93d-7bb5-47b5-9585-bf32be9a05f6\") " pod="openshift-apiserver/apiserver-76f77b778f-jb5sz" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.899871 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/2c2ecf12-49bd-4baf-90df-68486f27be1b-profile-collector-cert\") pod \"catalog-operator-68c6474976-4kjbr\" (UID: \"2c2ecf12-49bd-4baf-90df-68486f27be1b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4kjbr" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.900282 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.900446 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/064de93d-7bb5-47b5-9585-bf32be9a05f6-trusted-ca-bundle\") pod \"apiserver-76f77b778f-jb5sz\" (UID: \"064de93d-7bb5-47b5-9585-bf32be9a05f6\") " pod="openshift-apiserver/apiserver-76f77b778f-jb5sz" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.901084 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7fe78aa4-571c-488b-96bd-23ae4b6333db-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-zvztq\" (UID: \"7fe78aa4-571c-488b-96bd-23ae4b6333db\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zvztq" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.901091 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/064de93d-7bb5-47b5-9585-bf32be9a05f6-serving-cert\") pod \"apiserver-76f77b778f-jb5sz\" (UID: \"064de93d-7bb5-47b5-9585-bf32be9a05f6\") " pod="openshift-apiserver/apiserver-76f77b778f-jb5sz" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.901769 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/2c2ecf12-49bd-4baf-90df-68486f27be1b-srv-cert\") pod \"catalog-operator-68c6474976-4kjbr\" (UID: \"2c2ecf12-49bd-4baf-90df-68486f27be1b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4kjbr" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.903886 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/402cf28b-25d2-428a-ab7b-1d70345a921e-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-p2jmh\" (UID: \"402cf28b-25d2-428a-ab7b-1d70345a921e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-p2jmh" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.903978 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8234c8fc-a19c-45c6-8d25-e65668a244ea-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-wxp2r\" (UID: \"8234c8fc-a19c-45c6-8d25-e65668a244ea\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wxp2r" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.904051 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/24708c13-efe2-435b-acde-b29b48871f16-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-zj4s2\" (UID: \"24708c13-efe2-435b-acde-b29b48871f16\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zj4s2" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.904064 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/155595fc-76b3-4561-bd3f-c540076d4cd6-profile-collector-cert\") pod \"olm-operator-6b444d44fb-v88lg\" (UID: \"155595fc-76b3-4561-bd3f-c540076d4cd6\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-v88lg" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.904053 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/513166e2-0ac9-4299-bf84-d994a46da192-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-k622z\" (UID: \"513166e2-0ac9-4299-bf84-d994a46da192\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k622z" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.904263 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b0e34420-095d-42aa-89c4-5a80331e1c01-metrics-tls\") pod \"dns-operator-744455d44c-hddc7\" (UID: \"b0e34420-095d-42aa-89c4-5a80331e1c01\") " pod="openshift-dns-operator/dns-operator-744455d44c-hddc7" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.918827 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.928281 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4878872d-8035-442a-a355-7c966f37e8c1-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-n4lbr\" (UID: \"4878872d-8035-442a-a355-7c966f37e8c1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-n4lbr" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.937717 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.944409 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4878872d-8035-442a-a355-7c966f37e8c1-config\") pod \"kube-apiserver-operator-766d6c64bb-n4lbr\" (UID: \"4878872d-8035-442a-a355-7c966f37e8c1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-n4lbr" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.958612 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.978077 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.995988 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5ed0f2b3-fd6d-4892-8198-2b81d5ea1a7f-trusted-ca\") pod \"ingress-operator-5b745b69d9-ng2nv\" (UID: \"5ed0f2b3-fd6d-4892-8198-2b81d5ea1a7f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ng2nv" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.996035 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/4a74ce9e-b145-4e40-ad4e-67ecdf722195-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-2stv6\" (UID: \"4a74ce9e-b145-4e40-ad4e-67ecdf722195\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2stv6" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.996053 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/590f742c-e0d3-4394-8ef3-08bde5d57d01-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-dxf4x\" (UID: \"590f742c-e0d3-4394-8ef3-08bde5d57d01\") " pod="openshift-marketplace/marketplace-operator-79b997595-dxf4x" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.996073 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r2cg5\" (UniqueName: \"kubernetes.io/projected/2d3eed24-29f1-414a-a5d8-f8b021e6aff4-kube-api-access-r2cg5\") pod \"cluster-image-registry-operator-dc59b4c8b-tkgh9\" (UID: \"2d3eed24-29f1-414a-a5d8-f8b021e6aff4\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tkgh9" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.996114 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xgpbf\" (UniqueName: \"kubernetes.io/projected/d7e79465-8df8-41a7-ab16-5c730f9ceca2-kube-api-access-xgpbf\") pod \"package-server-manager-789f6589d5-f2tjh\" (UID: \"d7e79465-8df8-41a7-ab16-5c730f9ceca2\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-f2tjh" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.996129 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/590f742c-e0d3-4394-8ef3-08bde5d57d01-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-dxf4x\" (UID: \"590f742c-e0d3-4394-8ef3-08bde5d57d01\") " pod="openshift-marketplace/marketplace-operator-79b997595-dxf4x" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.997491 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2nzbr\" (UniqueName: \"kubernetes.io/projected/499ccac5-6595-4d67-93ed-dd2d19d045bf-kube-api-access-2nzbr\") pod \"service-ca-9c57cc56f-7792m\" (UID: \"499ccac5-6595-4d67-93ed-dd2d19d045bf\") " pod="openshift-service-ca/service-ca-9c57cc56f-7792m" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.997548 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bwgwh\" (UniqueName: \"kubernetes.io/projected/5ed0f2b3-fd6d-4892-8198-2b81d5ea1a7f-kube-api-access-bwgwh\") pod \"ingress-operator-5b745b69d9-ng2nv\" (UID: \"5ed0f2b3-fd6d-4892-8198-2b81d5ea1a7f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ng2nv" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.997582 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.997601 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b18917c3-98cb-407b-97e5-5333aa5ca797-webhook-cert\") pod \"packageserver-d55dfcdfc-d8x5t\" (UID: \"b18917c3-98cb-407b-97e5-5333aa5ca797\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d8x5t" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.997680 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/2d3eed24-29f1-414a-a5d8-f8b021e6aff4-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-tkgh9\" (UID: \"2d3eed24-29f1-414a-a5d8-f8b021e6aff4\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tkgh9" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.997752 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h5s8b\" (UniqueName: \"kubernetes.io/projected/b18917c3-98cb-407b-97e5-5333aa5ca797-kube-api-access-h5s8b\") pod \"packageserver-d55dfcdfc-d8x5t\" (UID: \"b18917c3-98cb-407b-97e5-5333aa5ca797\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d8x5t" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.997839 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5ed0f2b3-fd6d-4892-8198-2b81d5ea1a7f-bound-sa-token\") pod \"ingress-operator-5b745b69d9-ng2nv\" (UID: \"5ed0f2b3-fd6d-4892-8198-2b81d5ea1a7f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ng2nv" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.997880 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2d3eed24-29f1-414a-a5d8-f8b021e6aff4-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-tkgh9\" (UID: \"2d3eed24-29f1-414a-a5d8-f8b021e6aff4\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tkgh9" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.997911 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2d3eed24-29f1-414a-a5d8-f8b021e6aff4-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-tkgh9\" (UID: \"2d3eed24-29f1-414a-a5d8-f8b021e6aff4\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tkgh9" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.997949 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/b18917c3-98cb-407b-97e5-5333aa5ca797-tmpfs\") pod \"packageserver-d55dfcdfc-d8x5t\" (UID: \"b18917c3-98cb-407b-97e5-5333aa5ca797\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d8x5t" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.997972 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b18917c3-98cb-407b-97e5-5333aa5ca797-apiservice-cert\") pod \"packageserver-d55dfcdfc-d8x5t\" (UID: \"b18917c3-98cb-407b-97e5-5333aa5ca797\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d8x5t" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.998046 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9pxr7\" (UniqueName: \"kubernetes.io/projected/590f742c-e0d3-4394-8ef3-08bde5d57d01-kube-api-access-9pxr7\") pod \"marketplace-operator-79b997595-dxf4x\" (UID: \"590f742c-e0d3-4394-8ef3-08bde5d57d01\") " pod="openshift-marketplace/marketplace-operator-79b997595-dxf4x" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.998140 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/499ccac5-6595-4d67-93ed-dd2d19d045bf-signing-cabundle\") pod \"service-ca-9c57cc56f-7792m\" (UID: \"499ccac5-6595-4d67-93ed-dd2d19d045bf\") " pod="openshift-service-ca/service-ca-9c57cc56f-7792m" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.998227 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/499ccac5-6595-4d67-93ed-dd2d19d045bf-signing-key\") pod \"service-ca-9c57cc56f-7792m\" (UID: \"499ccac5-6595-4d67-93ed-dd2d19d045bf\") " pod="openshift-service-ca/service-ca-9c57cc56f-7792m" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.998251 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5ed0f2b3-fd6d-4892-8198-2b81d5ea1a7f-metrics-tls\") pod \"ingress-operator-5b745b69d9-ng2nv\" (UID: \"5ed0f2b3-fd6d-4892-8198-2b81d5ea1a7f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ng2nv" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.998313 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/d7e79465-8df8-41a7-ab16-5c730f9ceca2-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-f2tjh\" (UID: \"d7e79465-8df8-41a7-ab16-5c730f9ceca2\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-f2tjh" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.998339 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gfvzz\" (UniqueName: \"kubernetes.io/projected/4a74ce9e-b145-4e40-ad4e-67ecdf722195-kube-api-access-gfvzz\") pod \"control-plane-machine-set-operator-78cbb6b69f-2stv6\" (UID: \"4a74ce9e-b145-4e40-ad4e-67ecdf722195\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2stv6" Nov 25 10:34:28 crc kubenswrapper[4821]: I1125 10:34:28.998539 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/b18917c3-98cb-407b-97e5-5333aa5ca797-tmpfs\") pod \"packageserver-d55dfcdfc-d8x5t\" (UID: \"b18917c3-98cb-407b-97e5-5333aa5ca797\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d8x5t" Nov 25 10:34:29 crc kubenswrapper[4821]: I1125 10:34:29.007846 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0e82d98b-7c99-496d-910d-e9bfae2ca6ec-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-6xxs8\" (UID: \"0e82d98b-7c99-496d-910d-e9bfae2ca6ec\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6xxs8" Nov 25 10:34:29 crc kubenswrapper[4821]: I1125 10:34:29.017672 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Nov 25 10:34:29 crc kubenswrapper[4821]: I1125 10:34:29.027993 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e82d98b-7c99-496d-910d-e9bfae2ca6ec-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-6xxs8\" (UID: \"0e82d98b-7c99-496d-910d-e9bfae2ca6ec\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6xxs8" Nov 25 10:34:29 crc kubenswrapper[4821]: I1125 10:34:29.038515 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Nov 25 10:34:29 crc kubenswrapper[4821]: I1125 10:34:29.057624 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Nov 25 10:34:29 crc kubenswrapper[4821]: I1125 10:34:29.078708 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Nov 25 10:34:29 crc kubenswrapper[4821]: I1125 10:34:29.097687 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Nov 25 10:34:29 crc kubenswrapper[4821]: I1125 10:34:29.101852 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5ed0f2b3-fd6d-4892-8198-2b81d5ea1a7f-metrics-tls\") pod \"ingress-operator-5b745b69d9-ng2nv\" (UID: \"5ed0f2b3-fd6d-4892-8198-2b81d5ea1a7f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ng2nv" Nov 25 10:34:29 crc kubenswrapper[4821]: I1125 10:34:29.124449 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Nov 25 10:34:29 crc kubenswrapper[4821]: I1125 10:34:29.128359 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5ed0f2b3-fd6d-4892-8198-2b81d5ea1a7f-trusted-ca\") pod \"ingress-operator-5b745b69d9-ng2nv\" (UID: \"5ed0f2b3-fd6d-4892-8198-2b81d5ea1a7f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ng2nv" Nov 25 10:34:29 crc kubenswrapper[4821]: I1125 10:34:29.138830 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Nov 25 10:34:29 crc kubenswrapper[4821]: I1125 10:34:29.165033 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Nov 25 10:34:29 crc kubenswrapper[4821]: I1125 10:34:29.171136 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2d3eed24-29f1-414a-a5d8-f8b021e6aff4-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-tkgh9\" (UID: \"2d3eed24-29f1-414a-a5d8-f8b021e6aff4\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tkgh9" Nov 25 10:34:29 crc kubenswrapper[4821]: I1125 10:34:29.177645 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Nov 25 10:34:29 crc kubenswrapper[4821]: I1125 10:34:29.198592 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Nov 25 10:34:29 crc kubenswrapper[4821]: I1125 10:34:29.212938 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/2d3eed24-29f1-414a-a5d8-f8b021e6aff4-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-tkgh9\" (UID: \"2d3eed24-29f1-414a-a5d8-f8b021e6aff4\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tkgh9" Nov 25 10:34:29 crc kubenswrapper[4821]: I1125 10:34:29.217971 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Nov 25 10:34:29 crc kubenswrapper[4821]: I1125 10:34:29.237593 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Nov 25 10:34:29 crc kubenswrapper[4821]: I1125 10:34:29.258018 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Nov 25 10:34:29 crc kubenswrapper[4821]: I1125 10:34:29.277678 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Nov 25 10:34:29 crc kubenswrapper[4821]: I1125 10:34:29.298491 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Nov 25 10:34:29 crc kubenswrapper[4821]: I1125 10:34:29.318037 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Nov 25 10:34:29 crc kubenswrapper[4821]: I1125 10:34:29.337848 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Nov 25 10:34:29 crc kubenswrapper[4821]: I1125 10:34:29.358339 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Nov 25 10:34:29 crc kubenswrapper[4821]: I1125 10:34:29.378513 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Nov 25 10:34:29 crc kubenswrapper[4821]: I1125 10:34:29.398920 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Nov 25 10:34:29 crc kubenswrapper[4821]: I1125 10:34:29.418347 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Nov 25 10:34:29 crc kubenswrapper[4821]: I1125 10:34:29.437874 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Nov 25 10:34:29 crc kubenswrapper[4821]: I1125 10:34:29.458818 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Nov 25 10:34:29 crc kubenswrapper[4821]: I1125 10:34:29.477952 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Nov 25 10:34:29 crc kubenswrapper[4821]: I1125 10:34:29.497622 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Nov 25 10:34:29 crc kubenswrapper[4821]: I1125 10:34:29.518140 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Nov 25 10:34:29 crc kubenswrapper[4821]: I1125 10:34:29.538426 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Nov 25 10:34:29 crc kubenswrapper[4821]: I1125 10:34:29.557680 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Nov 25 10:34:29 crc kubenswrapper[4821]: I1125 10:34:29.578143 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Nov 25 10:34:29 crc kubenswrapper[4821]: I1125 10:34:29.591049 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/590f742c-e0d3-4394-8ef3-08bde5d57d01-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-dxf4x\" (UID: \"590f742c-e0d3-4394-8ef3-08bde5d57d01\") " pod="openshift-marketplace/marketplace-operator-79b997595-dxf4x" Nov 25 10:34:29 crc kubenswrapper[4821]: I1125 10:34:29.598319 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Nov 25 10:34:29 crc kubenswrapper[4821]: I1125 10:34:29.617837 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Nov 25 10:34:29 crc kubenswrapper[4821]: I1125 10:34:29.646891 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Nov 25 10:34:29 crc kubenswrapper[4821]: I1125 10:34:29.649757 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/590f742c-e0d3-4394-8ef3-08bde5d57d01-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-dxf4x\" (UID: \"590f742c-e0d3-4394-8ef3-08bde5d57d01\") " pod="openshift-marketplace/marketplace-operator-79b997595-dxf4x" Nov 25 10:34:29 crc kubenswrapper[4821]: I1125 10:34:29.657730 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Nov 25 10:34:29 crc kubenswrapper[4821]: I1125 10:34:29.678080 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Nov 25 10:34:29 crc kubenswrapper[4821]: I1125 10:34:29.698199 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Nov 25 10:34:29 crc kubenswrapper[4821]: I1125 10:34:29.717136 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Nov 25 10:34:29 crc kubenswrapper[4821]: I1125 10:34:29.738663 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Nov 25 10:34:29 crc kubenswrapper[4821]: I1125 10:34:29.758139 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Nov 25 10:34:29 crc kubenswrapper[4821]: I1125 10:34:29.777053 4821 request.go:700] Waited for 1.000747113s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-service-ca-operator/configmaps?fieldSelector=metadata.name%3Dservice-ca-operator-config&limit=500&resourceVersion=0 Nov 25 10:34:29 crc kubenswrapper[4821]: I1125 10:34:29.778358 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Nov 25 10:34:29 crc kubenswrapper[4821]: I1125 10:34:29.797863 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Nov 25 10:34:29 crc kubenswrapper[4821]: I1125 10:34:29.818380 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Nov 25 10:34:29 crc kubenswrapper[4821]: I1125 10:34:29.838491 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Nov 25 10:34:29 crc kubenswrapper[4821]: I1125 10:34:29.850794 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b18917c3-98cb-407b-97e5-5333aa5ca797-apiservice-cert\") pod \"packageserver-d55dfcdfc-d8x5t\" (UID: \"b18917c3-98cb-407b-97e5-5333aa5ca797\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d8x5t" Nov 25 10:34:29 crc kubenswrapper[4821]: I1125 10:34:29.850829 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b18917c3-98cb-407b-97e5-5333aa5ca797-webhook-cert\") pod \"packageserver-d55dfcdfc-d8x5t\" (UID: \"b18917c3-98cb-407b-97e5-5333aa5ca797\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d8x5t" Nov 25 10:34:29 crc kubenswrapper[4821]: I1125 10:34:29.858255 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Nov 25 10:34:29 crc kubenswrapper[4821]: I1125 10:34:29.861902 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/499ccac5-6595-4d67-93ed-dd2d19d045bf-signing-key\") pod \"service-ca-9c57cc56f-7792m\" (UID: \"499ccac5-6595-4d67-93ed-dd2d19d045bf\") " pod="openshift-service-ca/service-ca-9c57cc56f-7792m" Nov 25 10:34:29 crc kubenswrapper[4821]: I1125 10:34:29.878389 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Nov 25 10:34:29 crc kubenswrapper[4821]: I1125 10:34:29.897549 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Nov 25 10:34:29 crc kubenswrapper[4821]: I1125 10:34:29.899237 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/499ccac5-6595-4d67-93ed-dd2d19d045bf-signing-cabundle\") pod \"service-ca-9c57cc56f-7792m\" (UID: \"499ccac5-6595-4d67-93ed-dd2d19d045bf\") " pod="openshift-service-ca/service-ca-9c57cc56f-7792m" Nov 25 10:34:29 crc kubenswrapper[4821]: I1125 10:34:29.918067 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Nov 25 10:34:29 crc kubenswrapper[4821]: I1125 10:34:29.938747 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Nov 25 10:34:29 crc kubenswrapper[4821]: I1125 10:34:29.978022 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5x4xr\" (UniqueName: \"kubernetes.io/projected/88725123-3da9-4127-9506-80c90c9accd5-kube-api-access-5x4xr\") pod \"console-f9d7485db-hf49d\" (UID: \"88725123-3da9-4127-9506-80c90c9accd5\") " pod="openshift-console/console-f9d7485db-hf49d" Nov 25 10:34:29 crc kubenswrapper[4821]: I1125 10:34:29.992941 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njkct\" (UniqueName: \"kubernetes.io/projected/cb0f66e4-2c0f-4a95-89db-1aef53ae5b3b-kube-api-access-njkct\") pod \"openshift-apiserver-operator-796bbdcf4f-5m5gj\" (UID: \"cb0f66e4-2c0f-4a95-89db-1aef53ae5b3b\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5m5gj" Nov 25 10:34:29 crc kubenswrapper[4821]: E1125 10:34:29.996642 4821 secret.go:188] Couldn't get secret openshift-machine-api/control-plane-machine-set-operator-tls: failed to sync secret cache: timed out waiting for the condition Nov 25 10:34:29 crc kubenswrapper[4821]: E1125 10:34:29.996702 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4a74ce9e-b145-4e40-ad4e-67ecdf722195-control-plane-machine-set-operator-tls podName:4a74ce9e-b145-4e40-ad4e-67ecdf722195 nodeName:}" failed. No retries permitted until 2025-11-25 10:34:30.496686373 +0000 UTC m=+141.033006220 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "control-plane-machine-set-operator-tls" (UniqueName: "kubernetes.io/secret/4a74ce9e-b145-4e40-ad4e-67ecdf722195-control-plane-machine-set-operator-tls") pod "control-plane-machine-set-operator-78cbb6b69f-2stv6" (UID: "4a74ce9e-b145-4e40-ad4e-67ecdf722195") : failed to sync secret cache: timed out waiting for the condition Nov 25 10:34:29 crc kubenswrapper[4821]: E1125 10:34:29.998809 4821 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/package-server-manager-serving-cert: failed to sync secret cache: timed out waiting for the condition Nov 25 10:34:29 crc kubenswrapper[4821]: E1125 10:34:29.998894 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d7e79465-8df8-41a7-ab16-5c730f9ceca2-package-server-manager-serving-cert podName:d7e79465-8df8-41a7-ab16-5c730f9ceca2 nodeName:}" failed. No retries permitted until 2025-11-25 10:34:30.498877406 +0000 UTC m=+141.035197253 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "package-server-manager-serving-cert" (UniqueName: "kubernetes.io/secret/d7e79465-8df8-41a7-ab16-5c730f9ceca2-package-server-manager-serving-cert") pod "package-server-manager-789f6589d5-f2tjh" (UID: "d7e79465-8df8-41a7-ab16-5c730f9ceca2") : failed to sync secret cache: timed out waiting for the condition Nov 25 10:34:30 crc kubenswrapper[4821]: I1125 10:34:30.016588 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ktx4k\" (UniqueName: \"kubernetes.io/projected/5a7c1a05-b29c-4ff3-a4d6-90fe59eb1a01-kube-api-access-ktx4k\") pod \"openshift-config-operator-7777fb866f-q8tq8\" (UID: \"5a7c1a05-b29c-4ff3-a4d6-90fe59eb1a01\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-q8tq8" Nov 25 10:34:30 crc kubenswrapper[4821]: I1125 10:34:30.034451 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gzl9t\" (UniqueName: \"kubernetes.io/projected/ba27af47-1944-4970-a015-e96c7cee4862-kube-api-access-gzl9t\") pod \"route-controller-manager-6576b87f9c-dv9nt\" (UID: \"ba27af47-1944-4970-a015-e96c7cee4862\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dv9nt" Nov 25 10:34:30 crc kubenswrapper[4821]: I1125 10:34:30.053611 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wpnl\" (UniqueName: \"kubernetes.io/projected/59ec70c2-c208-4529-a763-a0d7ea7e1dfc-kube-api-access-9wpnl\") pod \"machine-approver-56656f9798-86cnx\" (UID: \"59ec70c2-c208-4529-a763-a0d7ea7e1dfc\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-86cnx" Nov 25 10:34:30 crc kubenswrapper[4821]: I1125 10:34:30.073385 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8g76m\" (UniqueName: \"kubernetes.io/projected/b75a7773-a689-46c1-b98e-e7c1be0878e6-kube-api-access-8g76m\") pod \"console-operator-58897d9998-t5hm7\" (UID: \"b75a7773-a689-46c1-b98e-e7c1be0878e6\") " pod="openshift-console-operator/console-operator-58897d9998-t5hm7" Nov 25 10:34:30 crc kubenswrapper[4821]: I1125 10:34:30.085426 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-86cnx" Nov 25 10:34:30 crc kubenswrapper[4821]: I1125 10:34:30.096925 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7b5z\" (UniqueName: \"kubernetes.io/projected/88feae6a-5031-48ed-93ec-34ef8d2eff40-kube-api-access-q7b5z\") pod \"oauth-openshift-558db77b4-mbn4k\" (UID: \"88feae6a-5031-48ed-93ec-34ef8d2eff40\") " pod="openshift-authentication/oauth-openshift-558db77b4-mbn4k" Nov 25 10:34:30 crc kubenswrapper[4821]: W1125 10:34:30.098410 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod59ec70c2_c208_4529_a763_a0d7ea7e1dfc.slice/crio-142b35c907fc8f063d743e92486505bec38497c0f59a67a78ba5ad4737ebc3c8 WatchSource:0}: Error finding container 142b35c907fc8f063d743e92486505bec38497c0f59a67a78ba5ad4737ebc3c8: Status 404 returned error can't find the container with id 142b35c907fc8f063d743e92486505bec38497c0f59a67a78ba5ad4737ebc3c8 Nov 25 10:34:30 crc kubenswrapper[4821]: I1125 10:34:30.104615 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-t5hm7" Nov 25 10:34:30 crc kubenswrapper[4821]: I1125 10:34:30.112549 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-q8tq8" Nov 25 10:34:30 crc kubenswrapper[4821]: I1125 10:34:30.113046 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-trdqk\" (UniqueName: \"kubernetes.io/projected/518db853-3bbe-421b-a2dd-6d1b2e965120-kube-api-access-trdqk\") pod \"controller-manager-879f6c89f-58htt\" (UID: \"518db853-3bbe-421b-a2dd-6d1b2e965120\") " pod="openshift-controller-manager/controller-manager-879f6c89f-58htt" Nov 25 10:34:30 crc kubenswrapper[4821]: I1125 10:34:30.117676 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Nov 25 10:34:30 crc kubenswrapper[4821]: I1125 10:34:30.125364 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-hf49d" Nov 25 10:34:30 crc kubenswrapper[4821]: I1125 10:34:30.133653 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-58htt" Nov 25 10:34:30 crc kubenswrapper[4821]: I1125 10:34:30.137847 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 25 10:34:30 crc kubenswrapper[4821]: I1125 10:34:30.147088 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5m5gj" Nov 25 10:34:30 crc kubenswrapper[4821]: I1125 10:34:30.157779 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 25 10:34:30 crc kubenswrapper[4821]: I1125 10:34:30.174380 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dv9nt" Nov 25 10:34:30 crc kubenswrapper[4821]: I1125 10:34:30.178468 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Nov 25 10:34:30 crc kubenswrapper[4821]: I1125 10:34:30.198300 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Nov 25 10:34:30 crc kubenswrapper[4821]: I1125 10:34:30.225254 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Nov 25 10:34:30 crc kubenswrapper[4821]: I1125 10:34:30.238719 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Nov 25 10:34:30 crc kubenswrapper[4821]: I1125 10:34:30.259284 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Nov 25 10:34:30 crc kubenswrapper[4821]: I1125 10:34:30.278721 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Nov 25 10:34:30 crc kubenswrapper[4821]: I1125 10:34:30.297918 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Nov 25 10:34:30 crc kubenswrapper[4821]: I1125 10:34:30.322249 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Nov 25 10:34:30 crc kubenswrapper[4821]: I1125 10:34:30.338050 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Nov 25 10:34:30 crc kubenswrapper[4821]: I1125 10:34:30.361834 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Nov 25 10:34:30 crc kubenswrapper[4821]: I1125 10:34:30.378514 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Nov 25 10:34:30 crc kubenswrapper[4821]: I1125 10:34:30.392918 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-mbn4k" Nov 25 10:34:30 crc kubenswrapper[4821]: I1125 10:34:30.398686 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Nov 25 10:34:30 crc kubenswrapper[4821]: I1125 10:34:30.418084 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Nov 25 10:34:30 crc kubenswrapper[4821]: I1125 10:34:30.438279 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Nov 25 10:34:30 crc kubenswrapper[4821]: I1125 10:34:30.457986 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Nov 25 10:34:30 crc kubenswrapper[4821]: I1125 10:34:30.478308 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Nov 25 10:34:30 crc kubenswrapper[4821]: I1125 10:34:30.517377 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/d7e79465-8df8-41a7-ab16-5c730f9ceca2-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-f2tjh\" (UID: \"d7e79465-8df8-41a7-ab16-5c730f9ceca2\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-f2tjh" Nov 25 10:34:30 crc kubenswrapper[4821]: I1125 10:34:30.517455 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/4a74ce9e-b145-4e40-ad4e-67ecdf722195-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-2stv6\" (UID: \"4a74ce9e-b145-4e40-ad4e-67ecdf722195\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2stv6" Nov 25 10:34:30 crc kubenswrapper[4821]: I1125 10:34:30.521038 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/4a74ce9e-b145-4e40-ad4e-67ecdf722195-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-2stv6\" (UID: \"4a74ce9e-b145-4e40-ad4e-67ecdf722195\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2stv6" Nov 25 10:34:30 crc kubenswrapper[4821]: I1125 10:34:30.523433 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Nov 25 10:34:30 crc kubenswrapper[4821]: I1125 10:34:30.525372 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/d7e79465-8df8-41a7-ab16-5c730f9ceca2-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-f2tjh\" (UID: \"d7e79465-8df8-41a7-ab16-5c730f9ceca2\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-f2tjh" Nov 25 10:34:30 crc kubenswrapper[4821]: I1125 10:34:30.538215 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Nov 25 10:34:30 crc kubenswrapper[4821]: I1125 10:34:30.559065 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Nov 25 10:34:30 crc kubenswrapper[4821]: I1125 10:34:30.578667 4821 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Nov 25 10:34:30 crc kubenswrapper[4821]: I1125 10:34:30.599413 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Nov 25 10:34:30 crc kubenswrapper[4821]: I1125 10:34:30.618551 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Nov 25 10:34:30 crc kubenswrapper[4821]: I1125 10:34:30.640429 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Nov 25 10:34:30 crc kubenswrapper[4821]: I1125 10:34:30.660459 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Nov 25 10:34:30 crc kubenswrapper[4821]: I1125 10:34:30.667830 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-86cnx" event={"ID":"59ec70c2-c208-4529-a763-a0d7ea7e1dfc","Type":"ContainerStarted","Data":"142b35c907fc8f063d743e92486505bec38497c0f59a67a78ba5ad4737ebc3c8"} Nov 25 10:34:30 crc kubenswrapper[4821]: I1125 10:34:30.679242 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Nov 25 10:34:30 crc kubenswrapper[4821]: I1125 10:34:30.718677 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Nov 25 10:34:30 crc kubenswrapper[4821]: I1125 10:34:30.734365 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-q8tq8"] Nov 25 10:34:30 crc kubenswrapper[4821]: I1125 10:34:30.740359 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Nov 25 10:34:30 crc kubenswrapper[4821]: I1125 10:34:30.759448 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Nov 25 10:34:30 crc kubenswrapper[4821]: I1125 10:34:30.777920 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Nov 25 10:34:30 crc kubenswrapper[4821]: I1125 10:34:30.795995 4821 request.go:700] Waited for 1.901999541s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-api/serviceaccounts/machine-api-operator/token Nov 25 10:34:30 crc kubenswrapper[4821]: I1125 10:34:30.830591 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tnlzh\" (UniqueName: \"kubernetes.io/projected/24708c13-efe2-435b-acde-b29b48871f16-kube-api-access-tnlzh\") pod \"machine-api-operator-5694c8668f-zj4s2\" (UID: \"24708c13-efe2-435b-acde-b29b48871f16\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zj4s2" Nov 25 10:34:30 crc kubenswrapper[4821]: I1125 10:34:30.832562 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dt66w\" (UniqueName: \"kubernetes.io/projected/7fe78aa4-571c-488b-96bd-23ae4b6333db-kube-api-access-dt66w\") pod \"openshift-controller-manager-operator-756b6f6bc6-zvztq\" (UID: \"7fe78aa4-571c-488b-96bd-23ae4b6333db\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zvztq" Nov 25 10:34:30 crc kubenswrapper[4821]: I1125 10:34:30.854182 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rf8tm\" (UniqueName: \"kubernetes.io/projected/155595fc-76b3-4561-bd3f-c540076d4cd6-kube-api-access-rf8tm\") pod \"olm-operator-6b444d44fb-v88lg\" (UID: \"155595fc-76b3-4561-bd3f-c540076d4cd6\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-v88lg" Nov 25 10:34:30 crc kubenswrapper[4821]: I1125 10:34:30.869542 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-zj4s2" Nov 25 10:34:30 crc kubenswrapper[4821]: I1125 10:34:30.873002 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tpsds\" (UniqueName: \"kubernetes.io/projected/513166e2-0ac9-4299-bf84-d994a46da192-kube-api-access-tpsds\") pod \"cluster-samples-operator-665b6dd947-k622z\" (UID: \"513166e2-0ac9-4299-bf84-d994a46da192\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k622z" Nov 25 10:34:30 crc kubenswrapper[4821]: I1125 10:34:30.887337 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k622z" Nov 25 10:34:30 crc kubenswrapper[4821]: I1125 10:34:30.891207 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6cq5w\" (UniqueName: \"kubernetes.io/projected/2c2ecf12-49bd-4baf-90df-68486f27be1b-kube-api-access-6cq5w\") pod \"catalog-operator-68c6474976-4kjbr\" (UID: \"2c2ecf12-49bd-4baf-90df-68486f27be1b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4kjbr" Nov 25 10:34:30 crc kubenswrapper[4821]: I1125 10:34:30.912566 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4qsq\" (UniqueName: \"kubernetes.io/projected/402cf28b-25d2-428a-ab7b-1d70345a921e-kube-api-access-r4qsq\") pod \"multus-admission-controller-857f4d67dd-p2jmh\" (UID: \"402cf28b-25d2-428a-ab7b-1d70345a921e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-p2jmh" Nov 25 10:34:30 crc kubenswrapper[4821]: I1125 10:34:30.938851 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fq9pc\" (UniqueName: \"kubernetes.io/projected/b0e34420-095d-42aa-89c4-5a80331e1c01-kube-api-access-fq9pc\") pod \"dns-operator-744455d44c-hddc7\" (UID: \"b0e34420-095d-42aa-89c4-5a80331e1c01\") " pod="openshift-dns-operator/dns-operator-744455d44c-hddc7" Nov 25 10:34:30 crc kubenswrapper[4821]: I1125 10:34:30.951475 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v86wz\" (UniqueName: \"kubernetes.io/projected/064de93d-7bb5-47b5-9585-bf32be9a05f6-kube-api-access-v86wz\") pod \"apiserver-76f77b778f-jb5sz\" (UID: \"064de93d-7bb5-47b5-9585-bf32be9a05f6\") " pod="openshift-apiserver/apiserver-76f77b778f-jb5sz" Nov 25 10:34:30 crc kubenswrapper[4821]: I1125 10:34:30.952296 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zvztq" Nov 25 10:34:30 crc kubenswrapper[4821]: I1125 10:34:30.953689 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-mbn4k"] Nov 25 10:34:30 crc kubenswrapper[4821]: I1125 10:34:30.954807 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-dv9nt"] Nov 25 10:34:30 crc kubenswrapper[4821]: I1125 10:34:30.962186 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4kjbr" Nov 25 10:34:30 crc kubenswrapper[4821]: I1125 10:34:30.969891 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-hddc7" Nov 25 10:34:30 crc kubenswrapper[4821]: I1125 10:34:30.971153 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rc8sv\" (UniqueName: \"kubernetes.io/projected/0e82d98b-7c99-496d-910d-e9bfae2ca6ec-kube-api-access-rc8sv\") pod \"kube-storage-version-migrator-operator-b67b599dd-6xxs8\" (UID: \"0e82d98b-7c99-496d-910d-e9bfae2ca6ec\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6xxs8" Nov 25 10:34:30 crc kubenswrapper[4821]: I1125 10:34:30.978919 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-p2jmh" Nov 25 10:34:30 crc kubenswrapper[4821]: I1125 10:34:30.986299 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-v88lg" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:30.998906 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bdd88e7d-edf6-41c1-a5a3-47746bb3ab99-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-whcwc\" (UID: \"bdd88e7d-edf6-41c1-a5a3-47746bb3ab99\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-whcwc" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.027114 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4878872d-8035-442a-a355-7c966f37e8c1-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-n4lbr\" (UID: \"4878872d-8035-442a-a355-7c966f37e8c1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-n4lbr" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.032555 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5m5gj"] Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.032791 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-n4lbr" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.034943 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-t5hm7"] Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.035234 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kbrfs\" (UniqueName: \"kubernetes.io/projected/9f6c1e2f-b7e7-40e7-92c7-8ca1b57eb9f1-kube-api-access-kbrfs\") pod \"authentication-operator-69f744f599-tpbzb\" (UID: \"9f6c1e2f-b7e7-40e7-92c7-8ca1b57eb9f1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tpbzb" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.037986 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-58htt"] Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.038201 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6xxs8" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.040930 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-hf49d"] Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.054104 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-44zw6\" (UniqueName: \"kubernetes.io/projected/f954533e-e8a0-4899-be72-241760feae95-kube-api-access-44zw6\") pod \"downloads-7954f5f757-hjvt9\" (UID: \"f954533e-e8a0-4899-be72-241760feae95\") " pod="openshift-console/downloads-7954f5f757-hjvt9" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.083068 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8234c8fc-a19c-45c6-8d25-e65668a244ea-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-wxp2r\" (UID: \"8234c8fc-a19c-45c6-8d25-e65668a244ea\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wxp2r" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.097487 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r2cg5\" (UniqueName: \"kubernetes.io/projected/2d3eed24-29f1-414a-a5d8-f8b021e6aff4-kube-api-access-r2cg5\") pod \"cluster-image-registry-operator-dc59b4c8b-tkgh9\" (UID: \"2d3eed24-29f1-414a-a5d8-f8b021e6aff4\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tkgh9" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.123047 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-hjvt9" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.125410 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xgpbf\" (UniqueName: \"kubernetes.io/projected/d7e79465-8df8-41a7-ab16-5c730f9ceca2-kube-api-access-xgpbf\") pod \"package-server-manager-789f6589d5-f2tjh\" (UID: \"d7e79465-8df8-41a7-ab16-5c730f9ceca2\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-f2tjh" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.126035 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-f2tjh" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.133594 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-zj4s2"] Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.141923 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2nzbr\" (UniqueName: \"kubernetes.io/projected/499ccac5-6595-4d67-93ed-dd2d19d045bf-kube-api-access-2nzbr\") pod \"service-ca-9c57cc56f-7792m\" (UID: \"499ccac5-6595-4d67-93ed-dd2d19d045bf\") " pod="openshift-service-ca/service-ca-9c57cc56f-7792m" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.145881 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-tpbzb" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.147097 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k622z"] Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.154916 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bwgwh\" (UniqueName: \"kubernetes.io/projected/5ed0f2b3-fd6d-4892-8198-2b81d5ea1a7f-kube-api-access-bwgwh\") pod \"ingress-operator-5b745b69d9-ng2nv\" (UID: \"5ed0f2b3-fd6d-4892-8198-2b81d5ea1a7f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ng2nv" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.182211 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h5s8b\" (UniqueName: \"kubernetes.io/projected/b18917c3-98cb-407b-97e5-5333aa5ca797-kube-api-access-h5s8b\") pod \"packageserver-d55dfcdfc-d8x5t\" (UID: \"b18917c3-98cb-407b-97e5-5333aa5ca797\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d8x5t" Nov 25 10:34:31 crc kubenswrapper[4821]: W1125 10:34:31.193849 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcb0f66e4_2c0f_4a95_89db_1aef53ae5b3b.slice/crio-3b4d1a301ae80bb352cffc0a098d0a292d7c7b887a5eb9b155bf3518c2313cf0 WatchSource:0}: Error finding container 3b4d1a301ae80bb352cffc0a098d0a292d7c7b887a5eb9b155bf3518c2313cf0: Status 404 returned error can't find the container with id 3b4d1a301ae80bb352cffc0a098d0a292d7c7b887a5eb9b155bf3518c2313cf0 Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.196527 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5ed0f2b3-fd6d-4892-8198-2b81d5ea1a7f-bound-sa-token\") pod \"ingress-operator-5b745b69d9-ng2nv\" (UID: \"5ed0f2b3-fd6d-4892-8198-2b81d5ea1a7f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ng2nv" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.212915 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-jb5sz" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.222182 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2d3eed24-29f1-414a-a5d8-f8b021e6aff4-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-tkgh9\" (UID: \"2d3eed24-29f1-414a-a5d8-f8b021e6aff4\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tkgh9" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.242572 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9pxr7\" (UniqueName: \"kubernetes.io/projected/590f742c-e0d3-4394-8ef3-08bde5d57d01-kube-api-access-9pxr7\") pod \"marketplace-operator-79b997595-dxf4x\" (UID: \"590f742c-e0d3-4394-8ef3-08bde5d57d01\") " pod="openshift-marketplace/marketplace-operator-79b997595-dxf4x" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.254379 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gfvzz\" (UniqueName: \"kubernetes.io/projected/4a74ce9e-b145-4e40-ad4e-67ecdf722195-kube-api-access-gfvzz\") pod \"control-plane-machine-set-operator-78cbb6b69f-2stv6\" (UID: \"4a74ce9e-b145-4e40-ad4e-67ecdf722195\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2stv6" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.372825 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-whcwc" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.372889 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tkgh9" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.372916 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ng2nv" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.372825 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wxp2r" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.373981 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/3fb2355d-5ec2-44f1-8665-a3d73c24f8a6-ca-trust-extracted\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.374032 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/eba7a4e4-d755-4368-a487-43f200d258b4-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-kps6n\" (UID: \"eba7a4e4-d755-4368-a487-43f200d258b4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kps6n" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.374074 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3fb2355d-5ec2-44f1-8665-a3d73c24f8a6-bound-sa-token\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.374098 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/db41ea16-6481-4ed9-a935-d14544e552a0-etcd-ca\") pod \"etcd-operator-b45778765-v6pjk\" (UID: \"db41ea16-6481-4ed9-a935-d14544e552a0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-v6pjk" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.374429 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/eba7a4e4-d755-4368-a487-43f200d258b4-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-kps6n\" (UID: \"eba7a4e4-d755-4368-a487-43f200d258b4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kps6n" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.374504 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qk4p5\" (UniqueName: \"kubernetes.io/projected/486e2f09-ddfc-4058-8749-bf747fc829ca-kube-api-access-qk4p5\") pod \"machine-config-controller-84d6567774-w5k28\" (UID: \"486e2f09-ddfc-4058-8749-bf747fc829ca\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w5k28" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.374539 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/3fb2355d-5ec2-44f1-8665-a3d73c24f8a6-registry-tls\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.374583 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/486e2f09-ddfc-4058-8749-bf747fc829ca-proxy-tls\") pod \"machine-config-controller-84d6567774-w5k28\" (UID: \"486e2f09-ddfc-4058-8749-bf747fc829ca\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w5k28" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.374630 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/eba7a4e4-d755-4368-a487-43f200d258b4-encryption-config\") pod \"apiserver-7bbb656c7d-kps6n\" (UID: \"eba7a4e4-d755-4368-a487-43f200d258b4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kps6n" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.374685 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/db41ea16-6481-4ed9-a935-d14544e552a0-serving-cert\") pod \"etcd-operator-b45778765-v6pjk\" (UID: \"db41ea16-6481-4ed9-a935-d14544e552a0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-v6pjk" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.374714 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3fb2355d-5ec2-44f1-8665-a3d73c24f8a6-trusted-ca\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.374769 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/db41ea16-6481-4ed9-a935-d14544e552a0-etcd-service-ca\") pod \"etcd-operator-b45778765-v6pjk\" (UID: \"db41ea16-6481-4ed9-a935-d14544e552a0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-v6pjk" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.374810 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7bl5\" (UniqueName: \"kubernetes.io/projected/db41ea16-6481-4ed9-a935-d14544e552a0-kube-api-access-c7bl5\") pod \"etcd-operator-b45778765-v6pjk\" (UID: \"db41ea16-6481-4ed9-a935-d14544e552a0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-v6pjk" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.374879 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/486e2f09-ddfc-4058-8749-bf747fc829ca-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-w5k28\" (UID: \"486e2f09-ddfc-4058-8749-bf747fc829ca\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w5k28" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.374965 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/db41ea16-6481-4ed9-a935-d14544e552a0-etcd-client\") pod \"etcd-operator-b45778765-v6pjk\" (UID: \"db41ea16-6481-4ed9-a935-d14544e552a0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-v6pjk" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.375017 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.375066 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/eba7a4e4-d755-4368-a487-43f200d258b4-etcd-client\") pod \"apiserver-7bbb656c7d-kps6n\" (UID: \"eba7a4e4-d755-4368-a487-43f200d258b4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kps6n" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.375092 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db778f33-712d-42fc-849e-20e0af125a5a-config\") pod \"service-ca-operator-777779d784-prvdm\" (UID: \"db778f33-712d-42fc-849e-20e0af125a5a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-prvdm" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.375134 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgwn7\" (UniqueName: \"kubernetes.io/projected/3fb2355d-5ec2-44f1-8665-a3d73c24f8a6-kube-api-access-lgwn7\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.375149 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cq9vq\" (UniqueName: \"kubernetes.io/projected/db778f33-712d-42fc-849e-20e0af125a5a-kube-api-access-cq9vq\") pod \"service-ca-operator-777779d784-prvdm\" (UID: \"db778f33-712d-42fc-849e-20e0af125a5a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-prvdm" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.375201 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/3fb2355d-5ec2-44f1-8665-a3d73c24f8a6-installation-pull-secrets\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.375229 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eba7a4e4-d755-4368-a487-43f200d258b4-serving-cert\") pod \"apiserver-7bbb656c7d-kps6n\" (UID: \"eba7a4e4-d755-4368-a487-43f200d258b4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kps6n" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.375249 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/eba7a4e4-d755-4368-a487-43f200d258b4-audit-policies\") pod \"apiserver-7bbb656c7d-kps6n\" (UID: \"eba7a4e4-d755-4368-a487-43f200d258b4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kps6n" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.375278 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db41ea16-6481-4ed9-a935-d14544e552a0-config\") pod \"etcd-operator-b45778765-v6pjk\" (UID: \"db41ea16-6481-4ed9-a935-d14544e552a0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-v6pjk" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.375294 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/db778f33-712d-42fc-849e-20e0af125a5a-serving-cert\") pod \"service-ca-operator-777779d784-prvdm\" (UID: \"db778f33-712d-42fc-849e-20e0af125a5a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-prvdm" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.375310 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/eba7a4e4-d755-4368-a487-43f200d258b4-audit-dir\") pod \"apiserver-7bbb656c7d-kps6n\" (UID: \"eba7a4e4-d755-4368-a487-43f200d258b4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kps6n" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.375328 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8rvk\" (UniqueName: \"kubernetes.io/projected/eba7a4e4-d755-4368-a487-43f200d258b4-kube-api-access-m8rvk\") pod \"apiserver-7bbb656c7d-kps6n\" (UID: \"eba7a4e4-d755-4368-a487-43f200d258b4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kps6n" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.375345 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/3fb2355d-5ec2-44f1-8665-a3d73c24f8a6-registry-certificates\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:31 crc kubenswrapper[4821]: E1125 10:34:31.375574 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:31.875558952 +0000 UTC m=+142.411878869 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.387651 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-dxf4x" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.411510 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d8x5t" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.419100 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-7792m" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.450876 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-hddc7"] Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.461429 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2stv6" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.465347 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4kjbr"] Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.469267 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zvztq"] Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.475822 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:31 crc kubenswrapper[4821]: E1125 10:34:31.475894 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:31.975880189 +0000 UTC m=+142.512200036 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.477100 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/66798226-d262-4925-82db-2742d96a33fc-default-certificate\") pod \"router-default-5444994796-p92sp\" (UID: \"66798226-d262-4925-82db-2742d96a33fc\") " pod="openshift-ingress/router-default-5444994796-p92sp" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.477131 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/db41ea16-6481-4ed9-a935-d14544e552a0-etcd-client\") pod \"etcd-operator-b45778765-v6pjk\" (UID: \"db41ea16-6481-4ed9-a935-d14544e552a0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-v6pjk" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.477148 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/d1e46310-3643-4e6a-9624-467e4e00b093-images\") pod \"machine-config-operator-74547568cd-s7j8m\" (UID: \"d1e46310-3643-4e6a-9624-467e4e00b093\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s7j8m" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.477199 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d1e46310-3643-4e6a-9624-467e4e00b093-auth-proxy-config\") pod \"machine-config-operator-74547568cd-s7j8m\" (UID: \"d1e46310-3643-4e6a-9624-467e4e00b093\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s7j8m" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.477222 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.477281 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7v7zm\" (UniqueName: \"kubernetes.io/projected/9276b1a7-e1c8-4cdb-a77c-4b09df17be76-kube-api-access-7v7zm\") pod \"migrator-59844c95c7-898jb\" (UID: \"9276b1a7-e1c8-4cdb-a77c-4b09df17be76\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-898jb" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.477381 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/66798226-d262-4925-82db-2742d96a33fc-service-ca-bundle\") pod \"router-default-5444994796-p92sp\" (UID: \"66798226-d262-4925-82db-2742d96a33fc\") " pod="openshift-ingress/router-default-5444994796-p92sp" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.477441 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/eba7a4e4-d755-4368-a487-43f200d258b4-etcd-client\") pod \"apiserver-7bbb656c7d-kps6n\" (UID: \"eba7a4e4-d755-4368-a487-43f200d258b4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kps6n" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.477459 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/c82133bf-deb1-4b25-b002-f435a64847cd-certs\") pod \"machine-config-server-kc2wp\" (UID: \"c82133bf-deb1-4b25-b002-f435a64847cd\") " pod="openshift-machine-config-operator/machine-config-server-kc2wp" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.478002 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db778f33-712d-42fc-849e-20e0af125a5a-config\") pod \"service-ca-operator-777779d784-prvdm\" (UID: \"db778f33-712d-42fc-849e-20e0af125a5a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-prvdm" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.478050 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7lg4\" (UniqueName: \"kubernetes.io/projected/d1e46310-3643-4e6a-9624-467e4e00b093-kube-api-access-z7lg4\") pod \"machine-config-operator-74547568cd-s7j8m\" (UID: \"d1e46310-3643-4e6a-9624-467e4e00b093\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s7j8m" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.478069 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g52j5\" (UniqueName: \"kubernetes.io/projected/a5033129-ba8c-488f-982e-78e12b2f1664-kube-api-access-g52j5\") pod \"collect-profiles-29401110-dm86l\" (UID: \"a5033129-ba8c-488f-982e-78e12b2f1664\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401110-dm86l" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.478089 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a5033129-ba8c-488f-982e-78e12b2f1664-secret-volume\") pod \"collect-profiles-29401110-dm86l\" (UID: \"a5033129-ba8c-488f-982e-78e12b2f1664\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401110-dm86l" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.478119 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgwn7\" (UniqueName: \"kubernetes.io/projected/3fb2355d-5ec2-44f1-8665-a3d73c24f8a6-kube-api-access-lgwn7\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.478136 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cq9vq\" (UniqueName: \"kubernetes.io/projected/db778f33-712d-42fc-849e-20e0af125a5a-kube-api-access-cq9vq\") pod \"service-ca-operator-777779d784-prvdm\" (UID: \"db778f33-712d-42fc-849e-20e0af125a5a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-prvdm" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.478153 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/66798226-d262-4925-82db-2742d96a33fc-stats-auth\") pod \"router-default-5444994796-p92sp\" (UID: \"66798226-d262-4925-82db-2742d96a33fc\") " pod="openshift-ingress/router-default-5444994796-p92sp" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.478241 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/3fb2355d-5ec2-44f1-8665-a3d73c24f8a6-installation-pull-secrets\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.478270 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eba7a4e4-d755-4368-a487-43f200d258b4-serving-cert\") pod \"apiserver-7bbb656c7d-kps6n\" (UID: \"eba7a4e4-d755-4368-a487-43f200d258b4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kps6n" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.478298 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/c82133bf-deb1-4b25-b002-f435a64847cd-node-bootstrap-token\") pod \"machine-config-server-kc2wp\" (UID: \"c82133bf-deb1-4b25-b002-f435a64847cd\") " pod="openshift-machine-config-operator/machine-config-server-kc2wp" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.478321 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/eba7a4e4-d755-4368-a487-43f200d258b4-audit-policies\") pod \"apiserver-7bbb656c7d-kps6n\" (UID: \"eba7a4e4-d755-4368-a487-43f200d258b4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kps6n" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.478371 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/f97c276b-2b69-49a0-96d1-b1c5171a542e-registration-dir\") pod \"csi-hostpathplugin-tsb45\" (UID: \"f97c276b-2b69-49a0-96d1-b1c5171a542e\") " pod="hostpath-provisioner/csi-hostpathplugin-tsb45" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.478402 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db41ea16-6481-4ed9-a935-d14544e552a0-config\") pod \"etcd-operator-b45778765-v6pjk\" (UID: \"db41ea16-6481-4ed9-a935-d14544e552a0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-v6pjk" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.478419 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qh7h7\" (UniqueName: \"kubernetes.io/projected/f97c276b-2b69-49a0-96d1-b1c5171a542e-kube-api-access-qh7h7\") pod \"csi-hostpathplugin-tsb45\" (UID: \"f97c276b-2b69-49a0-96d1-b1c5171a542e\") " pod="hostpath-provisioner/csi-hostpathplugin-tsb45" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.478448 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/db778f33-712d-42fc-849e-20e0af125a5a-serving-cert\") pod \"service-ca-operator-777779d784-prvdm\" (UID: \"db778f33-712d-42fc-849e-20e0af125a5a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-prvdm" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.478467 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/eba7a4e4-d755-4368-a487-43f200d258b4-audit-dir\") pod \"apiserver-7bbb656c7d-kps6n\" (UID: \"eba7a4e4-d755-4368-a487-43f200d258b4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kps6n" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.478491 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/475ce242-f515-4f81-838c-5b0684222eb4-metrics-tls\") pod \"dns-default-rx6sh\" (UID: \"475ce242-f515-4f81-838c-5b0684222eb4\") " pod="openshift-dns/dns-default-rx6sh" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.478517 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8rvk\" (UniqueName: \"kubernetes.io/projected/eba7a4e4-d755-4368-a487-43f200d258b4-kube-api-access-m8rvk\") pod \"apiserver-7bbb656c7d-kps6n\" (UID: \"eba7a4e4-d755-4368-a487-43f200d258b4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kps6n" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.478533 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/f97c276b-2b69-49a0-96d1-b1c5171a542e-csi-data-dir\") pod \"csi-hostpathplugin-tsb45\" (UID: \"f97c276b-2b69-49a0-96d1-b1c5171a542e\") " pod="hostpath-provisioner/csi-hostpathplugin-tsb45" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.478771 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db778f33-712d-42fc-849e-20e0af125a5a-config\") pod \"service-ca-operator-777779d784-prvdm\" (UID: \"db778f33-712d-42fc-849e-20e0af125a5a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-prvdm" Nov 25 10:34:31 crc kubenswrapper[4821]: E1125 10:34:31.479110 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:31.979092179 +0000 UTC m=+142.515412126 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.479426 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/eba7a4e4-d755-4368-a487-43f200d258b4-audit-dir\") pod \"apiserver-7bbb656c7d-kps6n\" (UID: \"eba7a4e4-d755-4368-a487-43f200d258b4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kps6n" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.479692 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/3fb2355d-5ec2-44f1-8665-a3d73c24f8a6-registry-certificates\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.479831 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/eba7a4e4-d755-4368-a487-43f200d258b4-audit-policies\") pod \"apiserver-7bbb656c7d-kps6n\" (UID: \"eba7a4e4-d755-4368-a487-43f200d258b4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kps6n" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.480137 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db41ea16-6481-4ed9-a935-d14544e552a0-config\") pod \"etcd-operator-b45778765-v6pjk\" (UID: \"db41ea16-6481-4ed9-a935-d14544e552a0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-v6pjk" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.480293 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/3fb2355d-5ec2-44f1-8665-a3d73c24f8a6-ca-trust-extracted\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.480415 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/eba7a4e4-d755-4368-a487-43f200d258b4-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-kps6n\" (UID: \"eba7a4e4-d755-4368-a487-43f200d258b4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kps6n" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.480465 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3fb2355d-5ec2-44f1-8665-a3d73c24f8a6-bound-sa-token\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.480510 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/db41ea16-6481-4ed9-a935-d14544e552a0-etcd-ca\") pod \"etcd-operator-b45778765-v6pjk\" (UID: \"db41ea16-6481-4ed9-a935-d14544e552a0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-v6pjk" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.480534 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3e66624b-9d0d-4440-9cd6-06ca77cb7566-cert\") pod \"ingress-canary-cx6h6\" (UID: \"3e66624b-9d0d-4440-9cd6-06ca77cb7566\") " pod="openshift-ingress-canary/ingress-canary-cx6h6" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.480623 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/eba7a4e4-d755-4368-a487-43f200d258b4-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-kps6n\" (UID: \"eba7a4e4-d755-4368-a487-43f200d258b4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kps6n" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.480647 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4mnz\" (UniqueName: \"kubernetes.io/projected/475ce242-f515-4f81-838c-5b0684222eb4-kube-api-access-q4mnz\") pod \"dns-default-rx6sh\" (UID: \"475ce242-f515-4f81-838c-5b0684222eb4\") " pod="openshift-dns/dns-default-rx6sh" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.481050 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/eba7a4e4-d755-4368-a487-43f200d258b4-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-kps6n\" (UID: \"eba7a4e4-d755-4368-a487-43f200d258b4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kps6n" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.481259 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/db41ea16-6481-4ed9-a935-d14544e552a0-etcd-ca\") pod \"etcd-operator-b45778765-v6pjk\" (UID: \"db41ea16-6481-4ed9-a935-d14544e552a0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-v6pjk" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.480649 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/3fb2355d-5ec2-44f1-8665-a3d73c24f8a6-ca-trust-extracted\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.481273 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/db41ea16-6481-4ed9-a935-d14544e552a0-etcd-client\") pod \"etcd-operator-b45778765-v6pjk\" (UID: \"db41ea16-6481-4ed9-a935-d14544e552a0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-v6pjk" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.481313 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qk4p5\" (UniqueName: \"kubernetes.io/projected/486e2f09-ddfc-4058-8749-bf747fc829ca-kube-api-access-qk4p5\") pod \"machine-config-controller-84d6567774-w5k28\" (UID: \"486e2f09-ddfc-4058-8749-bf747fc829ca\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w5k28" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.481422 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/3fb2355d-5ec2-44f1-8665-a3d73c24f8a6-registry-certificates\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.481539 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eba7a4e4-d755-4368-a487-43f200d258b4-serving-cert\") pod \"apiserver-7bbb656c7d-kps6n\" (UID: \"eba7a4e4-d755-4368-a487-43f200d258b4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kps6n" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.481688 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/3fb2355d-5ec2-44f1-8665-a3d73c24f8a6-installation-pull-secrets\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.482037 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/eba7a4e4-d755-4368-a487-43f200d258b4-etcd-client\") pod \"apiserver-7bbb656c7d-kps6n\" (UID: \"eba7a4e4-d755-4368-a487-43f200d258b4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kps6n" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.482389 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/3fb2355d-5ec2-44f1-8665-a3d73c24f8a6-registry-tls\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.482417 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/66798226-d262-4925-82db-2742d96a33fc-metrics-certs\") pod \"router-default-5444994796-p92sp\" (UID: \"66798226-d262-4925-82db-2742d96a33fc\") " pod="openshift-ingress/router-default-5444994796-p92sp" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.482437 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fbtbb\" (UniqueName: \"kubernetes.io/projected/c82133bf-deb1-4b25-b002-f435a64847cd-kube-api-access-fbtbb\") pod \"machine-config-server-kc2wp\" (UID: \"c82133bf-deb1-4b25-b002-f435a64847cd\") " pod="openshift-machine-config-operator/machine-config-server-kc2wp" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.482495 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/486e2f09-ddfc-4058-8749-bf747fc829ca-proxy-tls\") pod \"machine-config-controller-84d6567774-w5k28\" (UID: \"486e2f09-ddfc-4058-8749-bf747fc829ca\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w5k28" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.482547 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/db778f33-712d-42fc-849e-20e0af125a5a-serving-cert\") pod \"service-ca-operator-777779d784-prvdm\" (UID: \"db778f33-712d-42fc-849e-20e0af125a5a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-prvdm" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.482791 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/eba7a4e4-d755-4368-a487-43f200d258b4-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-kps6n\" (UID: \"eba7a4e4-d755-4368-a487-43f200d258b4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kps6n" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.482930 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/eba7a4e4-d755-4368-a487-43f200d258b4-encryption-config\") pod \"apiserver-7bbb656c7d-kps6n\" (UID: \"eba7a4e4-d755-4368-a487-43f200d258b4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kps6n" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.483199 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/db41ea16-6481-4ed9-a935-d14544e552a0-serving-cert\") pod \"etcd-operator-b45778765-v6pjk\" (UID: \"db41ea16-6481-4ed9-a935-d14544e552a0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-v6pjk" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.483860 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3fb2355d-5ec2-44f1-8665-a3d73c24f8a6-trusted-ca\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.484126 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/db41ea16-6481-4ed9-a935-d14544e552a0-etcd-service-ca\") pod \"etcd-operator-b45778765-v6pjk\" (UID: \"db41ea16-6481-4ed9-a935-d14544e552a0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-v6pjk" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.484194 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jj84j\" (UniqueName: \"kubernetes.io/projected/3e66624b-9d0d-4440-9cd6-06ca77cb7566-kube-api-access-jj84j\") pod \"ingress-canary-cx6h6\" (UID: \"3e66624b-9d0d-4440-9cd6-06ca77cb7566\") " pod="openshift-ingress-canary/ingress-canary-cx6h6" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.484276 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/f97c276b-2b69-49a0-96d1-b1c5171a542e-plugins-dir\") pod \"csi-hostpathplugin-tsb45\" (UID: \"f97c276b-2b69-49a0-96d1-b1c5171a542e\") " pod="hostpath-provisioner/csi-hostpathplugin-tsb45" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.484537 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7bl5\" (UniqueName: \"kubernetes.io/projected/db41ea16-6481-4ed9-a935-d14544e552a0-kube-api-access-c7bl5\") pod \"etcd-operator-b45778765-v6pjk\" (UID: \"db41ea16-6481-4ed9-a935-d14544e552a0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-v6pjk" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.484565 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a5033129-ba8c-488f-982e-78e12b2f1664-config-volume\") pod \"collect-profiles-29401110-dm86l\" (UID: \"a5033129-ba8c-488f-982e-78e12b2f1664\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401110-dm86l" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.484569 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/db41ea16-6481-4ed9-a935-d14544e552a0-etcd-service-ca\") pod \"etcd-operator-b45778765-v6pjk\" (UID: \"db41ea16-6481-4ed9-a935-d14544e552a0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-v6pjk" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.484590 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/486e2f09-ddfc-4058-8749-bf747fc829ca-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-w5k28\" (UID: \"486e2f09-ddfc-4058-8749-bf747fc829ca\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w5k28" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.484615 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d1e46310-3643-4e6a-9624-467e4e00b093-proxy-tls\") pod \"machine-config-operator-74547568cd-s7j8m\" (UID: \"d1e46310-3643-4e6a-9624-467e4e00b093\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s7j8m" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.484684 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/f97c276b-2b69-49a0-96d1-b1c5171a542e-socket-dir\") pod \"csi-hostpathplugin-tsb45\" (UID: \"f97c276b-2b69-49a0-96d1-b1c5171a542e\") " pod="hostpath-provisioner/csi-hostpathplugin-tsb45" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.484708 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/f97c276b-2b69-49a0-96d1-b1c5171a542e-mountpoint-dir\") pod \"csi-hostpathplugin-tsb45\" (UID: \"f97c276b-2b69-49a0-96d1-b1c5171a542e\") " pod="hostpath-provisioner/csi-hostpathplugin-tsb45" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.484740 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/475ce242-f515-4f81-838c-5b0684222eb4-config-volume\") pod \"dns-default-rx6sh\" (UID: \"475ce242-f515-4f81-838c-5b0684222eb4\") " pod="openshift-dns/dns-default-rx6sh" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.484762 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slgnm\" (UniqueName: \"kubernetes.io/projected/66798226-d262-4925-82db-2742d96a33fc-kube-api-access-slgnm\") pod \"router-default-5444994796-p92sp\" (UID: \"66798226-d262-4925-82db-2742d96a33fc\") " pod="openshift-ingress/router-default-5444994796-p92sp" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.484952 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3fb2355d-5ec2-44f1-8665-a3d73c24f8a6-trusted-ca\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.485266 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/eba7a4e4-d755-4368-a487-43f200d258b4-encryption-config\") pod \"apiserver-7bbb656c7d-kps6n\" (UID: \"eba7a4e4-d755-4368-a487-43f200d258b4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kps6n" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.485480 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/db41ea16-6481-4ed9-a935-d14544e552a0-serving-cert\") pod \"etcd-operator-b45778765-v6pjk\" (UID: \"db41ea16-6481-4ed9-a935-d14544e552a0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-v6pjk" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.485619 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/486e2f09-ddfc-4058-8749-bf747fc829ca-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-w5k28\" (UID: \"486e2f09-ddfc-4058-8749-bf747fc829ca\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w5k28" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.486450 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/3fb2355d-5ec2-44f1-8665-a3d73c24f8a6-registry-tls\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.511343 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cq9vq\" (UniqueName: \"kubernetes.io/projected/db778f33-712d-42fc-849e-20e0af125a5a-kube-api-access-cq9vq\") pod \"service-ca-operator-777779d784-prvdm\" (UID: \"db778f33-712d-42fc-849e-20e0af125a5a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-prvdm" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.536010 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-v88lg"] Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.536142 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgwn7\" (UniqueName: \"kubernetes.io/projected/3fb2355d-5ec2-44f1-8665-a3d73c24f8a6-kube-api-access-lgwn7\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.538228 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/486e2f09-ddfc-4058-8749-bf747fc829ca-proxy-tls\") pod \"machine-config-controller-84d6567774-w5k28\" (UID: \"486e2f09-ddfc-4058-8749-bf747fc829ca\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w5k28" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.538265 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-p2jmh"] Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.578210 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8rvk\" (UniqueName: \"kubernetes.io/projected/eba7a4e4-d755-4368-a487-43f200d258b4-kube-api-access-m8rvk\") pod \"apiserver-7bbb656c7d-kps6n\" (UID: \"eba7a4e4-d755-4368-a487-43f200d258b4\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kps6n" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.585448 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.585788 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/66798226-d262-4925-82db-2742d96a33fc-default-certificate\") pod \"router-default-5444994796-p92sp\" (UID: \"66798226-d262-4925-82db-2742d96a33fc\") " pod="openshift-ingress/router-default-5444994796-p92sp" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.585813 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/d1e46310-3643-4e6a-9624-467e4e00b093-images\") pod \"machine-config-operator-74547568cd-s7j8m\" (UID: \"d1e46310-3643-4e6a-9624-467e4e00b093\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s7j8m" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.585834 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d1e46310-3643-4e6a-9624-467e4e00b093-auth-proxy-config\") pod \"machine-config-operator-74547568cd-s7j8m\" (UID: \"d1e46310-3643-4e6a-9624-467e4e00b093\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s7j8m" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.585887 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7v7zm\" (UniqueName: \"kubernetes.io/projected/9276b1a7-e1c8-4cdb-a77c-4b09df17be76-kube-api-access-7v7zm\") pod \"migrator-59844c95c7-898jb\" (UID: \"9276b1a7-e1c8-4cdb-a77c-4b09df17be76\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-898jb" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.585908 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/66798226-d262-4925-82db-2742d96a33fc-service-ca-bundle\") pod \"router-default-5444994796-p92sp\" (UID: \"66798226-d262-4925-82db-2742d96a33fc\") " pod="openshift-ingress/router-default-5444994796-p92sp" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.585949 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/c82133bf-deb1-4b25-b002-f435a64847cd-certs\") pod \"machine-config-server-kc2wp\" (UID: \"c82133bf-deb1-4b25-b002-f435a64847cd\") " pod="openshift-machine-config-operator/machine-config-server-kc2wp" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.585968 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7lg4\" (UniqueName: \"kubernetes.io/projected/d1e46310-3643-4e6a-9624-467e4e00b093-kube-api-access-z7lg4\") pod \"machine-config-operator-74547568cd-s7j8m\" (UID: \"d1e46310-3643-4e6a-9624-467e4e00b093\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s7j8m" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.585984 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g52j5\" (UniqueName: \"kubernetes.io/projected/a5033129-ba8c-488f-982e-78e12b2f1664-kube-api-access-g52j5\") pod \"collect-profiles-29401110-dm86l\" (UID: \"a5033129-ba8c-488f-982e-78e12b2f1664\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401110-dm86l" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.586021 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a5033129-ba8c-488f-982e-78e12b2f1664-secret-volume\") pod \"collect-profiles-29401110-dm86l\" (UID: \"a5033129-ba8c-488f-982e-78e12b2f1664\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401110-dm86l" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.586042 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/66798226-d262-4925-82db-2742d96a33fc-stats-auth\") pod \"router-default-5444994796-p92sp\" (UID: \"66798226-d262-4925-82db-2742d96a33fc\") " pod="openshift-ingress/router-default-5444994796-p92sp" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.586070 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/c82133bf-deb1-4b25-b002-f435a64847cd-node-bootstrap-token\") pod \"machine-config-server-kc2wp\" (UID: \"c82133bf-deb1-4b25-b002-f435a64847cd\") " pod="openshift-machine-config-operator/machine-config-server-kc2wp" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.586119 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/f97c276b-2b69-49a0-96d1-b1c5171a542e-registration-dir\") pod \"csi-hostpathplugin-tsb45\" (UID: \"f97c276b-2b69-49a0-96d1-b1c5171a542e\") " pod="hostpath-provisioner/csi-hostpathplugin-tsb45" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.586136 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qh7h7\" (UniqueName: \"kubernetes.io/projected/f97c276b-2b69-49a0-96d1-b1c5171a542e-kube-api-access-qh7h7\") pod \"csi-hostpathplugin-tsb45\" (UID: \"f97c276b-2b69-49a0-96d1-b1c5171a542e\") " pod="hostpath-provisioner/csi-hostpathplugin-tsb45" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.586153 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/475ce242-f515-4f81-838c-5b0684222eb4-metrics-tls\") pod \"dns-default-rx6sh\" (UID: \"475ce242-f515-4f81-838c-5b0684222eb4\") " pod="openshift-dns/dns-default-rx6sh" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.586232 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/f97c276b-2b69-49a0-96d1-b1c5171a542e-csi-data-dir\") pod \"csi-hostpathplugin-tsb45\" (UID: \"f97c276b-2b69-49a0-96d1-b1c5171a542e\") " pod="hostpath-provisioner/csi-hostpathplugin-tsb45" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.586285 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3e66624b-9d0d-4440-9cd6-06ca77cb7566-cert\") pod \"ingress-canary-cx6h6\" (UID: \"3e66624b-9d0d-4440-9cd6-06ca77cb7566\") " pod="openshift-ingress-canary/ingress-canary-cx6h6" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.586307 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4mnz\" (UniqueName: \"kubernetes.io/projected/475ce242-f515-4f81-838c-5b0684222eb4-kube-api-access-q4mnz\") pod \"dns-default-rx6sh\" (UID: \"475ce242-f515-4f81-838c-5b0684222eb4\") " pod="openshift-dns/dns-default-rx6sh" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.586334 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/66798226-d262-4925-82db-2742d96a33fc-metrics-certs\") pod \"router-default-5444994796-p92sp\" (UID: \"66798226-d262-4925-82db-2742d96a33fc\") " pod="openshift-ingress/router-default-5444994796-p92sp" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.586384 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fbtbb\" (UniqueName: \"kubernetes.io/projected/c82133bf-deb1-4b25-b002-f435a64847cd-kube-api-access-fbtbb\") pod \"machine-config-server-kc2wp\" (UID: \"c82133bf-deb1-4b25-b002-f435a64847cd\") " pod="openshift-machine-config-operator/machine-config-server-kc2wp" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.586430 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jj84j\" (UniqueName: \"kubernetes.io/projected/3e66624b-9d0d-4440-9cd6-06ca77cb7566-kube-api-access-jj84j\") pod \"ingress-canary-cx6h6\" (UID: \"3e66624b-9d0d-4440-9cd6-06ca77cb7566\") " pod="openshift-ingress-canary/ingress-canary-cx6h6" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.586451 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/f97c276b-2b69-49a0-96d1-b1c5171a542e-plugins-dir\") pod \"csi-hostpathplugin-tsb45\" (UID: \"f97c276b-2b69-49a0-96d1-b1c5171a542e\") " pod="hostpath-provisioner/csi-hostpathplugin-tsb45" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.586484 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a5033129-ba8c-488f-982e-78e12b2f1664-config-volume\") pod \"collect-profiles-29401110-dm86l\" (UID: \"a5033129-ba8c-488f-982e-78e12b2f1664\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401110-dm86l" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.586506 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d1e46310-3643-4e6a-9624-467e4e00b093-proxy-tls\") pod \"machine-config-operator-74547568cd-s7j8m\" (UID: \"d1e46310-3643-4e6a-9624-467e4e00b093\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s7j8m" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.586531 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/f97c276b-2b69-49a0-96d1-b1c5171a542e-socket-dir\") pod \"csi-hostpathplugin-tsb45\" (UID: \"f97c276b-2b69-49a0-96d1-b1c5171a542e\") " pod="hostpath-provisioner/csi-hostpathplugin-tsb45" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.586551 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/f97c276b-2b69-49a0-96d1-b1c5171a542e-mountpoint-dir\") pod \"csi-hostpathplugin-tsb45\" (UID: \"f97c276b-2b69-49a0-96d1-b1c5171a542e\") " pod="hostpath-provisioner/csi-hostpathplugin-tsb45" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.586573 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/475ce242-f515-4f81-838c-5b0684222eb4-config-volume\") pod \"dns-default-rx6sh\" (UID: \"475ce242-f515-4f81-838c-5b0684222eb4\") " pod="openshift-dns/dns-default-rx6sh" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.586596 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slgnm\" (UniqueName: \"kubernetes.io/projected/66798226-d262-4925-82db-2742d96a33fc-kube-api-access-slgnm\") pod \"router-default-5444994796-p92sp\" (UID: \"66798226-d262-4925-82db-2742d96a33fc\") " pod="openshift-ingress/router-default-5444994796-p92sp" Nov 25 10:34:31 crc kubenswrapper[4821]: E1125 10:34:31.586834 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:32.08681402 +0000 UTC m=+142.623133867 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.587556 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/f97c276b-2b69-49a0-96d1-b1c5171a542e-csi-data-dir\") pod \"csi-hostpathplugin-tsb45\" (UID: \"f97c276b-2b69-49a0-96d1-b1c5171a542e\") " pod="hostpath-provisioner/csi-hostpathplugin-tsb45" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.588505 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/d1e46310-3643-4e6a-9624-467e4e00b093-images\") pod \"machine-config-operator-74547568cd-s7j8m\" (UID: \"d1e46310-3643-4e6a-9624-467e4e00b093\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s7j8m" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.589051 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d1e46310-3643-4e6a-9624-467e4e00b093-auth-proxy-config\") pod \"machine-config-operator-74547568cd-s7j8m\" (UID: \"d1e46310-3643-4e6a-9624-467e4e00b093\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s7j8m" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.589573 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/f97c276b-2b69-49a0-96d1-b1c5171a542e-registration-dir\") pod \"csi-hostpathplugin-tsb45\" (UID: \"f97c276b-2b69-49a0-96d1-b1c5171a542e\") " pod="hostpath-provisioner/csi-hostpathplugin-tsb45" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.589913 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/66798226-d262-4925-82db-2742d96a33fc-service-ca-bundle\") pod \"router-default-5444994796-p92sp\" (UID: \"66798226-d262-4925-82db-2742d96a33fc\") " pod="openshift-ingress/router-default-5444994796-p92sp" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.590604 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/f97c276b-2b69-49a0-96d1-b1c5171a542e-mountpoint-dir\") pod \"csi-hostpathplugin-tsb45\" (UID: \"f97c276b-2b69-49a0-96d1-b1c5171a542e\") " pod="hostpath-provisioner/csi-hostpathplugin-tsb45" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.590702 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/f97c276b-2b69-49a0-96d1-b1c5171a542e-socket-dir\") pod \"csi-hostpathplugin-tsb45\" (UID: \"f97c276b-2b69-49a0-96d1-b1c5171a542e\") " pod="hostpath-provisioner/csi-hostpathplugin-tsb45" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.591016 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/f97c276b-2b69-49a0-96d1-b1c5171a542e-plugins-dir\") pod \"csi-hostpathplugin-tsb45\" (UID: \"f97c276b-2b69-49a0-96d1-b1c5171a542e\") " pod="hostpath-provisioner/csi-hostpathplugin-tsb45" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.591414 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/475ce242-f515-4f81-838c-5b0684222eb4-config-volume\") pod \"dns-default-rx6sh\" (UID: \"475ce242-f515-4f81-838c-5b0684222eb4\") " pod="openshift-dns/dns-default-rx6sh" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.592156 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a5033129-ba8c-488f-982e-78e12b2f1664-config-volume\") pod \"collect-profiles-29401110-dm86l\" (UID: \"a5033129-ba8c-488f-982e-78e12b2f1664\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401110-dm86l" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.594144 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/475ce242-f515-4f81-838c-5b0684222eb4-metrics-tls\") pod \"dns-default-rx6sh\" (UID: \"475ce242-f515-4f81-838c-5b0684222eb4\") " pod="openshift-dns/dns-default-rx6sh" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.594852 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a5033129-ba8c-488f-982e-78e12b2f1664-secret-volume\") pod \"collect-profiles-29401110-dm86l\" (UID: \"a5033129-ba8c-488f-982e-78e12b2f1664\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401110-dm86l" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.594886 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/c82133bf-deb1-4b25-b002-f435a64847cd-node-bootstrap-token\") pod \"machine-config-server-kc2wp\" (UID: \"c82133bf-deb1-4b25-b002-f435a64847cd\") " pod="openshift-machine-config-operator/machine-config-server-kc2wp" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.598414 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/c82133bf-deb1-4b25-b002-f435a64847cd-certs\") pod \"machine-config-server-kc2wp\" (UID: \"c82133bf-deb1-4b25-b002-f435a64847cd\") " pod="openshift-machine-config-operator/machine-config-server-kc2wp" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.598754 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3e66624b-9d0d-4440-9cd6-06ca77cb7566-cert\") pod \"ingress-canary-cx6h6\" (UID: \"3e66624b-9d0d-4440-9cd6-06ca77cb7566\") " pod="openshift-ingress-canary/ingress-canary-cx6h6" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.599297 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qk4p5\" (UniqueName: \"kubernetes.io/projected/486e2f09-ddfc-4058-8749-bf747fc829ca-kube-api-access-qk4p5\") pod \"machine-config-controller-84d6567774-w5k28\" (UID: \"486e2f09-ddfc-4058-8749-bf747fc829ca\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w5k28" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.612645 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d1e46310-3643-4e6a-9624-467e4e00b093-proxy-tls\") pod \"machine-config-operator-74547568cd-s7j8m\" (UID: \"d1e46310-3643-4e6a-9624-467e4e00b093\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s7j8m" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.617417 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3fb2355d-5ec2-44f1-8665-a3d73c24f8a6-bound-sa-token\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.663937 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w5k28" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.672187 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kps6n" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.674592 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-t5hm7" event={"ID":"b75a7773-a689-46c1-b98e-e7c1be0878e6","Type":"ContainerStarted","Data":"78978133f1fc390c08abab9e06855a43321af09301ff908d66cb3dcd5e78513a"} Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.675506 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-zj4s2" event={"ID":"24708c13-efe2-435b-acde-b29b48871f16","Type":"ContainerStarted","Data":"bc4bd7d35a2da4dde6059e18357604abf39d9dbd5f8d4e38bdd3242798d2eb38"} Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.676816 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5m5gj" event={"ID":"cb0f66e4-2c0f-4a95-89db-1aef53ae5b3b","Type":"ContainerStarted","Data":"3b4d1a301ae80bb352cffc0a098d0a292d7c7b887a5eb9b155bf3518c2313cf0"} Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.678379 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4kjbr" event={"ID":"2c2ecf12-49bd-4baf-90df-68486f27be1b","Type":"ContainerStarted","Data":"7f8c7a28f9eed13092ecec1048f8f01fc120a25a1b9605daee4507f83b5c103b"} Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.687327 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:31 crc kubenswrapper[4821]: E1125 10:34:31.687730 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:32.187711302 +0000 UTC m=+142.724031199 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.687883 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-hddc7" event={"ID":"b0e34420-095d-42aa-89c4-5a80331e1c01","Type":"ContainerStarted","Data":"3785ec3b855b5e19b0eae30b7de1f7463643536065864be2d45853297c7c2cd7"} Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.689560 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-86cnx" event={"ID":"59ec70c2-c208-4529-a763-a0d7ea7e1dfc","Type":"ContainerStarted","Data":"a663ce0339455eec407ece5b4822f07c80965045739b8a080e36da281d2c728d"} Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.690214 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zvztq" event={"ID":"7fe78aa4-571c-488b-96bd-23ae4b6333db","Type":"ContainerStarted","Data":"5d9be55890aa29ed5a7b436a5598936483cc0c9742c84e6731f95412ab3db098"} Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.705197 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-hf49d" event={"ID":"88725123-3da9-4127-9506-80c90c9accd5","Type":"ContainerStarted","Data":"fa6ed8bb408c16d09537a52f993d7362b7b95f6699f1285cbfb2a8959b734528"} Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.705295 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-prvdm" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.706213 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jj84j\" (UniqueName: \"kubernetes.io/projected/3e66624b-9d0d-4440-9cd6-06ca77cb7566-kube-api-access-jj84j\") pod \"ingress-canary-cx6h6\" (UID: \"3e66624b-9d0d-4440-9cd6-06ca77cb7566\") " pod="openshift-ingress-canary/ingress-canary-cx6h6" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.707101 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-p2jmh" event={"ID":"402cf28b-25d2-428a-ab7b-1d70345a921e","Type":"ContainerStarted","Data":"b1d7f7fbe220468012de8bc79ce5ea95922234566eda370f427518dce1e8beca"} Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.708205 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-mbn4k" event={"ID":"88feae6a-5031-48ed-93ec-34ef8d2eff40","Type":"ContainerStarted","Data":"d9d52c11eb28b942ee927fd912e4bb08f1ffc0b522c3507677b9105434e099d9"} Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.710226 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-v88lg" event={"ID":"155595fc-76b3-4561-bd3f-c540076d4cd6","Type":"ContainerStarted","Data":"637f9c31e944fa641f3b53a33e2683c244a541220e9f8495564c82732666b154"} Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.711262 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dv9nt" event={"ID":"ba27af47-1944-4970-a015-e96c7cee4862","Type":"ContainerStarted","Data":"21ee34f3e4c08e218c5e2422993cd523869563687b4357d81391e29fc85026c3"} Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.712011 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/66798226-d262-4925-82db-2742d96a33fc-stats-auth\") pod \"router-default-5444994796-p92sp\" (UID: \"66798226-d262-4925-82db-2742d96a33fc\") " pod="openshift-ingress/router-default-5444994796-p92sp" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.712304 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/66798226-d262-4925-82db-2742d96a33fc-metrics-certs\") pod \"router-default-5444994796-p92sp\" (UID: \"66798226-d262-4925-82db-2742d96a33fc\") " pod="openshift-ingress/router-default-5444994796-p92sp" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.712463 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/66798226-d262-4925-82db-2742d96a33fc-default-certificate\") pod \"router-default-5444994796-p92sp\" (UID: \"66798226-d262-4925-82db-2742d96a33fc\") " pod="openshift-ingress/router-default-5444994796-p92sp" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.716026 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slgnm\" (UniqueName: \"kubernetes.io/projected/66798226-d262-4925-82db-2742d96a33fc-kube-api-access-slgnm\") pod \"router-default-5444994796-p92sp\" (UID: \"66798226-d262-4925-82db-2742d96a33fc\") " pod="openshift-ingress/router-default-5444994796-p92sp" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.716028 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4mnz\" (UniqueName: \"kubernetes.io/projected/475ce242-f515-4f81-838c-5b0684222eb4-kube-api-access-q4mnz\") pod \"dns-default-rx6sh\" (UID: \"475ce242-f515-4f81-838c-5b0684222eb4\") " pod="openshift-dns/dns-default-rx6sh" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.727188 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7bl5\" (UniqueName: \"kubernetes.io/projected/db41ea16-6481-4ed9-a935-d14544e552a0-kube-api-access-c7bl5\") pod \"etcd-operator-b45778765-v6pjk\" (UID: \"db41ea16-6481-4ed9-a935-d14544e552a0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-v6pjk" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.729677 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-58htt" event={"ID":"518db853-3bbe-421b-a2dd-6d1b2e965120","Type":"ContainerStarted","Data":"903cb8e95459b26e7c1e4806fc6c432eac8bea8cdc1e84013d939af20b308de5"} Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.734151 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-q8tq8" event={"ID":"5a7c1a05-b29c-4ff3-a4d6-90fe59eb1a01","Type":"ContainerStarted","Data":"9aa0f58c3cdca0630e0cf35cf767332e225651691b81534eab3363f683cccc60"} Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.734682 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7lg4\" (UniqueName: \"kubernetes.io/projected/d1e46310-3643-4e6a-9624-467e4e00b093-kube-api-access-z7lg4\") pod \"machine-config-operator-74547568cd-s7j8m\" (UID: \"d1e46310-3643-4e6a-9624-467e4e00b093\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s7j8m" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.756923 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s7j8m" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.759124 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fbtbb\" (UniqueName: \"kubernetes.io/projected/c82133bf-deb1-4b25-b002-f435a64847cd-kube-api-access-fbtbb\") pod \"machine-config-server-kc2wp\" (UID: \"c82133bf-deb1-4b25-b002-f435a64847cd\") " pod="openshift-machine-config-operator/machine-config-server-kc2wp" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.769932 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-v6pjk" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.775645 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-n4lbr"] Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.777712 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7v7zm\" (UniqueName: \"kubernetes.io/projected/9276b1a7-e1c8-4cdb-a77c-4b09df17be76-kube-api-access-7v7zm\") pod \"migrator-59844c95c7-898jb\" (UID: \"9276b1a7-e1c8-4cdb-a77c-4b09df17be76\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-898jb" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.782045 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-rx6sh" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.788300 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:31 crc kubenswrapper[4821]: E1125 10:34:31.788507 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:32.28846987 +0000 UTC m=+142.824789717 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.788801 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:31 crc kubenswrapper[4821]: E1125 10:34:31.789361 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:32.289351232 +0000 UTC m=+142.825671079 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.794791 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g52j5\" (UniqueName: \"kubernetes.io/projected/a5033129-ba8c-488f-982e-78e12b2f1664-kube-api-access-g52j5\") pod \"collect-profiles-29401110-dm86l\" (UID: \"a5033129-ba8c-488f-982e-78e12b2f1664\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401110-dm86l" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.810383 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-kc2wp" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.815378 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qh7h7\" (UniqueName: \"kubernetes.io/projected/f97c276b-2b69-49a0-96d1-b1c5171a542e-kube-api-access-qh7h7\") pod \"csi-hostpathplugin-tsb45\" (UID: \"f97c276b-2b69-49a0-96d1-b1c5171a542e\") " pod="hostpath-provisioner/csi-hostpathplugin-tsb45" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.818115 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-cx6h6" Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.840149 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6xxs8"] Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.890277 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:31 crc kubenswrapper[4821]: E1125 10:34:31.890566 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:32.390528441 +0000 UTC m=+142.926848328 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.954485 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-tpbzb"] Nov 25 10:34:31 crc kubenswrapper[4821]: W1125 10:34:31.980361 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4878872d_8035_442a_a355_7c966f37e8c1.slice/crio-45707b44bd2686acec3d3bc4a075ca7ef887924e14702770dbc6a4e819ec99d5 WatchSource:0}: Error finding container 45707b44bd2686acec3d3bc4a075ca7ef887924e14702770dbc6a4e819ec99d5: Status 404 returned error can't find the container with id 45707b44bd2686acec3d3bc4a075ca7ef887924e14702770dbc6a4e819ec99d5 Nov 25 10:34:31 crc kubenswrapper[4821]: I1125 10:34:31.980794 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-p92sp" Nov 25 10:34:32 crc kubenswrapper[4821]: I1125 10:34:32.001840 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:32 crc kubenswrapper[4821]: E1125 10:34:32.002454 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:32.502438605 +0000 UTC m=+143.038758452 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:32 crc kubenswrapper[4821]: I1125 10:34:32.017005 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-hjvt9"] Nov 25 10:34:32 crc kubenswrapper[4821]: I1125 10:34:32.037659 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401110-dm86l" Nov 25 10:34:32 crc kubenswrapper[4821]: I1125 10:34:32.043378 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-898jb" Nov 25 10:34:32 crc kubenswrapper[4821]: I1125 10:34:32.082946 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-f2tjh"] Nov 25 10:34:32 crc kubenswrapper[4821]: I1125 10:34:32.100523 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-tsb45" Nov 25 10:34:32 crc kubenswrapper[4821]: I1125 10:34:32.103831 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:32 crc kubenswrapper[4821]: E1125 10:34:32.104264 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:32.60424572 +0000 UTC m=+143.140565567 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:32 crc kubenswrapper[4821]: I1125 10:34:32.205838 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:32 crc kubenswrapper[4821]: E1125 10:34:32.206317 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:32.70629867 +0000 UTC m=+143.242618607 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:32 crc kubenswrapper[4821]: I1125 10:34:32.225002 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-jb5sz"] Nov 25 10:34:32 crc kubenswrapper[4821]: I1125 10:34:32.309693 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:32 crc kubenswrapper[4821]: E1125 10:34:32.309835 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:32.809814567 +0000 UTC m=+143.346134414 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:32 crc kubenswrapper[4821]: I1125 10:34:32.310389 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:32 crc kubenswrapper[4821]: E1125 10:34:32.310769 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:32.81075015 +0000 UTC m=+143.347070057 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:32 crc kubenswrapper[4821]: I1125 10:34:32.413852 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:32 crc kubenswrapper[4821]: E1125 10:34:32.414240 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:32.914223956 +0000 UTC m=+143.450543803 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:32 crc kubenswrapper[4821]: I1125 10:34:32.520066 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:32 crc kubenswrapper[4821]: E1125 10:34:32.520646 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:33.020632965 +0000 UTC m=+143.556952812 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:32 crc kubenswrapper[4821]: I1125 10:34:32.629471 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:32 crc kubenswrapper[4821]: E1125 10:34:32.629854 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:33.129835882 +0000 UTC m=+143.666155739 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:32 crc kubenswrapper[4821]: I1125 10:34:32.722633 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2stv6"] Nov 25 10:34:32 crc kubenswrapper[4821]: I1125 10:34:32.734434 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:32 crc kubenswrapper[4821]: E1125 10:34:32.739728 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:33.239710566 +0000 UTC m=+143.776030413 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:32 crc kubenswrapper[4821]: I1125 10:34:32.746964 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-kc2wp" event={"ID":"c82133bf-deb1-4b25-b002-f435a64847cd","Type":"ContainerStarted","Data":"26c4b09953d2805dc100c237ab6abcc234f026f2fc434881dc03f4f6a641d7c2"} Nov 25 10:34:32 crc kubenswrapper[4821]: I1125 10:34:32.751299 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-whcwc"] Nov 25 10:34:32 crc kubenswrapper[4821]: I1125 10:34:32.751655 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-hjvt9" event={"ID":"f954533e-e8a0-4899-be72-241760feae95","Type":"ContainerStarted","Data":"3e39e33e9657337ae7776e7fdb545656d2a2578ba246672857cbdd9d2994a0b4"} Nov 25 10:34:32 crc kubenswrapper[4821]: I1125 10:34:32.755302 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k622z" event={"ID":"513166e2-0ac9-4299-bf84-d994a46da192","Type":"ContainerStarted","Data":"631e4f9a3b5b7e3daa6cc35cc8cab7759db678ddaaf4b8a8951bc1ad81263dd4"} Nov 25 10:34:32 crc kubenswrapper[4821]: I1125 10:34:32.757007 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d8x5t"] Nov 25 10:34:32 crc kubenswrapper[4821]: I1125 10:34:32.762997 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-prvdm"] Nov 25 10:34:32 crc kubenswrapper[4821]: I1125 10:34:32.764905 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-t5hm7" event={"ID":"b75a7773-a689-46c1-b98e-e7c1be0878e6","Type":"ContainerStarted","Data":"db05a13923848cff0148e545acb0c93ec76e08aa031620da486b8643220d7b77"} Nov 25 10:34:32 crc kubenswrapper[4821]: I1125 10:34:32.765149 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-t5hm7" Nov 25 10:34:32 crc kubenswrapper[4821]: I1125 10:34:32.768173 4821 patch_prober.go:28] interesting pod/console-operator-58897d9998-t5hm7 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/readyz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Nov 25 10:34:32 crc kubenswrapper[4821]: I1125 10:34:32.768221 4821 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-t5hm7" podUID="b75a7773-a689-46c1-b98e-e7c1be0878e6" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.6:8443/readyz\": dial tcp 10.217.0.6:8443: connect: connection refused" Nov 25 10:34:32 crc kubenswrapper[4821]: I1125 10:34:32.769431 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5m5gj" event={"ID":"cb0f66e4-2c0f-4a95-89db-1aef53ae5b3b","Type":"ContainerStarted","Data":"1c8741874c24e79014d0ca54a32d493fe7ac379a8bc7465cd1fe572da031900d"} Nov 25 10:34:32 crc kubenswrapper[4821]: I1125 10:34:32.772774 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-hf49d" event={"ID":"88725123-3da9-4127-9506-80c90c9accd5","Type":"ContainerStarted","Data":"3dd116757d3852f02a7dd5f0a100fafa105ad3f81d3d6be8c6c22129dab37315"} Nov 25 10:34:32 crc kubenswrapper[4821]: I1125 10:34:32.774060 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-7792m"] Nov 25 10:34:32 crc kubenswrapper[4821]: I1125 10:34:32.778937 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-tpbzb" event={"ID":"9f6c1e2f-b7e7-40e7-92c7-8ca1b57eb9f1","Type":"ContainerStarted","Data":"d12fa41011a89baf2925f772ccafcb1efbf3a675e1f42109a41192480cb3c243"} Nov 25 10:34:32 crc kubenswrapper[4821]: I1125 10:34:32.780900 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-mbn4k" event={"ID":"88feae6a-5031-48ed-93ec-34ef8d2eff40","Type":"ContainerStarted","Data":"694393c7f8c46b2f1b6aa9777e6755df30be0d3f493bb5ab58f77f1e24aeb086"} Nov 25 10:34:32 crc kubenswrapper[4821]: I1125 10:34:32.781210 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-mbn4k" Nov 25 10:34:32 crc kubenswrapper[4821]: I1125 10:34:32.782749 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-jb5sz" event={"ID":"064de93d-7bb5-47b5-9585-bf32be9a05f6","Type":"ContainerStarted","Data":"34aeb7f4a3e22cd3dd2e26e301b405fdb21e9775fb3b47c9ca0e0a0255fb4646"} Nov 25 10:34:32 crc kubenswrapper[4821]: I1125 10:34:32.783001 4821 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-mbn4k container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.5:6443/healthz\": dial tcp 10.217.0.5:6443: connect: connection refused" start-of-body= Nov 25 10:34:32 crc kubenswrapper[4821]: I1125 10:34:32.783033 4821 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-mbn4k" podUID="88feae6a-5031-48ed-93ec-34ef8d2eff40" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.5:6443/healthz\": dial tcp 10.217.0.5:6443: connect: connection refused" Nov 25 10:34:32 crc kubenswrapper[4821]: I1125 10:34:32.784026 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-n4lbr" event={"ID":"4878872d-8035-442a-a355-7c966f37e8c1","Type":"ContainerStarted","Data":"45707b44bd2686acec3d3bc4a075ca7ef887924e14702770dbc6a4e819ec99d5"} Nov 25 10:34:32 crc kubenswrapper[4821]: I1125 10:34:32.786463 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-f2tjh" event={"ID":"d7e79465-8df8-41a7-ab16-5c730f9ceca2","Type":"ContainerStarted","Data":"7bc2214a8be9382a1c41cf54df462a6a31842f1344e0f3a9165b0baad41d1f78"} Nov 25 10:34:32 crc kubenswrapper[4821]: I1125 10:34:32.787684 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6xxs8" event={"ID":"0e82d98b-7c99-496d-910d-e9bfae2ca6ec","Type":"ContainerStarted","Data":"343ab166fbe1c1b1c8203ed7e6be8fcd9ccc32d93634cfae8dc1035e919109af"} Nov 25 10:34:32 crc kubenswrapper[4821]: I1125 10:34:32.790192 4821 generic.go:334] "Generic (PLEG): container finished" podID="5a7c1a05-b29c-4ff3-a4d6-90fe59eb1a01" containerID="742a4d14a7744ba39fcb32dc298bfadf41d5a9fd0f2c8b62f2d2f49d073da12c" exitCode=0 Nov 25 10:34:32 crc kubenswrapper[4821]: I1125 10:34:32.790250 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-q8tq8" event={"ID":"5a7c1a05-b29c-4ff3-a4d6-90fe59eb1a01","Type":"ContainerDied","Data":"742a4d14a7744ba39fcb32dc298bfadf41d5a9fd0f2c8b62f2d2f49d073da12c"} Nov 25 10:34:32 crc kubenswrapper[4821]: I1125 10:34:32.798520 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4kjbr" event={"ID":"2c2ecf12-49bd-4baf-90df-68486f27be1b","Type":"ContainerStarted","Data":"b56ba4b9adba6e4c6e1d467ea64d65ff9c660c31742218714c182d7881e2d495"} Nov 25 10:34:32 crc kubenswrapper[4821]: I1125 10:34:32.798855 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4kjbr" Nov 25 10:34:32 crc kubenswrapper[4821]: I1125 10:34:32.814790 4821 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-4kjbr container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.29:8443/healthz\": dial tcp 10.217.0.29:8443: connect: connection refused" start-of-body= Nov 25 10:34:32 crc kubenswrapper[4821]: I1125 10:34:32.815204 4821 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4kjbr" podUID="2c2ecf12-49bd-4baf-90df-68486f27be1b" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.29:8443/healthz\": dial tcp 10.217.0.29:8443: connect: connection refused" Nov 25 10:34:32 crc kubenswrapper[4821]: I1125 10:34:32.835710 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:32 crc kubenswrapper[4821]: E1125 10:34:32.836194 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:33.336158308 +0000 UTC m=+143.872478145 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:32 crc kubenswrapper[4821]: I1125 10:34:32.836650 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:32 crc kubenswrapper[4821]: E1125 10:34:32.840155 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:33.340141067 +0000 UTC m=+143.876460914 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:32 crc kubenswrapper[4821]: W1125 10:34:32.844021 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4a74ce9e_b145_4e40_ad4e_67ecdf722195.slice/crio-3d5abd23747fb90e7e1e636a330508109a09e95c0b47833b71e225275a9ef40c WatchSource:0}: Error finding container 3d5abd23747fb90e7e1e636a330508109a09e95c0b47833b71e225275a9ef40c: Status 404 returned error can't find the container with id 3d5abd23747fb90e7e1e636a330508109a09e95c0b47833b71e225275a9ef40c Nov 25 10:34:32 crc kubenswrapper[4821]: W1125 10:34:32.871763 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb18917c3_98cb_407b_97e5_5333aa5ca797.slice/crio-6e322c7619c8a09ae256d79caec90543b2c4220486027879d089751315f3bf88 WatchSource:0}: Error finding container 6e322c7619c8a09ae256d79caec90543b2c4220486027879d089751315f3bf88: Status 404 returned error can't find the container with id 6e322c7619c8a09ae256d79caec90543b2c4220486027879d089751315f3bf88 Nov 25 10:34:32 crc kubenswrapper[4821]: W1125 10:34:32.898216 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbdd88e7d_edf6_41c1_a5a3_47746bb3ab99.slice/crio-a67627cfb0e3116bef25404b270ac01cecbc65ff01961cdf2e660c4e6445724b WatchSource:0}: Error finding container a67627cfb0e3116bef25404b270ac01cecbc65ff01961cdf2e660c4e6445724b: Status 404 returned error can't find the container with id a67627cfb0e3116bef25404b270ac01cecbc65ff01961cdf2e660c4e6445724b Nov 25 10:34:32 crc kubenswrapper[4821]: W1125 10:34:32.920122 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod499ccac5_6595_4d67_93ed_dd2d19d045bf.slice/crio-a7f84213d8f358aebc938c840f63081a47c3b7ad32fea9bb437fb937e832b114 WatchSource:0}: Error finding container a7f84213d8f358aebc938c840f63081a47c3b7ad32fea9bb437fb937e832b114: Status 404 returned error can't find the container with id a7f84213d8f358aebc938c840f63081a47c3b7ad32fea9bb437fb937e832b114 Nov 25 10:34:32 crc kubenswrapper[4821]: W1125 10:34:32.920603 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddb778f33_712d_42fc_849e_20e0af125a5a.slice/crio-2ac2400aa91050e7b6331171ab8ba2370aa04605a268613880341df13c27b9a4 WatchSource:0}: Error finding container 2ac2400aa91050e7b6331171ab8ba2370aa04605a268613880341df13c27b9a4: Status 404 returned error can't find the container with id 2ac2400aa91050e7b6331171ab8ba2370aa04605a268613880341df13c27b9a4 Nov 25 10:34:32 crc kubenswrapper[4821]: I1125 10:34:32.937682 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:32 crc kubenswrapper[4821]: E1125 10:34:32.937918 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:33.43787179 +0000 UTC m=+143.974191637 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:32 crc kubenswrapper[4821]: I1125 10:34:32.938033 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:32 crc kubenswrapper[4821]: E1125 10:34:32.938535 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:33.438522776 +0000 UTC m=+143.974842623 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:33 crc kubenswrapper[4821]: I1125 10:34:33.039070 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:33 crc kubenswrapper[4821]: E1125 10:34:33.039186 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:33.539155831 +0000 UTC m=+144.075475678 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:33 crc kubenswrapper[4821]: I1125 10:34:33.039505 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:33 crc kubenswrapper[4821]: E1125 10:34:33.039796 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:33.539788746 +0000 UTC m=+144.076108593 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:33 crc kubenswrapper[4821]: I1125 10:34:33.140180 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:33 crc kubenswrapper[4821]: E1125 10:34:33.140447 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:33.640416552 +0000 UTC m=+144.176736419 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:33 crc kubenswrapper[4821]: I1125 10:34:33.140837 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:33 crc kubenswrapper[4821]: E1125 10:34:33.141182 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:33.64115218 +0000 UTC m=+144.177472027 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:33 crc kubenswrapper[4821]: I1125 10:34:33.151229 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-kps6n"] Nov 25 10:34:33 crc kubenswrapper[4821]: I1125 10:34:33.180684 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tkgh9"] Nov 25 10:34:33 crc kubenswrapper[4821]: I1125 10:34:33.190278 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-dxf4x"] Nov 25 10:34:33 crc kubenswrapper[4821]: I1125 10:34:33.198695 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" Nov 25 10:34:33 crc kubenswrapper[4821]: I1125 10:34:33.200409 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-ng2nv"] Nov 25 10:34:33 crc kubenswrapper[4821]: I1125 10:34:33.217395 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wxp2r"] Nov 25 10:34:33 crc kubenswrapper[4821]: I1125 10:34:33.242751 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:33 crc kubenswrapper[4821]: E1125 10:34:33.246654 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:33.746630965 +0000 UTC m=+144.282950842 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:33 crc kubenswrapper[4821]: I1125 10:34:33.269251 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-w5k28"] Nov 25 10:34:33 crc kubenswrapper[4821]: I1125 10:34:33.276993 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-cx6h6"] Nov 25 10:34:33 crc kubenswrapper[4821]: I1125 10:34:33.279971 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-v6pjk"] Nov 25 10:34:33 crc kubenswrapper[4821]: I1125 10:34:33.287208 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-s7j8m"] Nov 25 10:34:33 crc kubenswrapper[4821]: W1125 10:34:33.319370 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddb41ea16_6481_4ed9_a935_d14544e552a0.slice/crio-ac9ff53bf3ed708f11377bd8b6d787c33f21aaa156f9d11a128631bfb487c70d WatchSource:0}: Error finding container ac9ff53bf3ed708f11377bd8b6d787c33f21aaa156f9d11a128631bfb487c70d: Status 404 returned error can't find the container with id ac9ff53bf3ed708f11377bd8b6d787c33f21aaa156f9d11a128631bfb487c70d Nov 25 10:34:33 crc kubenswrapper[4821]: I1125 10:34:33.344800 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:33 crc kubenswrapper[4821]: E1125 10:34:33.346735 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:33.846719147 +0000 UTC m=+144.383039004 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:33 crc kubenswrapper[4821]: I1125 10:34:33.349664 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-898jb"] Nov 25 10:34:33 crc kubenswrapper[4821]: I1125 10:34:33.393931 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-tsb45"] Nov 25 10:34:33 crc kubenswrapper[4821]: I1125 10:34:33.405992 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-rx6sh"] Nov 25 10:34:33 crc kubenswrapper[4821]: I1125 10:34:33.453754 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:33 crc kubenswrapper[4821]: E1125 10:34:33.454247 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:33.954226962 +0000 UTC m=+144.490546809 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:33 crc kubenswrapper[4821]: I1125 10:34:33.470611 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-mbn4k" podStartSLOduration=123.470594718 podStartE2EDuration="2m3.470594718s" podCreationTimestamp="2025-11-25 10:32:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:34:33.45610728 +0000 UTC m=+143.992427207" watchObservedRunningTime="2025-11-25 10:34:33.470594718 +0000 UTC m=+144.006914575" Nov 25 10:34:33 crc kubenswrapper[4821]: I1125 10:34:33.490982 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401110-dm86l"] Nov 25 10:34:33 crc kubenswrapper[4821]: W1125 10:34:33.525768 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda5033129_ba8c_488f_982e_78e12b2f1664.slice/crio-f647a6676a61dff0159b32ea289553f18ca7336aafb1fec95c26f25c594d6e9e WatchSource:0}: Error finding container f647a6676a61dff0159b32ea289553f18ca7336aafb1fec95c26f25c594d6e9e: Status 404 returned error can't find the container with id f647a6676a61dff0159b32ea289553f18ca7336aafb1fec95c26f25c594d6e9e Nov 25 10:34:33 crc kubenswrapper[4821]: I1125 10:34:33.533741 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4kjbr" podStartSLOduration=123.533707793 podStartE2EDuration="2m3.533707793s" podCreationTimestamp="2025-11-25 10:32:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:34:33.532393671 +0000 UTC m=+144.068713518" watchObservedRunningTime="2025-11-25 10:34:33.533707793 +0000 UTC m=+144.070027650" Nov 25 10:34:33 crc kubenswrapper[4821]: I1125 10:34:33.549779 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5m5gj" podStartSLOduration=123.549760082 podStartE2EDuration="2m3.549760082s" podCreationTimestamp="2025-11-25 10:32:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:34:33.547492785 +0000 UTC m=+144.083812632" watchObservedRunningTime="2025-11-25 10:34:33.549760082 +0000 UTC m=+144.086079929" Nov 25 10:34:33 crc kubenswrapper[4821]: I1125 10:34:33.555126 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:33 crc kubenswrapper[4821]: E1125 10:34:33.555530 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:34.055512764 +0000 UTC m=+144.591832681 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:33 crc kubenswrapper[4821]: I1125 10:34:33.571609 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-t5hm7" podStartSLOduration=123.571586552 podStartE2EDuration="2m3.571586552s" podCreationTimestamp="2025-11-25 10:32:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:34:33.568312332 +0000 UTC m=+144.104632179" watchObservedRunningTime="2025-11-25 10:34:33.571586552 +0000 UTC m=+144.107906399" Nov 25 10:34:33 crc kubenswrapper[4821]: I1125 10:34:33.664717 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:33 crc kubenswrapper[4821]: E1125 10:34:33.680852 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:34.180815741 +0000 UTC m=+144.717135598 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:33 crc kubenswrapper[4821]: I1125 10:34:33.781948 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:33 crc kubenswrapper[4821]: E1125 10:34:33.783004 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:34.282946703 +0000 UTC m=+144.819266550 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:33 crc kubenswrapper[4821]: I1125 10:34:33.832866 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d8x5t" event={"ID":"b18917c3-98cb-407b-97e5-5333aa5ca797","Type":"ContainerStarted","Data":"d42628a7283863ae412c28951d6a4d12da9f559569d1a091bc386b7168aab459"} Nov 25 10:34:33 crc kubenswrapper[4821]: I1125 10:34:33.832929 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d8x5t" event={"ID":"b18917c3-98cb-407b-97e5-5333aa5ca797","Type":"ContainerStarted","Data":"6e322c7619c8a09ae256d79caec90543b2c4220486027879d089751315f3bf88"} Nov 25 10:34:33 crc kubenswrapper[4821]: I1125 10:34:33.849341 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-whcwc" event={"ID":"bdd88e7d-edf6-41c1-a5a3-47746bb3ab99","Type":"ContainerStarted","Data":"a67627cfb0e3116bef25404b270ac01cecbc65ff01961cdf2e660c4e6445724b"} Nov 25 10:34:33 crc kubenswrapper[4821]: I1125 10:34:33.852657 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-q8tq8" event={"ID":"5a7c1a05-b29c-4ff3-a4d6-90fe59eb1a01","Type":"ContainerStarted","Data":"4ad9e19219a5b5d4acc9a1dee2c7b5dd0eae577c6099fffecce1b01300a3805d"} Nov 25 10:34:33 crc kubenswrapper[4821]: I1125 10:34:33.855489 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-58htt" event={"ID":"518db853-3bbe-421b-a2dd-6d1b2e965120","Type":"ContainerStarted","Data":"717e41ba50a71a53a80e70255a9aed3a28af1f2d24eb9efc4567403bbd4385f4"} Nov 25 10:34:33 crc kubenswrapper[4821]: I1125 10:34:33.855743 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-58htt" Nov 25 10:34:33 crc kubenswrapper[4821]: I1125 10:34:33.857749 4821 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-58htt container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.9:8443/healthz\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Nov 25 10:34:33 crc kubenswrapper[4821]: I1125 10:34:33.857796 4821 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-58htt" podUID="518db853-3bbe-421b-a2dd-6d1b2e965120" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.9:8443/healthz\": dial tcp 10.217.0.9:8443: connect: connection refused" Nov 25 10:34:33 crc kubenswrapper[4821]: I1125 10:34:33.858393 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-tpbzb" event={"ID":"9f6c1e2f-b7e7-40e7-92c7-8ca1b57eb9f1","Type":"ContainerStarted","Data":"533f7b06995aa096676caf8eebbfde56cd6eab2450ccf1aebae2670b30e4812e"} Nov 25 10:34:33 crc kubenswrapper[4821]: I1125 10:34:33.861394 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k622z" event={"ID":"513166e2-0ac9-4299-bf84-d994a46da192","Type":"ContainerStarted","Data":"af9266f5e9efa2c8c07a882e75ea15f8395929a1ddd7afe2d953df1b7bd6fbf3"} Nov 25 10:34:33 crc kubenswrapper[4821]: I1125 10:34:33.862509 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kps6n" event={"ID":"eba7a4e4-d755-4368-a487-43f200d258b4","Type":"ContainerStarted","Data":"59adcb56bf29822c0d4d3576db50737223a34a125e875f39ae92f3e518c24247"} Nov 25 10:34:33 crc kubenswrapper[4821]: I1125 10:34:33.866796 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-prvdm" event={"ID":"db778f33-712d-42fc-849e-20e0af125a5a","Type":"ContainerStarted","Data":"0fc9912e86a78739ec95d887fe6e464f4a1cf7f9178b3d3caa7d291d9a51218a"} Nov 25 10:34:33 crc kubenswrapper[4821]: I1125 10:34:33.866847 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-prvdm" event={"ID":"db778f33-712d-42fc-849e-20e0af125a5a","Type":"ContainerStarted","Data":"2ac2400aa91050e7b6331171ab8ba2370aa04605a268613880341df13c27b9a4"} Nov 25 10:34:33 crc kubenswrapper[4821]: I1125 10:34:33.869784 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zvztq" event={"ID":"7fe78aa4-571c-488b-96bd-23ae4b6333db","Type":"ContainerStarted","Data":"c3d4d3f5faaf611b17a76e9867bf74cc63dd4e2613ada5490c52fba63ea8de8e"} Nov 25 10:34:33 crc kubenswrapper[4821]: I1125 10:34:33.872316 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-hddc7" event={"ID":"b0e34420-095d-42aa-89c4-5a80331e1c01","Type":"ContainerStarted","Data":"a66fab01b8d1502e0fa78f4235e983084275a0d16c2fdaf8635a8f31e4d3f743"} Nov 25 10:34:33 crc kubenswrapper[4821]: I1125 10:34:33.877647 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-p92sp" event={"ID":"66798226-d262-4925-82db-2742d96a33fc","Type":"ContainerStarted","Data":"77e723f00c823d028aea92559eae3701ea990dbc0b1721bff80494a3af67baa2"} Nov 25 10:34:33 crc kubenswrapper[4821]: I1125 10:34:33.877689 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-p92sp" event={"ID":"66798226-d262-4925-82db-2742d96a33fc","Type":"ContainerStarted","Data":"feb5afc3b77a9d7ac0a54efe1b9ad23cad970329cea0ec6f3fabf01d608fcae9"} Nov 25 10:34:33 crc kubenswrapper[4821]: I1125 10:34:33.877789 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-58htt" podStartSLOduration=123.87777298500001 podStartE2EDuration="2m3.877772985s" podCreationTimestamp="2025-11-25 10:32:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:34:33.876752029 +0000 UTC m=+144.413071876" watchObservedRunningTime="2025-11-25 10:34:33.877772985 +0000 UTC m=+144.414092832" Nov 25 10:34:33 crc kubenswrapper[4821]: I1125 10:34:33.880214 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-kc2wp" event={"ID":"c82133bf-deb1-4b25-b002-f435a64847cd","Type":"ContainerStarted","Data":"9d6ecd17821bd306419f1e09b3ba9c765df07769029fc459b7f37556bdce6d46"} Nov 25 10:34:33 crc kubenswrapper[4821]: I1125 10:34:33.883026 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:33 crc kubenswrapper[4821]: I1125 10:34:33.883246 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tkgh9" event={"ID":"2d3eed24-29f1-414a-a5d8-f8b021e6aff4","Type":"ContainerStarted","Data":"879aa6da9549a617b9bcf15691f04f4aec4a6dd694f32b2ad2e7fdf8240f4b7f"} Nov 25 10:34:33 crc kubenswrapper[4821]: I1125 10:34:33.883284 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tkgh9" event={"ID":"2d3eed24-29f1-414a-a5d8-f8b021e6aff4","Type":"ContainerStarted","Data":"c1aa2b8e206dc391db5d9455f3d3d716800d9f9f5f0b4220482ae7de70b39448"} Nov 25 10:34:33 crc kubenswrapper[4821]: E1125 10:34:33.883346 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:34.383331903 +0000 UTC m=+144.919651750 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:33 crc kubenswrapper[4821]: I1125 10:34:33.885038 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w5k28" event={"ID":"486e2f09-ddfc-4058-8749-bf747fc829ca","Type":"ContainerStarted","Data":"58276a53c4c2ccf14c891a1659e70aadae5cde48b8f68941641f027da9773e94"} Nov 25 10:34:33 crc kubenswrapper[4821]: I1125 10:34:33.886826 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s7j8m" event={"ID":"d1e46310-3643-4e6a-9624-467e4e00b093","Type":"ContainerStarted","Data":"e796091d08990271d80ed91078fb919cda6e524111c846c7a1f8b027c65fb35a"} Nov 25 10:34:33 crc kubenswrapper[4821]: I1125 10:34:33.890554 4821 generic.go:334] "Generic (PLEG): container finished" podID="064de93d-7bb5-47b5-9585-bf32be9a05f6" containerID="95bb8af3d192dbc0390a16eb571a0902f6617bf5c88c7863c9bdca3771d2c7a7" exitCode=0 Nov 25 10:34:33 crc kubenswrapper[4821]: I1125 10:34:33.890641 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-jb5sz" event={"ID":"064de93d-7bb5-47b5-9585-bf32be9a05f6","Type":"ContainerDied","Data":"95bb8af3d192dbc0390a16eb571a0902f6617bf5c88c7863c9bdca3771d2c7a7"} Nov 25 10:34:33 crc kubenswrapper[4821]: I1125 10:34:33.895418 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-prvdm" podStartSLOduration=123.895398061 podStartE2EDuration="2m3.895398061s" podCreationTimestamp="2025-11-25 10:32:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:34:33.892249343 +0000 UTC m=+144.428569190" watchObservedRunningTime="2025-11-25 10:34:33.895398061 +0000 UTC m=+144.431717908" Nov 25 10:34:33 crc kubenswrapper[4821]: I1125 10:34:33.908086 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-v6pjk" event={"ID":"db41ea16-6481-4ed9-a935-d14544e552a0","Type":"ContainerStarted","Data":"ac9ff53bf3ed708f11377bd8b6d787c33f21aaa156f9d11a128631bfb487c70d"} Nov 25 10:34:33 crc kubenswrapper[4821]: I1125 10:34:33.910052 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-tsb45" event={"ID":"f97c276b-2b69-49a0-96d1-b1c5171a542e","Type":"ContainerStarted","Data":"585e030c852c79bbf7113be89ecc7dac0f1e61f32b138c6b7d852b3b494d47de"} Nov 25 10:34:33 crc kubenswrapper[4821]: I1125 10:34:33.911532 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dv9nt" event={"ID":"ba27af47-1944-4970-a015-e96c7cee4862","Type":"ContainerStarted","Data":"afa3b5464872e57b7a8921ecdae57338715ab15ac3e01f3eb570907c98771328"} Nov 25 10:34:33 crc kubenswrapper[4821]: I1125 10:34:33.912544 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dv9nt" Nov 25 10:34:33 crc kubenswrapper[4821]: I1125 10:34:33.920726 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ng2nv" event={"ID":"5ed0f2b3-fd6d-4892-8198-2b81d5ea1a7f","Type":"ContainerStarted","Data":"954f98ffed45f1a30a23811ff47a5b05979d28dcf5464775a7451e42690706b8"} Nov 25 10:34:33 crc kubenswrapper[4821]: I1125 10:34:33.924042 4821 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-dv9nt container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.11:8443/healthz\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Nov 25 10:34:33 crc kubenswrapper[4821]: I1125 10:34:33.925225 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401110-dm86l" event={"ID":"a5033129-ba8c-488f-982e-78e12b2f1664","Type":"ContainerStarted","Data":"f647a6676a61dff0159b32ea289553f18ca7336aafb1fec95c26f25c594d6e9e"} Nov 25 10:34:33 crc kubenswrapper[4821]: I1125 10:34:33.927493 4821 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dv9nt" podUID="ba27af47-1944-4970-a015-e96c7cee4862" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.11:8443/healthz\": dial tcp 10.217.0.11:8443: connect: connection refused" Nov 25 10:34:33 crc kubenswrapper[4821]: I1125 10:34:33.928388 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zvztq" podStartSLOduration=123.928367949 podStartE2EDuration="2m3.928367949s" podCreationTimestamp="2025-11-25 10:32:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:34:33.915628183 +0000 UTC m=+144.451948050" watchObservedRunningTime="2025-11-25 10:34:33.928367949 +0000 UTC m=+144.464687796" Nov 25 10:34:33 crc kubenswrapper[4821]: I1125 10:34:33.952907 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6xxs8" event={"ID":"0e82d98b-7c99-496d-910d-e9bfae2ca6ec","Type":"ContainerStarted","Data":"09f7bb4dc3e537fa92b8d31ab3a456c7f600cec92df535806cbcf016bacce8b0"} Nov 25 10:34:33 crc kubenswrapper[4821]: I1125 10:34:33.959865 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-tpbzb" podStartSLOduration=123.959843479 podStartE2EDuration="2m3.959843479s" podCreationTimestamp="2025-11-25 10:32:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:34:33.952098708 +0000 UTC m=+144.488418575" watchObservedRunningTime="2025-11-25 10:34:33.959843479 +0000 UTC m=+144.496163326" Nov 25 10:34:33 crc kubenswrapper[4821]: I1125 10:34:33.962450 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-898jb" event={"ID":"9276b1a7-e1c8-4cdb-a77c-4b09df17be76","Type":"ContainerStarted","Data":"999165c9c56225a04ac3160e2aea0a64b6092302a34607c297d6720c4752dfee"} Nov 25 10:34:33 crc kubenswrapper[4821]: I1125 10:34:33.965535 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-kc2wp" podStartSLOduration=5.965519511 podStartE2EDuration="5.965519511s" podCreationTimestamp="2025-11-25 10:34:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:34:33.963966011 +0000 UTC m=+144.500285868" watchObservedRunningTime="2025-11-25 10:34:33.965519511 +0000 UTC m=+144.501839358" Nov 25 10:34:33 crc kubenswrapper[4821]: I1125 10:34:33.984835 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dv9nt" podStartSLOduration=123.984792068 podStartE2EDuration="2m3.984792068s" podCreationTimestamp="2025-11-25 10:32:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:34:33.97642801 +0000 UTC m=+144.512747867" watchObservedRunningTime="2025-11-25 10:34:33.984792068 +0000 UTC m=+144.521111915" Nov 25 10:34:33 crc kubenswrapper[4821]: I1125 10:34:33.985050 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:33 crc kubenswrapper[4821]: E1125 10:34:33.988688 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:34.488675744 +0000 UTC m=+145.024995591 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:33 crc kubenswrapper[4821]: I1125 10:34:33.996517 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-7792m" event={"ID":"499ccac5-6595-4d67-93ed-dd2d19d045bf","Type":"ContainerStarted","Data":"9240aa0b4b0b04e6be8b91e5db1cba51e647d271a9d2a657efb8d9021c22cb6b"} Nov 25 10:34:33 crc kubenswrapper[4821]: I1125 10:34:33.996609 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-7792m" event={"ID":"499ccac5-6595-4d67-93ed-dd2d19d045bf","Type":"ContainerStarted","Data":"a7f84213d8f358aebc938c840f63081a47c3b7ad32fea9bb437fb937e832b114"} Nov 25 10:34:33 crc kubenswrapper[4821]: I1125 10:34:33.998131 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6xxs8" podStartSLOduration=123.998117399 podStartE2EDuration="2m3.998117399s" podCreationTimestamp="2025-11-25 10:32:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:34:33.997962375 +0000 UTC m=+144.534282232" watchObservedRunningTime="2025-11-25 10:34:33.998117399 +0000 UTC m=+144.534437246" Nov 25 10:34:34 crc kubenswrapper[4821]: I1125 10:34:34.013473 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-dxf4x" event={"ID":"590f742c-e0d3-4394-8ef3-08bde5d57d01","Type":"ContainerStarted","Data":"41b4dac6291ebdf0ed0a316060e54df84fbc3b2fc75af2abd1e432330754933d"} Nov 25 10:34:34 crc kubenswrapper[4821]: I1125 10:34:34.013639 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-dxf4x" event={"ID":"590f742c-e0d3-4394-8ef3-08bde5d57d01","Type":"ContainerStarted","Data":"d5fa1194e5ea743ba24948633dced0251487682c43e560805055518f64d3cc40"} Nov 25 10:34:34 crc kubenswrapper[4821]: I1125 10:34:34.019524 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-rx6sh" event={"ID":"475ce242-f515-4f81-838c-5b0684222eb4","Type":"ContainerStarted","Data":"f0a3bd2cb34ef0a28ca0f5690868f963b020746c4e7d2e62cbb275849453953d"} Nov 25 10:34:34 crc kubenswrapper[4821]: I1125 10:34:34.028520 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-p2jmh" event={"ID":"402cf28b-25d2-428a-ab7b-1d70345a921e","Type":"ContainerStarted","Data":"913bc2268ea0ff640ac33f6f83a7e0fa29817048ca2398b36d65dc6a1ba23aa2"} Nov 25 10:34:34 crc kubenswrapper[4821]: I1125 10:34:34.028585 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-p2jmh" event={"ID":"402cf28b-25d2-428a-ab7b-1d70345a921e","Type":"ContainerStarted","Data":"b5fd211ca7b035e74f963dc08874d7bb60fee190e3d034d69fdd2e3a3a3731ae"} Nov 25 10:34:34 crc kubenswrapper[4821]: I1125 10:34:34.030409 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wxp2r" event={"ID":"8234c8fc-a19c-45c6-8d25-e65668a244ea","Type":"ContainerStarted","Data":"668c16d95141728c6db3e0dd93adf50b19b92717a5c783623fef9a459cd30f1a"} Nov 25 10:34:34 crc kubenswrapper[4821]: I1125 10:34:34.033963 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-f2tjh" event={"ID":"d7e79465-8df8-41a7-ab16-5c730f9ceca2","Type":"ContainerStarted","Data":"c720803f6c289472cca89c823f111bbc7f33958563332e7758aa5e85cd8da024"} Nov 25 10:34:34 crc kubenswrapper[4821]: I1125 10:34:34.036027 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-cx6h6" event={"ID":"3e66624b-9d0d-4440-9cd6-06ca77cb7566","Type":"ContainerStarted","Data":"96c0d50c7db62ed3381b04645bbaeb324a7bc0fa611ef51c56125b856365254b"} Nov 25 10:34:34 crc kubenswrapper[4821]: I1125 10:34:34.038948 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-v88lg" event={"ID":"155595fc-76b3-4561-bd3f-c540076d4cd6","Type":"ContainerStarted","Data":"2005f6aa99c8677fe8dfb9aace739f576eafbb9788de798e7704139890b0abb5"} Nov 25 10:34:34 crc kubenswrapper[4821]: I1125 10:34:34.039529 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-v88lg" Nov 25 10:34:34 crc kubenswrapper[4821]: I1125 10:34:34.051338 4821 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-v88lg container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.21:8443/healthz\": dial tcp 10.217.0.21:8443: connect: connection refused" start-of-body= Nov 25 10:34:34 crc kubenswrapper[4821]: I1125 10:34:34.051387 4821 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-v88lg" podUID="155595fc-76b3-4561-bd3f-c540076d4cd6" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.21:8443/healthz\": dial tcp 10.217.0.21:8443: connect: connection refused" Nov 25 10:34:34 crc kubenswrapper[4821]: I1125 10:34:34.058493 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-v88lg" podStartSLOduration=124.058472595 podStartE2EDuration="2m4.058472595s" podCreationTimestamp="2025-11-25 10:32:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:34:34.057495401 +0000 UTC m=+144.593815248" watchObservedRunningTime="2025-11-25 10:34:34.058472595 +0000 UTC m=+144.594792442" Nov 25 10:34:34 crc kubenswrapper[4821]: I1125 10:34:34.060491 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-86cnx" event={"ID":"59ec70c2-c208-4529-a763-a0d7ea7e1dfc","Type":"ContainerStarted","Data":"ed4cfcd85e0fb3adc6038dc503023c4c0c24dcff0877a0740e2e6441a6deedd7"} Nov 25 10:34:34 crc kubenswrapper[4821]: I1125 10:34:34.064268 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-zj4s2" event={"ID":"24708c13-efe2-435b-acde-b29b48871f16","Type":"ContainerStarted","Data":"7596a6cebbb866bc614c521249ed464f41ada96f87797f02355d6cb93c3ebe6b"} Nov 25 10:34:34 crc kubenswrapper[4821]: I1125 10:34:34.069430 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-n4lbr" event={"ID":"4878872d-8035-442a-a355-7c966f37e8c1","Type":"ContainerStarted","Data":"11a46417988df7e6569bf2090e0b04312296a566fbc62760b766a237942db17a"} Nov 25 10:34:34 crc kubenswrapper[4821]: I1125 10:34:34.071049 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-hjvt9" event={"ID":"f954533e-e8a0-4899-be72-241760feae95","Type":"ContainerStarted","Data":"85476df01cdd233aa043cdab8538c2fa562637811fc58da7df8de836e9316399"} Nov 25 10:34:34 crc kubenswrapper[4821]: I1125 10:34:34.072148 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-hjvt9" Nov 25 10:34:34 crc kubenswrapper[4821]: I1125 10:34:34.075288 4821 patch_prober.go:28] interesting pod/downloads-7954f5f757-hjvt9 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Nov 25 10:34:34 crc kubenswrapper[4821]: I1125 10:34:34.075333 4821 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-hjvt9" podUID="f954533e-e8a0-4899-be72-241760feae95" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Nov 25 10:34:34 crc kubenswrapper[4821]: I1125 10:34:34.077519 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2stv6" event={"ID":"4a74ce9e-b145-4e40-ad4e-67ecdf722195","Type":"ContainerStarted","Data":"dd50d0750c93f27197659a477a861c8dc4dd6e981d44dceca2e4cc45aefdce65"} Nov 25 10:34:34 crc kubenswrapper[4821]: I1125 10:34:34.077542 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2stv6" event={"ID":"4a74ce9e-b145-4e40-ad4e-67ecdf722195","Type":"ContainerStarted","Data":"3d5abd23747fb90e7e1e636a330508109a09e95c0b47833b71e225275a9ef40c"} Nov 25 10:34:34 crc kubenswrapper[4821]: I1125 10:34:34.078204 4821 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-mbn4k container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.5:6443/healthz\": dial tcp 10.217.0.5:6443: connect: connection refused" start-of-body= Nov 25 10:34:34 crc kubenswrapper[4821]: I1125 10:34:34.078249 4821 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-mbn4k" podUID="88feae6a-5031-48ed-93ec-34ef8d2eff40" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.5:6443/healthz\": dial tcp 10.217.0.5:6443: connect: connection refused" Nov 25 10:34:34 crc kubenswrapper[4821]: I1125 10:34:34.078540 4821 patch_prober.go:28] interesting pod/console-operator-58897d9998-t5hm7 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/readyz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Nov 25 10:34:34 crc kubenswrapper[4821]: I1125 10:34:34.078588 4821 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-t5hm7" podUID="b75a7773-a689-46c1-b98e-e7c1be0878e6" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.6:8443/readyz\": dial tcp 10.217.0.6:8443: connect: connection refused" Nov 25 10:34:34 crc kubenswrapper[4821]: I1125 10:34:34.078996 4821 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-4kjbr container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.29:8443/healthz\": dial tcp 10.217.0.29:8443: connect: connection refused" start-of-body= Nov 25 10:34:34 crc kubenswrapper[4821]: I1125 10:34:34.079059 4821 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4kjbr" podUID="2c2ecf12-49bd-4baf-90df-68486f27be1b" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.29:8443/healthz\": dial tcp 10.217.0.29:8443: connect: connection refused" Nov 25 10:34:34 crc kubenswrapper[4821]: I1125 10:34:34.078991 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-86cnx" podStartSLOduration=124.078970094 podStartE2EDuration="2m4.078970094s" podCreationTimestamp="2025-11-25 10:32:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:34:34.077141458 +0000 UTC m=+144.613461305" watchObservedRunningTime="2025-11-25 10:34:34.078970094 +0000 UTC m=+144.615289941" Nov 25 10:34:34 crc kubenswrapper[4821]: I1125 10:34:34.088965 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:34 crc kubenswrapper[4821]: E1125 10:34:34.090601 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:34.590344356 +0000 UTC m=+145.126664203 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:34 crc kubenswrapper[4821]: I1125 10:34:34.096451 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-hjvt9" podStartSLOduration=124.096405666 podStartE2EDuration="2m4.096405666s" podCreationTimestamp="2025-11-25 10:32:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:34:34.093096884 +0000 UTC m=+144.629416731" watchObservedRunningTime="2025-11-25 10:34:34.096405666 +0000 UTC m=+144.632725513" Nov 25 10:34:34 crc kubenswrapper[4821]: I1125 10:34:34.136944 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-hf49d" podStartSLOduration=124.13692171 podStartE2EDuration="2m4.13692171s" podCreationTimestamp="2025-11-25 10:32:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:34:34.132346137 +0000 UTC m=+144.668665984" watchObservedRunningTime="2025-11-25 10:34:34.13692171 +0000 UTC m=+144.673241557" Nov 25 10:34:34 crc kubenswrapper[4821]: I1125 10:34:34.140254 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2stv6" podStartSLOduration=124.137997807 podStartE2EDuration="2m4.137997807s" podCreationTimestamp="2025-11-25 10:32:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:34:34.110850883 +0000 UTC m=+144.647170731" watchObservedRunningTime="2025-11-25 10:34:34.137997807 +0000 UTC m=+144.674317654" Nov 25 10:34:34 crc kubenswrapper[4821]: I1125 10:34:34.191236 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:34 crc kubenswrapper[4821]: E1125 10:34:34.200859 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:34.700817635 +0000 UTC m=+145.237137482 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:34 crc kubenswrapper[4821]: I1125 10:34:34.292417 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:34 crc kubenswrapper[4821]: E1125 10:34:34.292885 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:34.792869927 +0000 UTC m=+145.329189774 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:34 crc kubenswrapper[4821]: I1125 10:34:34.394239 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:34 crc kubenswrapper[4821]: E1125 10:34:34.394574 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:34.894561549 +0000 UTC m=+145.430881396 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:34 crc kubenswrapper[4821]: I1125 10:34:34.495780 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:34 crc kubenswrapper[4821]: E1125 10:34:34.495921 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:34.995896031 +0000 UTC m=+145.532215878 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:34 crc kubenswrapper[4821]: I1125 10:34:34.496469 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:34 crc kubenswrapper[4821]: E1125 10:34:34.496750 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:34.996739812 +0000 UTC m=+145.533059659 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:34 crc kubenswrapper[4821]: I1125 10:34:34.598007 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:34 crc kubenswrapper[4821]: E1125 10:34:34.598211 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:35.098187697 +0000 UTC m=+145.634507544 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:34 crc kubenswrapper[4821]: I1125 10:34:34.598350 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:34 crc kubenswrapper[4821]: E1125 10:34:34.598682 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:35.09867396 +0000 UTC m=+145.634993797 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:34 crc kubenswrapper[4821]: I1125 10:34:34.699261 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:34 crc kubenswrapper[4821]: E1125 10:34:34.699555 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:35.199540831 +0000 UTC m=+145.735860678 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:34 crc kubenswrapper[4821]: I1125 10:34:34.717073 4821 patch_prober.go:28] interesting pod/machine-config-daemon-2krbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:34:34 crc kubenswrapper[4821]: I1125 10:34:34.717135 4821 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:34:34 crc kubenswrapper[4821]: I1125 10:34:34.800614 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:34 crc kubenswrapper[4821]: E1125 10:34:34.803771 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:35.303753414 +0000 UTC m=+145.840073261 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:34 crc kubenswrapper[4821]: I1125 10:34:34.901346 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:34 crc kubenswrapper[4821]: E1125 10:34:34.901542 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:35.401520909 +0000 UTC m=+145.937840766 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:34 crc kubenswrapper[4821]: I1125 10:34:34.901845 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:34 crc kubenswrapper[4821]: E1125 10:34:34.902246 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:35.402236997 +0000 UTC m=+145.938556844 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:35 crc kubenswrapper[4821]: I1125 10:34:35.003735 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:35 crc kubenswrapper[4821]: E1125 10:34:35.003942 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:35.503912927 +0000 UTC m=+146.040232774 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:35 crc kubenswrapper[4821]: I1125 10:34:35.004156 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:35 crc kubenswrapper[4821]: E1125 10:34:35.004525 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:35.504516653 +0000 UTC m=+146.040836500 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:35 crc kubenswrapper[4821]: I1125 10:34:35.085824 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-898jb" event={"ID":"9276b1a7-e1c8-4cdb-a77c-4b09df17be76","Type":"ContainerStarted","Data":"dc96cd50f99e41aa398f9898c73ff12c596a945a913d846ad3df81f4f289289d"} Nov 25 10:34:35 crc kubenswrapper[4821]: I1125 10:34:35.085874 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-898jb" event={"ID":"9276b1a7-e1c8-4cdb-a77c-4b09df17be76","Type":"ContainerStarted","Data":"dbc84fcd66e485cc996aebf9255c566c2f7a22d1fd112f44f9bf1c535c3514db"} Nov 25 10:34:35 crc kubenswrapper[4821]: I1125 10:34:35.088856 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-cx6h6" event={"ID":"3e66624b-9d0d-4440-9cd6-06ca77cb7566","Type":"ContainerStarted","Data":"599dbc6634cf0b53225ed7c9d86c1ac0c31418c88b15695fffd35121a7d59cbc"} Nov 25 10:34:35 crc kubenswrapper[4821]: I1125 10:34:35.102370 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-f2tjh" event={"ID":"d7e79465-8df8-41a7-ab16-5c730f9ceca2","Type":"ContainerStarted","Data":"d81c7a581518cc7558ca5efc33e0412d675dc3c6a31dbec62eae60485be537a7"} Nov 25 10:34:35 crc kubenswrapper[4821]: I1125 10:34:35.102454 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-f2tjh" Nov 25 10:34:35 crc kubenswrapper[4821]: I1125 10:34:35.104808 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:35 crc kubenswrapper[4821]: E1125 10:34:35.105046 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:35.605018984 +0000 UTC m=+146.141338841 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:35 crc kubenswrapper[4821]: I1125 10:34:35.105277 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:35 crc kubenswrapper[4821]: E1125 10:34:35.105658 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:35.60562528 +0000 UTC m=+146.141945127 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:35 crc kubenswrapper[4821]: I1125 10:34:35.106935 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s7j8m" event={"ID":"d1e46310-3643-4e6a-9624-467e4e00b093","Type":"ContainerStarted","Data":"5372ea19e2bbd728610e97f54755d4153a7f8ac6097a325f7d72d3e38944beb4"} Nov 25 10:34:35 crc kubenswrapper[4821]: I1125 10:34:35.106968 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s7j8m" event={"ID":"d1e46310-3643-4e6a-9624-467e4e00b093","Type":"ContainerStarted","Data":"4a4c3eaeb81410c440071d2ed5f459297189040ee0198af53b2069b7093317d1"} Nov 25 10:34:35 crc kubenswrapper[4821]: I1125 10:34:35.111078 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k622z" event={"ID":"513166e2-0ac9-4299-bf84-d994a46da192","Type":"ContainerStarted","Data":"b236c3f122d3782981edb66544d4f5a35015c6991c9c46f5a56a226c41991668"} Nov 25 10:34:35 crc kubenswrapper[4821]: I1125 10:34:35.111389 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-898jb" podStartSLOduration=125.111369582 podStartE2EDuration="2m5.111369582s" podCreationTimestamp="2025-11-25 10:32:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:34:35.10848713 +0000 UTC m=+145.644806967" watchObservedRunningTime="2025-11-25 10:34:35.111369582 +0000 UTC m=+145.647689429" Nov 25 10:34:35 crc kubenswrapper[4821]: I1125 10:34:35.113829 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-v6pjk" event={"ID":"db41ea16-6481-4ed9-a935-d14544e552a0","Type":"ContainerStarted","Data":"881bc68e689048e4ab74fc341d161828c4d01bf2b911de263a7cbcb846b5b214"} Nov 25 10:34:35 crc kubenswrapper[4821]: I1125 10:34:35.115953 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401110-dm86l" event={"ID":"a5033129-ba8c-488f-982e-78e12b2f1664","Type":"ContainerStarted","Data":"21e6691e92fd38f9fb69206ae744da84aa6fde06b42b6560d4a335470deec9e3"} Nov 25 10:34:35 crc kubenswrapper[4821]: I1125 10:34:35.118289 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-zj4s2" event={"ID":"24708c13-efe2-435b-acde-b29b48871f16","Type":"ContainerStarted","Data":"cc7e0ce80a1b20945df9b442908582b4e49bc102de45cbb86536cfda2738f5fc"} Nov 25 10:34:35 crc kubenswrapper[4821]: I1125 10:34:35.120708 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-whcwc" event={"ID":"bdd88e7d-edf6-41c1-a5a3-47746bb3ab99","Type":"ContainerStarted","Data":"e1facc26ea20a2653665848fbc58b5de62bd0356dbb997ec1257d533ae35743d"} Nov 25 10:34:35 crc kubenswrapper[4821]: I1125 10:34:35.124604 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-hddc7" event={"ID":"b0e34420-095d-42aa-89c4-5a80331e1c01","Type":"ContainerStarted","Data":"bc543666db1a5bb0d9eba974c8510da00986aa5209103f7424a9bd442c5b9c96"} Nov 25 10:34:35 crc kubenswrapper[4821]: I1125 10:34:35.126478 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wxp2r" event={"ID":"8234c8fc-a19c-45c6-8d25-e65668a244ea","Type":"ContainerStarted","Data":"45968acdb8ed0ea734c4d19a69eebad59572b5bcb2e76d523a002e5a60bb1d53"} Nov 25 10:34:35 crc kubenswrapper[4821]: I1125 10:34:35.127980 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w5k28" event={"ID":"486e2f09-ddfc-4058-8749-bf747fc829ca","Type":"ContainerStarted","Data":"ec1b2c81f6544b0db406003fd18a5d902cb25227ecb33726e67fcc8cd838dc3f"} Nov 25 10:34:35 crc kubenswrapper[4821]: I1125 10:34:35.128005 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w5k28" event={"ID":"486e2f09-ddfc-4058-8749-bf747fc829ca","Type":"ContainerStarted","Data":"9b5003077e2ada8bf4ab800db708d47eb62115f98bfeea0a88da807c69e59a42"} Nov 25 10:34:35 crc kubenswrapper[4821]: I1125 10:34:35.129950 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ng2nv" event={"ID":"5ed0f2b3-fd6d-4892-8198-2b81d5ea1a7f","Type":"ContainerStarted","Data":"4893a1c01257bec4d96561164c456e714288a9846ddcff580f9d31fea75101a4"} Nov 25 10:34:35 crc kubenswrapper[4821]: I1125 10:34:35.130107 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ng2nv" event={"ID":"5ed0f2b3-fd6d-4892-8198-2b81d5ea1a7f","Type":"ContainerStarted","Data":"16ca13cb4deb12a3c9ffb76900e015f8a215af09b5449c347a99ac4b808349c2"} Nov 25 10:34:35 crc kubenswrapper[4821]: I1125 10:34:35.132727 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-rx6sh" event={"ID":"475ce242-f515-4f81-838c-5b0684222eb4","Type":"ContainerStarted","Data":"1763cc3823f472ce806224bc40d0128b2762aace226eaf007f8314157c666dad"} Nov 25 10:34:35 crc kubenswrapper[4821]: I1125 10:34:35.132897 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-rx6sh" Nov 25 10:34:35 crc kubenswrapper[4821]: I1125 10:34:35.138264 4821 generic.go:334] "Generic (PLEG): container finished" podID="eba7a4e4-d755-4368-a487-43f200d258b4" containerID="efe3cd5df887afa055d0bab9afa661007b4a822d1d75b7b57225b0ea131ee215" exitCode=0 Nov 25 10:34:35 crc kubenswrapper[4821]: I1125 10:34:35.138363 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kps6n" event={"ID":"eba7a4e4-d755-4368-a487-43f200d258b4","Type":"ContainerDied","Data":"efe3cd5df887afa055d0bab9afa661007b4a822d1d75b7b57225b0ea131ee215"} Nov 25 10:34:35 crc kubenswrapper[4821]: I1125 10:34:35.140208 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-dxf4x" Nov 25 10:34:35 crc kubenswrapper[4821]: I1125 10:34:35.141196 4821 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-dv9nt container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.11:8443/healthz\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Nov 25 10:34:35 crc kubenswrapper[4821]: I1125 10:34:35.141243 4821 patch_prober.go:28] interesting pod/downloads-7954f5f757-hjvt9 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Nov 25 10:34:35 crc kubenswrapper[4821]: I1125 10:34:35.141245 4821 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dv9nt" podUID="ba27af47-1944-4970-a015-e96c7cee4862" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.11:8443/healthz\": dial tcp 10.217.0.11:8443: connect: connection refused" Nov 25 10:34:35 crc kubenswrapper[4821]: I1125 10:34:35.141271 4821 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-hjvt9" podUID="f954533e-e8a0-4899-be72-241760feae95" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Nov 25 10:34:35 crc kubenswrapper[4821]: I1125 10:34:35.141333 4821 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-v88lg container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.21:8443/healthz\": dial tcp 10.217.0.21:8443: connect: connection refused" start-of-body= Nov 25 10:34:35 crc kubenswrapper[4821]: I1125 10:34:35.141351 4821 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-v88lg" podUID="155595fc-76b3-4561-bd3f-c540076d4cd6" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.21:8443/healthz\": dial tcp 10.217.0.21:8443: connect: connection refused" Nov 25 10:34:35 crc kubenswrapper[4821]: I1125 10:34:35.141196 4821 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-58htt container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.9:8443/healthz\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Nov 25 10:34:35 crc kubenswrapper[4821]: I1125 10:34:35.141381 4821 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-58htt" podUID="518db853-3bbe-421b-a2dd-6d1b2e965120" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.9:8443/healthz\": dial tcp 10.217.0.9:8443: connect: connection refused" Nov 25 10:34:35 crc kubenswrapper[4821]: I1125 10:34:35.145854 4821 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-dxf4x container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.35:8080/healthz\": dial tcp 10.217.0.35:8080: connect: connection refused" start-of-body= Nov 25 10:34:35 crc kubenswrapper[4821]: I1125 10:34:35.146036 4821 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-dxf4x" podUID="590f742c-e0d3-4394-8ef3-08bde5d57d01" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.35:8080/healthz\": dial tcp 10.217.0.35:8080: connect: connection refused" Nov 25 10:34:35 crc kubenswrapper[4821]: I1125 10:34:35.174924 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s7j8m" podStartSLOduration=125.174893307 podStartE2EDuration="2m5.174893307s" podCreationTimestamp="2025-11-25 10:32:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:34:35.152390359 +0000 UTC m=+145.688710206" watchObservedRunningTime="2025-11-25 10:34:35.174893307 +0000 UTC m=+145.711213154" Nov 25 10:34:35 crc kubenswrapper[4821]: I1125 10:34:35.205677 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-cx6h6" podStartSLOduration=7.205658099 podStartE2EDuration="7.205658099s" podCreationTimestamp="2025-11-25 10:34:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:34:35.205432374 +0000 UTC m=+145.741752231" watchObservedRunningTime="2025-11-25 10:34:35.205658099 +0000 UTC m=+145.741977946" Nov 25 10:34:35 crc kubenswrapper[4821]: I1125 10:34:35.206296 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:35 crc kubenswrapper[4821]: I1125 10:34:35.207103 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-f2tjh" podStartSLOduration=125.207094045 podStartE2EDuration="2m5.207094045s" podCreationTimestamp="2025-11-25 10:32:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:34:35.178567398 +0000 UTC m=+145.714887285" watchObservedRunningTime="2025-11-25 10:34:35.207094045 +0000 UTC m=+145.743413892" Nov 25 10:34:35 crc kubenswrapper[4821]: E1125 10:34:35.208624 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:35.708601452 +0000 UTC m=+146.244921299 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:35 crc kubenswrapper[4821]: I1125 10:34:35.229849 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:35 crc kubenswrapper[4821]: I1125 10:34:35.232019 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-hddc7" podStartSLOduration=125.232002243 podStartE2EDuration="2m5.232002243s" podCreationTimestamp="2025-11-25 10:32:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:34:35.230418114 +0000 UTC m=+145.766737971" watchObservedRunningTime="2025-11-25 10:34:35.232002243 +0000 UTC m=+145.768322090" Nov 25 10:34:35 crc kubenswrapper[4821]: E1125 10:34:35.237425 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:35.737410357 +0000 UTC m=+146.273730204 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:35 crc kubenswrapper[4821]: I1125 10:34:35.281917 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-p2jmh" podStartSLOduration=125.28189561 podStartE2EDuration="2m5.28189561s" podCreationTimestamp="2025-11-25 10:32:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:34:35.277096891 +0000 UTC m=+145.813416738" watchObservedRunningTime="2025-11-25 10:34:35.28189561 +0000 UTC m=+145.818215457" Nov 25 10:34:35 crc kubenswrapper[4821]: I1125 10:34:35.324430 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ng2nv" podStartSLOduration=125.324387924 podStartE2EDuration="2m5.324387924s" podCreationTimestamp="2025-11-25 10:32:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:34:35.305107865 +0000 UTC m=+145.841427722" watchObservedRunningTime="2025-11-25 10:34:35.324387924 +0000 UTC m=+145.860707771" Nov 25 10:34:35 crc kubenswrapper[4821]: I1125 10:34:35.335402 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-dxf4x" podStartSLOduration=125.335369806 podStartE2EDuration="2m5.335369806s" podCreationTimestamp="2025-11-25 10:32:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:34:35.330087535 +0000 UTC m=+145.866407402" watchObservedRunningTime="2025-11-25 10:34:35.335369806 +0000 UTC m=+145.871689663" Nov 25 10:34:35 crc kubenswrapper[4821]: I1125 10:34:35.349300 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:35 crc kubenswrapper[4821]: E1125 10:34:35.350141 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:35.850123351 +0000 UTC m=+146.386443198 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:35 crc kubenswrapper[4821]: I1125 10:34:35.361314 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wxp2r" podStartSLOduration=125.361297538 podStartE2EDuration="2m5.361297538s" podCreationTimestamp="2025-11-25 10:32:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:34:35.359191077 +0000 UTC m=+145.895510934" watchObservedRunningTime="2025-11-25 10:34:35.361297538 +0000 UTC m=+145.897617385" Nov 25 10:34:35 crc kubenswrapper[4821]: I1125 10:34:35.395721 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-n4lbr" podStartSLOduration=125.395687732 podStartE2EDuration="2m5.395687732s" podCreationTimestamp="2025-11-25 10:32:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:34:35.393407235 +0000 UTC m=+145.929727082" watchObservedRunningTime="2025-11-25 10:34:35.395687732 +0000 UTC m=+145.932007569" Nov 25 10:34:35 crc kubenswrapper[4821]: I1125 10:34:35.439894 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k622z" podStartSLOduration=125.439872247 podStartE2EDuration="2m5.439872247s" podCreationTimestamp="2025-11-25 10:32:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:34:35.4242751 +0000 UTC m=+145.960594967" watchObservedRunningTime="2025-11-25 10:34:35.439872247 +0000 UTC m=+145.976192104" Nov 25 10:34:35 crc kubenswrapper[4821]: I1125 10:34:35.451493 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:35 crc kubenswrapper[4821]: E1125 10:34:35.452069 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:35.952055409 +0000 UTC m=+146.488375256 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:35 crc kubenswrapper[4821]: I1125 10:34:35.466550 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-rx6sh" podStartSLOduration=7.466530778 podStartE2EDuration="7.466530778s" podCreationTimestamp="2025-11-25 10:34:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:34:35.466516768 +0000 UTC m=+146.002836615" watchObservedRunningTime="2025-11-25 10:34:35.466530778 +0000 UTC m=+146.002850625" Nov 25 10:34:35 crc kubenswrapper[4821]: I1125 10:34:35.531214 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-zj4s2" podStartSLOduration=125.531194292 podStartE2EDuration="2m5.531194292s" podCreationTimestamp="2025-11-25 10:32:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:34:35.52668219 +0000 UTC m=+146.063002027" watchObservedRunningTime="2025-11-25 10:34:35.531194292 +0000 UTC m=+146.067514139" Nov 25 10:34:35 crc kubenswrapper[4821]: I1125 10:34:35.552235 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:35 crc kubenswrapper[4821]: E1125 10:34:35.552606 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:36.052591822 +0000 UTC m=+146.588911669 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:35 crc kubenswrapper[4821]: I1125 10:34:35.586504 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tkgh9" podStartSLOduration=125.586480582 podStartE2EDuration="2m5.586480582s" podCreationTimestamp="2025-11-25 10:32:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:34:35.583095718 +0000 UTC m=+146.119415575" watchObservedRunningTime="2025-11-25 10:34:35.586480582 +0000 UTC m=+146.122800429" Nov 25 10:34:35 crc kubenswrapper[4821]: I1125 10:34:35.620481 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-v6pjk" podStartSLOduration=125.620461844 podStartE2EDuration="2m5.620461844s" podCreationTimestamp="2025-11-25 10:32:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:34:35.61867714 +0000 UTC m=+146.154996987" watchObservedRunningTime="2025-11-25 10:34:35.620461844 +0000 UTC m=+146.156781691" Nov 25 10:34:35 crc kubenswrapper[4821]: I1125 10:34:35.654040 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:35 crc kubenswrapper[4821]: E1125 10:34:35.654568 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:36.15455242 +0000 UTC m=+146.690872257 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:35 crc kubenswrapper[4821]: I1125 10:34:35.671546 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-whcwc" podStartSLOduration=125.671527611 podStartE2EDuration="2m5.671527611s" podCreationTimestamp="2025-11-25 10:32:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:34:35.66867044 +0000 UTC m=+146.204990307" watchObservedRunningTime="2025-11-25 10:34:35.671527611 +0000 UTC m=+146.207847458" Nov 25 10:34:35 crc kubenswrapper[4821]: I1125 10:34:35.704071 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29401110-dm86l" podStartSLOduration=125.704052147 podStartE2EDuration="2m5.704052147s" podCreationTimestamp="2025-11-25 10:32:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:34:35.702737005 +0000 UTC m=+146.239056862" watchObservedRunningTime="2025-11-25 10:34:35.704052147 +0000 UTC m=+146.240371994" Nov 25 10:34:35 crc kubenswrapper[4821]: I1125 10:34:35.723554 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-q8tq8" podStartSLOduration=125.72352973 podStartE2EDuration="2m5.72352973s" podCreationTimestamp="2025-11-25 10:32:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:34:35.719709695 +0000 UTC m=+146.256029542" watchObservedRunningTime="2025-11-25 10:34:35.72352973 +0000 UTC m=+146.259849577" Nov 25 10:34:35 crc kubenswrapper[4821]: I1125 10:34:35.735331 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d8x5t" podStartSLOduration=125.735310663 podStartE2EDuration="2m5.735310663s" podCreationTimestamp="2025-11-25 10:32:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:34:35.734288707 +0000 UTC m=+146.270608564" watchObservedRunningTime="2025-11-25 10:34:35.735310663 +0000 UTC m=+146.271630510" Nov 25 10:34:35 crc kubenswrapper[4821]: I1125 10:34:35.750615 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-7792m" podStartSLOduration=125.750598951 podStartE2EDuration="2m5.750598951s" podCreationTimestamp="2025-11-25 10:32:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:34:35.747345821 +0000 UTC m=+146.283665688" watchObservedRunningTime="2025-11-25 10:34:35.750598951 +0000 UTC m=+146.286918798" Nov 25 10:34:35 crc kubenswrapper[4821]: I1125 10:34:35.755582 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:35 crc kubenswrapper[4821]: E1125 10:34:35.755775 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:36.255748059 +0000 UTC m=+146.792067906 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:35 crc kubenswrapper[4821]: I1125 10:34:35.755843 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:35 crc kubenswrapper[4821]: E1125 10:34:35.756279 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:36.256266242 +0000 UTC m=+146.792586089 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:35 crc kubenswrapper[4821]: I1125 10:34:35.767336 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w5k28" podStartSLOduration=125.767319076 podStartE2EDuration="2m5.767319076s" podCreationTimestamp="2025-11-25 10:32:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:34:35.764944607 +0000 UTC m=+146.301264454" watchObservedRunningTime="2025-11-25 10:34:35.767319076 +0000 UTC m=+146.303638923" Nov 25 10:34:35 crc kubenswrapper[4821]: I1125 10:34:35.803073 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-p92sp" podStartSLOduration=125.803057982 podStartE2EDuration="2m5.803057982s" podCreationTimestamp="2025-11-25 10:32:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:34:35.802088108 +0000 UTC m=+146.338407955" watchObservedRunningTime="2025-11-25 10:34:35.803057982 +0000 UTC m=+146.339377829" Nov 25 10:34:35 crc kubenswrapper[4821]: I1125 10:34:35.856683 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:35 crc kubenswrapper[4821]: E1125 10:34:35.856876 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:36.356847556 +0000 UTC m=+146.893167413 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:35 crc kubenswrapper[4821]: I1125 10:34:35.857298 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:35 crc kubenswrapper[4821]: E1125 10:34:35.857833 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:36.35781962 +0000 UTC m=+146.894139527 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:35 crc kubenswrapper[4821]: I1125 10:34:35.958893 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:35 crc kubenswrapper[4821]: E1125 10:34:35.959100 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:36.45907184 +0000 UTC m=+146.995391697 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:35 crc kubenswrapper[4821]: I1125 10:34:35.959434 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:35 crc kubenswrapper[4821]: E1125 10:34:35.959768 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:36.459755578 +0000 UTC m=+146.996075425 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:35 crc kubenswrapper[4821]: I1125 10:34:35.981754 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-p92sp" Nov 25 10:34:36 crc kubenswrapper[4821]: I1125 10:34:36.008654 4821 patch_prober.go:28] interesting pod/router-default-5444994796-p92sp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 10:34:36 crc kubenswrapper[4821]: [-]has-synced failed: reason withheld Nov 25 10:34:36 crc kubenswrapper[4821]: [+]process-running ok Nov 25 10:34:36 crc kubenswrapper[4821]: healthz check failed Nov 25 10:34:36 crc kubenswrapper[4821]: I1125 10:34:36.008737 4821 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p92sp" podUID="66798226-d262-4925-82db-2742d96a33fc" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 10:34:36 crc kubenswrapper[4821]: I1125 10:34:36.060331 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:36 crc kubenswrapper[4821]: E1125 10:34:36.060468 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:36.560436624 +0000 UTC m=+147.096756471 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:36 crc kubenswrapper[4821]: I1125 10:34:36.060848 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:36 crc kubenswrapper[4821]: E1125 10:34:36.061226 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:36.561214993 +0000 UTC m=+147.097534910 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:36 crc kubenswrapper[4821]: I1125 10:34:36.113850 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-q8tq8" Nov 25 10:34:36 crc kubenswrapper[4821]: I1125 10:34:36.116326 4821 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-q8tq8 container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Nov 25 10:34:36 crc kubenswrapper[4821]: I1125 10:34:36.116384 4821 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-q8tq8" podUID="5a7c1a05-b29c-4ff3-a4d6-90fe59eb1a01" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" Nov 25 10:34:36 crc kubenswrapper[4821]: I1125 10:34:36.116518 4821 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-q8tq8 container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Nov 25 10:34:36 crc kubenswrapper[4821]: I1125 10:34:36.116607 4821 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-q8tq8" podUID="5a7c1a05-b29c-4ff3-a4d6-90fe59eb1a01" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" Nov 25 10:34:36 crc kubenswrapper[4821]: I1125 10:34:36.147251 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-rx6sh" event={"ID":"475ce242-f515-4f81-838c-5b0684222eb4","Type":"ContainerStarted","Data":"523d047c16806632be422e3fe3d6a3a1fab2c2f68d33610b5ee76d7713892c15"} Nov 25 10:34:36 crc kubenswrapper[4821]: I1125 10:34:36.149278 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kps6n" event={"ID":"eba7a4e4-d755-4368-a487-43f200d258b4","Type":"ContainerStarted","Data":"2052ef79e7ea59fa1cc79bf2add4d30a3d2b271750d6d0dca26cb7d0386b62af"} Nov 25 10:34:36 crc kubenswrapper[4821]: I1125 10:34:36.150812 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-jb5sz" event={"ID":"064de93d-7bb5-47b5-9585-bf32be9a05f6","Type":"ContainerStarted","Data":"75306a78e96c2c90ee48b97dc3e12e4ea48cab7981b2ba713f0419b55f395f1c"} Nov 25 10:34:36 crc kubenswrapper[4821]: I1125 10:34:36.151247 4821 patch_prober.go:28] interesting pod/downloads-7954f5f757-hjvt9 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Nov 25 10:34:36 crc kubenswrapper[4821]: I1125 10:34:36.151272 4821 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-dxf4x container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.35:8080/healthz\": dial tcp 10.217.0.35:8080: connect: connection refused" start-of-body= Nov 25 10:34:36 crc kubenswrapper[4821]: I1125 10:34:36.151293 4821 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-hjvt9" podUID="f954533e-e8a0-4899-be72-241760feae95" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Nov 25 10:34:36 crc kubenswrapper[4821]: I1125 10:34:36.151308 4821 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-dxf4x" podUID="590f742c-e0d3-4394-8ef3-08bde5d57d01" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.35:8080/healthz\": dial tcp 10.217.0.35:8080: connect: connection refused" Nov 25 10:34:36 crc kubenswrapper[4821]: I1125 10:34:36.152081 4821 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-q8tq8 container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Nov 25 10:34:36 crc kubenswrapper[4821]: I1125 10:34:36.152197 4821 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-q8tq8" podUID="5a7c1a05-b29c-4ff3-a4d6-90fe59eb1a01" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" Nov 25 10:34:36 crc kubenswrapper[4821]: I1125 10:34:36.156588 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-v88lg" Nov 25 10:34:36 crc kubenswrapper[4821]: I1125 10:34:36.162474 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:36 crc kubenswrapper[4821]: E1125 10:34:36.162579 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:36.662560036 +0000 UTC m=+147.198879883 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:36 crc kubenswrapper[4821]: I1125 10:34:36.162777 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:36 crc kubenswrapper[4821]: E1125 10:34:36.163317 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:36.663289434 +0000 UTC m=+147.199609281 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:36 crc kubenswrapper[4821]: I1125 10:34:36.264314 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:36 crc kubenswrapper[4821]: E1125 10:34:36.264634 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:36.764601257 +0000 UTC m=+147.300921104 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:36 crc kubenswrapper[4821]: I1125 10:34:36.265293 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:36 crc kubenswrapper[4821]: E1125 10:34:36.267988 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:36.76797214 +0000 UTC m=+147.304291987 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:36 crc kubenswrapper[4821]: I1125 10:34:36.367477 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:36 crc kubenswrapper[4821]: E1125 10:34:36.367650 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:36.867619731 +0000 UTC m=+147.403939588 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:36 crc kubenswrapper[4821]: I1125 10:34:36.368006 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:36 crc kubenswrapper[4821]: E1125 10:34:36.368406 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:36.86839102 +0000 UTC m=+147.404710867 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:36 crc kubenswrapper[4821]: I1125 10:34:36.469582 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:36 crc kubenswrapper[4821]: E1125 10:34:36.469766 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:36.969740173 +0000 UTC m=+147.506060020 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:36 crc kubenswrapper[4821]: I1125 10:34:36.470316 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:36 crc kubenswrapper[4821]: E1125 10:34:36.470701 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:36.970690057 +0000 UTC m=+147.507009904 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:36 crc kubenswrapper[4821]: I1125 10:34:36.572202 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:36 crc kubenswrapper[4821]: E1125 10:34:36.572380 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:37.072352937 +0000 UTC m=+147.608672784 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:36 crc kubenswrapper[4821]: I1125 10:34:36.572799 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:36 crc kubenswrapper[4821]: E1125 10:34:36.573232 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:37.073215638 +0000 UTC m=+147.609535485 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:36 crc kubenswrapper[4821]: I1125 10:34:36.651079 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dv9nt" Nov 25 10:34:36 crc kubenswrapper[4821]: I1125 10:34:36.673534 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:36 crc kubenswrapper[4821]: E1125 10:34:36.674281 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:37.174263144 +0000 UTC m=+147.710583001 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:36 crc kubenswrapper[4821]: I1125 10:34:36.776383 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:36 crc kubenswrapper[4821]: E1125 10:34:36.776805 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:37.276787046 +0000 UTC m=+147.813106893 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:36 crc kubenswrapper[4821]: I1125 10:34:36.879361 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:36 crc kubenswrapper[4821]: E1125 10:34:36.879562 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:37.379533774 +0000 UTC m=+147.915853631 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:36 crc kubenswrapper[4821]: I1125 10:34:36.880084 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:36 crc kubenswrapper[4821]: E1125 10:34:36.880461 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:37.380448626 +0000 UTC m=+147.916768533 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:36 crc kubenswrapper[4821]: I1125 10:34:36.981801 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:36 crc kubenswrapper[4821]: E1125 10:34:36.981983 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:37.481958503 +0000 UTC m=+148.018278350 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:36 crc kubenswrapper[4821]: I1125 10:34:36.982119 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:36 crc kubenswrapper[4821]: E1125 10:34:36.982479 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:37.482463026 +0000 UTC m=+148.018782923 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:36 crc kubenswrapper[4821]: I1125 10:34:36.990560 4821 patch_prober.go:28] interesting pod/router-default-5444994796-p92sp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 10:34:36 crc kubenswrapper[4821]: [-]has-synced failed: reason withheld Nov 25 10:34:36 crc kubenswrapper[4821]: [+]process-running ok Nov 25 10:34:36 crc kubenswrapper[4821]: healthz check failed Nov 25 10:34:36 crc kubenswrapper[4821]: I1125 10:34:36.990629 4821 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p92sp" podUID="66798226-d262-4925-82db-2742d96a33fc" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 10:34:37 crc kubenswrapper[4821]: I1125 10:34:37.082953 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:37 crc kubenswrapper[4821]: E1125 10:34:37.083141 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:37.583115041 +0000 UTC m=+148.119434888 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:37 crc kubenswrapper[4821]: I1125 10:34:37.083622 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:37 crc kubenswrapper[4821]: I1125 10:34:37.083864 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:34:37 crc kubenswrapper[4821]: E1125 10:34:37.083957 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:37.583947682 +0000 UTC m=+148.120267529 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:37 crc kubenswrapper[4821]: I1125 10:34:37.084147 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:34:37 crc kubenswrapper[4821]: I1125 10:34:37.085013 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:34:37 crc kubenswrapper[4821]: I1125 10:34:37.102134 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:34:37 crc kubenswrapper[4821]: I1125 10:34:37.156217 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-tsb45" event={"ID":"f97c276b-2b69-49a0-96d1-b1c5171a542e","Type":"ContainerStarted","Data":"0211cc7ef9453919289f643f08cb8e840c4aae2b54889bf7d9349227e5e7e5aa"} Nov 25 10:34:37 crc kubenswrapper[4821]: I1125 10:34:37.158567 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-jb5sz" event={"ID":"064de93d-7bb5-47b5-9585-bf32be9a05f6","Type":"ContainerStarted","Data":"993fea45f41fd1acc5b92695d9148120619f7093e6154eb03a9779549cef8967"} Nov 25 10:34:37 crc kubenswrapper[4821]: I1125 10:34:37.185151 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:37 crc kubenswrapper[4821]: I1125 10:34:37.185388 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:34:37 crc kubenswrapper[4821]: E1125 10:34:37.185440 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:37.685410508 +0000 UTC m=+148.221730395 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:37 crc kubenswrapper[4821]: I1125 10:34:37.185601 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:34:37 crc kubenswrapper[4821]: I1125 10:34:37.190752 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:34:37 crc kubenswrapper[4821]: I1125 10:34:37.191100 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:34:37 crc kubenswrapper[4821]: I1125 10:34:37.203851 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kps6n" podStartSLOduration=127.203834044 podStartE2EDuration="2m7.203834044s" podCreationTimestamp="2025-11-25 10:32:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:34:37.179765718 +0000 UTC m=+147.716085565" watchObservedRunningTime="2025-11-25 10:34:37.203834044 +0000 UTC m=+147.740153891" Nov 25 10:34:37 crc kubenswrapper[4821]: I1125 10:34:37.204771 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-jb5sz" podStartSLOduration=127.204765078 podStartE2EDuration="2m7.204765078s" podCreationTimestamp="2025-11-25 10:32:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:34:37.202568663 +0000 UTC m=+147.738888510" watchObservedRunningTime="2025-11-25 10:34:37.204765078 +0000 UTC m=+147.741084925" Nov 25 10:34:37 crc kubenswrapper[4821]: I1125 10:34:37.287642 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:37 crc kubenswrapper[4821]: E1125 10:34:37.288034 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:37.788018422 +0000 UTC m=+148.324338349 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:37 crc kubenswrapper[4821]: I1125 10:34:37.304870 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zrl8b"] Nov 25 10:34:37 crc kubenswrapper[4821]: I1125 10:34:37.306145 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zrl8b" Nov 25 10:34:37 crc kubenswrapper[4821]: I1125 10:34:37.311621 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 25 10:34:37 crc kubenswrapper[4821]: I1125 10:34:37.319560 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zrl8b"] Nov 25 10:34:37 crc kubenswrapper[4821]: I1125 10:34:37.326340 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 10:34:37 crc kubenswrapper[4821]: I1125 10:34:37.333108 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:34:37 crc kubenswrapper[4821]: I1125 10:34:37.340596 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 10:34:37 crc kubenswrapper[4821]: I1125 10:34:37.389480 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:37 crc kubenswrapper[4821]: E1125 10:34:37.389936 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:37.889916828 +0000 UTC m=+148.426236675 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:37 crc kubenswrapper[4821]: I1125 10:34:37.487107 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-ps9cl"] Nov 25 10:34:37 crc kubenswrapper[4821]: I1125 10:34:37.488746 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ps9cl" Nov 25 10:34:37 crc kubenswrapper[4821]: I1125 10:34:37.491616 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1d6804a-27b6-4ef8-90ce-32606c93869a-utilities\") pod \"community-operators-zrl8b\" (UID: \"d1d6804a-27b6-4ef8-90ce-32606c93869a\") " pod="openshift-marketplace/community-operators-zrl8b" Nov 25 10:34:37 crc kubenswrapper[4821]: I1125 10:34:37.491673 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fswpt\" (UniqueName: \"kubernetes.io/projected/d1d6804a-27b6-4ef8-90ce-32606c93869a-kube-api-access-fswpt\") pod \"community-operators-zrl8b\" (UID: \"d1d6804a-27b6-4ef8-90ce-32606c93869a\") " pod="openshift-marketplace/community-operators-zrl8b" Nov 25 10:34:37 crc kubenswrapper[4821]: I1125 10:34:37.491753 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1d6804a-27b6-4ef8-90ce-32606c93869a-catalog-content\") pod \"community-operators-zrl8b\" (UID: \"d1d6804a-27b6-4ef8-90ce-32606c93869a\") " pod="openshift-marketplace/community-operators-zrl8b" Nov 25 10:34:37 crc kubenswrapper[4821]: I1125 10:34:37.491787 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:37 crc kubenswrapper[4821]: E1125 10:34:37.492105 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:37.992091172 +0000 UTC m=+148.528411029 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:37 crc kubenswrapper[4821]: I1125 10:34:37.494550 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 25 10:34:37 crc kubenswrapper[4821]: I1125 10:34:37.499846 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ps9cl"] Nov 25 10:34:37 crc kubenswrapper[4821]: I1125 10:34:37.592912 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:37 crc kubenswrapper[4821]: I1125 10:34:37.593085 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1d6804a-27b6-4ef8-90ce-32606c93869a-utilities\") pod \"community-operators-zrl8b\" (UID: \"d1d6804a-27b6-4ef8-90ce-32606c93869a\") " pod="openshift-marketplace/community-operators-zrl8b" Nov 25 10:34:37 crc kubenswrapper[4821]: I1125 10:34:37.593121 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fswpt\" (UniqueName: \"kubernetes.io/projected/d1d6804a-27b6-4ef8-90ce-32606c93869a-kube-api-access-fswpt\") pod \"community-operators-zrl8b\" (UID: \"d1d6804a-27b6-4ef8-90ce-32606c93869a\") " pod="openshift-marketplace/community-operators-zrl8b" Nov 25 10:34:37 crc kubenswrapper[4821]: I1125 10:34:37.593146 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xn8qr\" (UniqueName: \"kubernetes.io/projected/645993fd-bdce-435e-a600-34bf1ff5b38b-kube-api-access-xn8qr\") pod \"certified-operators-ps9cl\" (UID: \"645993fd-bdce-435e-a600-34bf1ff5b38b\") " pod="openshift-marketplace/certified-operators-ps9cl" Nov 25 10:34:37 crc kubenswrapper[4821]: I1125 10:34:37.593208 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/645993fd-bdce-435e-a600-34bf1ff5b38b-utilities\") pod \"certified-operators-ps9cl\" (UID: \"645993fd-bdce-435e-a600-34bf1ff5b38b\") " pod="openshift-marketplace/certified-operators-ps9cl" Nov 25 10:34:37 crc kubenswrapper[4821]: I1125 10:34:37.593228 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1d6804a-27b6-4ef8-90ce-32606c93869a-catalog-content\") pod \"community-operators-zrl8b\" (UID: \"d1d6804a-27b6-4ef8-90ce-32606c93869a\") " pod="openshift-marketplace/community-operators-zrl8b" Nov 25 10:34:37 crc kubenswrapper[4821]: I1125 10:34:37.593258 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/645993fd-bdce-435e-a600-34bf1ff5b38b-catalog-content\") pod \"certified-operators-ps9cl\" (UID: \"645993fd-bdce-435e-a600-34bf1ff5b38b\") " pod="openshift-marketplace/certified-operators-ps9cl" Nov 25 10:34:37 crc kubenswrapper[4821]: E1125 10:34:37.593351 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:38.093334333 +0000 UTC m=+148.629654180 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:37 crc kubenswrapper[4821]: I1125 10:34:37.656011 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fswpt\" (UniqueName: \"kubernetes.io/projected/d1d6804a-27b6-4ef8-90ce-32606c93869a-kube-api-access-fswpt\") pod \"community-operators-zrl8b\" (UID: \"d1d6804a-27b6-4ef8-90ce-32606c93869a\") " pod="openshift-marketplace/community-operators-zrl8b" Nov 25 10:34:37 crc kubenswrapper[4821]: I1125 10:34:37.688588 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1d6804a-27b6-4ef8-90ce-32606c93869a-catalog-content\") pod \"community-operators-zrl8b\" (UID: \"d1d6804a-27b6-4ef8-90ce-32606c93869a\") " pod="openshift-marketplace/community-operators-zrl8b" Nov 25 10:34:37 crc kubenswrapper[4821]: I1125 10:34:37.694754 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/645993fd-bdce-435e-a600-34bf1ff5b38b-utilities\") pod \"certified-operators-ps9cl\" (UID: \"645993fd-bdce-435e-a600-34bf1ff5b38b\") " pod="openshift-marketplace/certified-operators-ps9cl" Nov 25 10:34:37 crc kubenswrapper[4821]: I1125 10:34:37.694846 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:37 crc kubenswrapper[4821]: I1125 10:34:37.694887 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/645993fd-bdce-435e-a600-34bf1ff5b38b-catalog-content\") pod \"certified-operators-ps9cl\" (UID: \"645993fd-bdce-435e-a600-34bf1ff5b38b\") " pod="openshift-marketplace/certified-operators-ps9cl" Nov 25 10:34:37 crc kubenswrapper[4821]: I1125 10:34:37.694977 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xn8qr\" (UniqueName: \"kubernetes.io/projected/645993fd-bdce-435e-a600-34bf1ff5b38b-kube-api-access-xn8qr\") pod \"certified-operators-ps9cl\" (UID: \"645993fd-bdce-435e-a600-34bf1ff5b38b\") " pod="openshift-marketplace/certified-operators-ps9cl" Nov 25 10:34:37 crc kubenswrapper[4821]: E1125 10:34:37.712641 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:38.21261639 +0000 UTC m=+148.748936237 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:37 crc kubenswrapper[4821]: I1125 10:34:37.712913 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-cgdxc"] Nov 25 10:34:37 crc kubenswrapper[4821]: I1125 10:34:37.721313 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cgdxc"] Nov 25 10:34:37 crc kubenswrapper[4821]: I1125 10:34:37.723237 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cgdxc" Nov 25 10:34:37 crc kubenswrapper[4821]: I1125 10:34:37.742422 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/645993fd-bdce-435e-a600-34bf1ff5b38b-utilities\") pod \"certified-operators-ps9cl\" (UID: \"645993fd-bdce-435e-a600-34bf1ff5b38b\") " pod="openshift-marketplace/certified-operators-ps9cl" Nov 25 10:34:37 crc kubenswrapper[4821]: I1125 10:34:37.742622 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1d6804a-27b6-4ef8-90ce-32606c93869a-utilities\") pod \"community-operators-zrl8b\" (UID: \"d1d6804a-27b6-4ef8-90ce-32606c93869a\") " pod="openshift-marketplace/community-operators-zrl8b" Nov 25 10:34:37 crc kubenswrapper[4821]: I1125 10:34:37.742685 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/645993fd-bdce-435e-a600-34bf1ff5b38b-catalog-content\") pod \"certified-operators-ps9cl\" (UID: \"645993fd-bdce-435e-a600-34bf1ff5b38b\") " pod="openshift-marketplace/certified-operators-ps9cl" Nov 25 10:34:37 crc kubenswrapper[4821]: I1125 10:34:37.748739 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xn8qr\" (UniqueName: \"kubernetes.io/projected/645993fd-bdce-435e-a600-34bf1ff5b38b-kube-api-access-xn8qr\") pod \"certified-operators-ps9cl\" (UID: \"645993fd-bdce-435e-a600-34bf1ff5b38b\") " pod="openshift-marketplace/certified-operators-ps9cl" Nov 25 10:34:37 crc kubenswrapper[4821]: W1125 10:34:37.754962 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-a549834d84cc714c9be0741f4c2daadab34fd572e0d3ad653eea91e610227a8c WatchSource:0}: Error finding container a549834d84cc714c9be0741f4c2daadab34fd572e0d3ad653eea91e610227a8c: Status 404 returned error can't find the container with id a549834d84cc714c9be0741f4c2daadab34fd572e0d3ad653eea91e610227a8c Nov 25 10:34:37 crc kubenswrapper[4821]: I1125 10:34:37.796092 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:37 crc kubenswrapper[4821]: I1125 10:34:37.796338 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njv9h\" (UniqueName: \"kubernetes.io/projected/b6e72ee2-50e4-4643-aee6-8be89d2e4589-kube-api-access-njv9h\") pod \"community-operators-cgdxc\" (UID: \"b6e72ee2-50e4-4643-aee6-8be89d2e4589\") " pod="openshift-marketplace/community-operators-cgdxc" Nov 25 10:34:37 crc kubenswrapper[4821]: I1125 10:34:37.796362 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6e72ee2-50e4-4643-aee6-8be89d2e4589-catalog-content\") pod \"community-operators-cgdxc\" (UID: \"b6e72ee2-50e4-4643-aee6-8be89d2e4589\") " pod="openshift-marketplace/community-operators-cgdxc" Nov 25 10:34:37 crc kubenswrapper[4821]: I1125 10:34:37.796382 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6e72ee2-50e4-4643-aee6-8be89d2e4589-utilities\") pod \"community-operators-cgdxc\" (UID: \"b6e72ee2-50e4-4643-aee6-8be89d2e4589\") " pod="openshift-marketplace/community-operators-cgdxc" Nov 25 10:34:37 crc kubenswrapper[4821]: E1125 10:34:37.796530 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:38.296515211 +0000 UTC m=+148.832835058 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:37 crc kubenswrapper[4821]: I1125 10:34:37.898914 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njv9h\" (UniqueName: \"kubernetes.io/projected/b6e72ee2-50e4-4643-aee6-8be89d2e4589-kube-api-access-njv9h\") pod \"community-operators-cgdxc\" (UID: \"b6e72ee2-50e4-4643-aee6-8be89d2e4589\") " pod="openshift-marketplace/community-operators-cgdxc" Nov 25 10:34:37 crc kubenswrapper[4821]: I1125 10:34:37.898946 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6e72ee2-50e4-4643-aee6-8be89d2e4589-catalog-content\") pod \"community-operators-cgdxc\" (UID: \"b6e72ee2-50e4-4643-aee6-8be89d2e4589\") " pod="openshift-marketplace/community-operators-cgdxc" Nov 25 10:34:37 crc kubenswrapper[4821]: I1125 10:34:37.898964 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6e72ee2-50e4-4643-aee6-8be89d2e4589-utilities\") pod \"community-operators-cgdxc\" (UID: \"b6e72ee2-50e4-4643-aee6-8be89d2e4589\") " pod="openshift-marketplace/community-operators-cgdxc" Nov 25 10:34:37 crc kubenswrapper[4821]: I1125 10:34:37.899005 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:37 crc kubenswrapper[4821]: E1125 10:34:37.899352 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:38.39934079 +0000 UTC m=+148.935660637 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:37 crc kubenswrapper[4821]: I1125 10:34:37.900013 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6e72ee2-50e4-4643-aee6-8be89d2e4589-catalog-content\") pod \"community-operators-cgdxc\" (UID: \"b6e72ee2-50e4-4643-aee6-8be89d2e4589\") " pod="openshift-marketplace/community-operators-cgdxc" Nov 25 10:34:37 crc kubenswrapper[4821]: I1125 10:34:37.900208 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6e72ee2-50e4-4643-aee6-8be89d2e4589-utilities\") pod \"community-operators-cgdxc\" (UID: \"b6e72ee2-50e4-4643-aee6-8be89d2e4589\") " pod="openshift-marketplace/community-operators-cgdxc" Nov 25 10:34:37 crc kubenswrapper[4821]: I1125 10:34:37.900336 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-r5nmb"] Nov 25 10:34:37 crc kubenswrapper[4821]: I1125 10:34:37.902518 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r5nmb" Nov 25 10:34:37 crc kubenswrapper[4821]: I1125 10:34:37.921109 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zrl8b" Nov 25 10:34:37 crc kubenswrapper[4821]: I1125 10:34:37.953798 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njv9h\" (UniqueName: \"kubernetes.io/projected/b6e72ee2-50e4-4643-aee6-8be89d2e4589-kube-api-access-njv9h\") pod \"community-operators-cgdxc\" (UID: \"b6e72ee2-50e4-4643-aee6-8be89d2e4589\") " pod="openshift-marketplace/community-operators-cgdxc" Nov 25 10:34:37 crc kubenswrapper[4821]: I1125 10:34:37.965642 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ps9cl" Nov 25 10:34:37 crc kubenswrapper[4821]: I1125 10:34:37.969142 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-r5nmb"] Nov 25 10:34:37 crc kubenswrapper[4821]: I1125 10:34:37.988681 4821 patch_prober.go:28] interesting pod/router-default-5444994796-p92sp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 10:34:37 crc kubenswrapper[4821]: [-]has-synced failed: reason withheld Nov 25 10:34:37 crc kubenswrapper[4821]: [+]process-running ok Nov 25 10:34:37 crc kubenswrapper[4821]: healthz check failed Nov 25 10:34:37 crc kubenswrapper[4821]: I1125 10:34:37.988728 4821 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p92sp" podUID="66798226-d262-4925-82db-2742d96a33fc" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 10:34:38 crc kubenswrapper[4821]: I1125 10:34:38.002892 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:38 crc kubenswrapper[4821]: I1125 10:34:38.003108 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f6dp4\" (UniqueName: \"kubernetes.io/projected/e1391070-3aad-47c3-a2f1-0c1697b014bf-kube-api-access-f6dp4\") pod \"certified-operators-r5nmb\" (UID: \"e1391070-3aad-47c3-a2f1-0c1697b014bf\") " pod="openshift-marketplace/certified-operators-r5nmb" Nov 25 10:34:38 crc kubenswrapper[4821]: I1125 10:34:38.003193 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1391070-3aad-47c3-a2f1-0c1697b014bf-catalog-content\") pod \"certified-operators-r5nmb\" (UID: \"e1391070-3aad-47c3-a2f1-0c1697b014bf\") " pod="openshift-marketplace/certified-operators-r5nmb" Nov 25 10:34:38 crc kubenswrapper[4821]: I1125 10:34:38.003220 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1391070-3aad-47c3-a2f1-0c1697b014bf-utilities\") pod \"certified-operators-r5nmb\" (UID: \"e1391070-3aad-47c3-a2f1-0c1697b014bf\") " pod="openshift-marketplace/certified-operators-r5nmb" Nov 25 10:34:38 crc kubenswrapper[4821]: E1125 10:34:38.003377 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:38.503361389 +0000 UTC m=+149.039681236 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:38 crc kubenswrapper[4821]: I1125 10:34:38.059314 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cgdxc" Nov 25 10:34:38 crc kubenswrapper[4821]: I1125 10:34:38.104879 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:38 crc kubenswrapper[4821]: I1125 10:34:38.104934 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f6dp4\" (UniqueName: \"kubernetes.io/projected/e1391070-3aad-47c3-a2f1-0c1697b014bf-kube-api-access-f6dp4\") pod \"certified-operators-r5nmb\" (UID: \"e1391070-3aad-47c3-a2f1-0c1697b014bf\") " pod="openshift-marketplace/certified-operators-r5nmb" Nov 25 10:34:38 crc kubenswrapper[4821]: I1125 10:34:38.105022 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1391070-3aad-47c3-a2f1-0c1697b014bf-catalog-content\") pod \"certified-operators-r5nmb\" (UID: \"e1391070-3aad-47c3-a2f1-0c1697b014bf\") " pod="openshift-marketplace/certified-operators-r5nmb" Nov 25 10:34:38 crc kubenswrapper[4821]: I1125 10:34:38.105060 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1391070-3aad-47c3-a2f1-0c1697b014bf-utilities\") pod \"certified-operators-r5nmb\" (UID: \"e1391070-3aad-47c3-a2f1-0c1697b014bf\") " pod="openshift-marketplace/certified-operators-r5nmb" Nov 25 10:34:38 crc kubenswrapper[4821]: I1125 10:34:38.105910 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1391070-3aad-47c3-a2f1-0c1697b014bf-utilities\") pod \"certified-operators-r5nmb\" (UID: \"e1391070-3aad-47c3-a2f1-0c1697b014bf\") " pod="openshift-marketplace/certified-operators-r5nmb" Nov 25 10:34:38 crc kubenswrapper[4821]: E1125 10:34:38.106213 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:38.606196369 +0000 UTC m=+149.142516216 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:38 crc kubenswrapper[4821]: I1125 10:34:38.106796 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1391070-3aad-47c3-a2f1-0c1697b014bf-catalog-content\") pod \"certified-operators-r5nmb\" (UID: \"e1391070-3aad-47c3-a2f1-0c1697b014bf\") " pod="openshift-marketplace/certified-operators-r5nmb" Nov 25 10:34:38 crc kubenswrapper[4821]: I1125 10:34:38.142772 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f6dp4\" (UniqueName: \"kubernetes.io/projected/e1391070-3aad-47c3-a2f1-0c1697b014bf-kube-api-access-f6dp4\") pod \"certified-operators-r5nmb\" (UID: \"e1391070-3aad-47c3-a2f1-0c1697b014bf\") " pod="openshift-marketplace/certified-operators-r5nmb" Nov 25 10:34:38 crc kubenswrapper[4821]: I1125 10:34:38.209677 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:38 crc kubenswrapper[4821]: E1125 10:34:38.209903 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:38.709872 +0000 UTC m=+149.246191867 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:38 crc kubenswrapper[4821]: I1125 10:34:38.234897 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"a549834d84cc714c9be0741f4c2daadab34fd572e0d3ad653eea91e610227a8c"} Nov 25 10:34:38 crc kubenswrapper[4821]: I1125 10:34:38.248451 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r5nmb" Nov 25 10:34:38 crc kubenswrapper[4821]: I1125 10:34:38.257411 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"a1a7b65b104dca4d839864f9da470ecba40a6ba9f6ceeeda3f527a1e24d4e71d"} Nov 25 10:34:38 crc kubenswrapper[4821]: I1125 10:34:38.311750 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:38 crc kubenswrapper[4821]: E1125 10:34:38.312199 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:38.812184346 +0000 UTC m=+149.348504193 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:38 crc kubenswrapper[4821]: I1125 10:34:38.413224 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:38 crc kubenswrapper[4821]: E1125 10:34:38.413395 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:38.913374305 +0000 UTC m=+149.449694152 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:38 crc kubenswrapper[4821]: I1125 10:34:38.413658 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:38 crc kubenswrapper[4821]: E1125 10:34:38.414820 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:38.914810391 +0000 UTC m=+149.451130348 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:38 crc kubenswrapper[4821]: I1125 10:34:38.523364 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:38 crc kubenswrapper[4821]: E1125 10:34:38.523601 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:39.023575028 +0000 UTC m=+149.559894875 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:38 crc kubenswrapper[4821]: I1125 10:34:38.523893 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:38 crc kubenswrapper[4821]: E1125 10:34:38.524364 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:39.024352947 +0000 UTC m=+149.560672794 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:38 crc kubenswrapper[4821]: I1125 10:34:38.624798 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:38 crc kubenswrapper[4821]: E1125 10:34:38.625220 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:39.125205428 +0000 UTC m=+149.661525275 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:38 crc kubenswrapper[4821]: I1125 10:34:38.653201 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zrl8b"] Nov 25 10:34:38 crc kubenswrapper[4821]: I1125 10:34:38.654603 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ps9cl"] Nov 25 10:34:38 crc kubenswrapper[4821]: W1125 10:34:38.682329 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod645993fd_bdce_435e_a600_34bf1ff5b38b.slice/crio-623f8448370b9aaeb71701b509f83954b1d63115c2e739cb4387ec020094e308 WatchSource:0}: Error finding container 623f8448370b9aaeb71701b509f83954b1d63115c2e739cb4387ec020094e308: Status 404 returned error can't find the container with id 623f8448370b9aaeb71701b509f83954b1d63115c2e739cb4387ec020094e308 Nov 25 10:34:38 crc kubenswrapper[4821]: I1125 10:34:38.726980 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:38 crc kubenswrapper[4821]: E1125 10:34:38.727330 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:39.227316609 +0000 UTC m=+149.763636466 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:38 crc kubenswrapper[4821]: I1125 10:34:38.791711 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cgdxc"] Nov 25 10:34:38 crc kubenswrapper[4821]: I1125 10:34:38.827886 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:38 crc kubenswrapper[4821]: E1125 10:34:38.828426 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:39.328404566 +0000 UTC m=+149.864724413 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:38 crc kubenswrapper[4821]: I1125 10:34:38.930437 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:38 crc kubenswrapper[4821]: E1125 10:34:38.930798 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:39.430782625 +0000 UTC m=+149.967102472 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:38 crc kubenswrapper[4821]: I1125 10:34:38.955110 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-r5nmb"] Nov 25 10:34:38 crc kubenswrapper[4821]: W1125 10:34:38.973331 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode1391070_3aad_47c3_a2f1_0c1697b014bf.slice/crio-00c07aed46d32e6e7ec0b81b24b2b64f7f0828080061143f799ce995c5d37992 WatchSource:0}: Error finding container 00c07aed46d32e6e7ec0b81b24b2b64f7f0828080061143f799ce995c5d37992: Status 404 returned error can't find the container with id 00c07aed46d32e6e7ec0b81b24b2b64f7f0828080061143f799ce995c5d37992 Nov 25 10:34:38 crc kubenswrapper[4821]: I1125 10:34:38.990540 4821 patch_prober.go:28] interesting pod/router-default-5444994796-p92sp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 10:34:38 crc kubenswrapper[4821]: [-]has-synced failed: reason withheld Nov 25 10:34:38 crc kubenswrapper[4821]: [+]process-running ok Nov 25 10:34:38 crc kubenswrapper[4821]: healthz check failed Nov 25 10:34:38 crc kubenswrapper[4821]: I1125 10:34:38.990585 4821 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p92sp" podUID="66798226-d262-4925-82db-2742d96a33fc" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 10:34:39 crc kubenswrapper[4821]: I1125 10:34:39.032760 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:39 crc kubenswrapper[4821]: E1125 10:34:39.033196 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:39.533177773 +0000 UTC m=+150.069497640 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:39 crc kubenswrapper[4821]: I1125 10:34:39.122038 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-q8tq8" Nov 25 10:34:39 crc kubenswrapper[4821]: I1125 10:34:39.134096 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:39 crc kubenswrapper[4821]: E1125 10:34:39.134670 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:39.634655199 +0000 UTC m=+150.170975046 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:39 crc kubenswrapper[4821]: I1125 10:34:39.166492 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 25 10:34:39 crc kubenswrapper[4821]: I1125 10:34:39.167456 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 10:34:39 crc kubenswrapper[4821]: I1125 10:34:39.171919 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Nov 25 10:34:39 crc kubenswrapper[4821]: I1125 10:34:39.172472 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Nov 25 10:34:39 crc kubenswrapper[4821]: I1125 10:34:39.182922 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 25 10:34:39 crc kubenswrapper[4821]: I1125 10:34:39.237765 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:39 crc kubenswrapper[4821]: E1125 10:34:39.238416 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:39.738398021 +0000 UTC m=+150.274717868 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:39 crc kubenswrapper[4821]: I1125 10:34:39.263803 4821 generic.go:334] "Generic (PLEG): container finished" podID="a5033129-ba8c-488f-982e-78e12b2f1664" containerID="21e6691e92fd38f9fb69206ae744da84aa6fde06b42b6560d4a335470deec9e3" exitCode=0 Nov 25 10:34:39 crc kubenswrapper[4821]: I1125 10:34:39.264087 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401110-dm86l" event={"ID":"a5033129-ba8c-488f-982e-78e12b2f1664","Type":"ContainerDied","Data":"21e6691e92fd38f9fb69206ae744da84aa6fde06b42b6560d4a335470deec9e3"} Nov 25 10:34:39 crc kubenswrapper[4821]: I1125 10:34:39.266791 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"d59dbbc9c351fcc1df15f15ddde8231b9f423b7e7f7056bc83ff208709a86adc"} Nov 25 10:34:39 crc kubenswrapper[4821]: I1125 10:34:39.269797 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ps9cl" event={"ID":"645993fd-bdce-435e-a600-34bf1ff5b38b","Type":"ContainerStarted","Data":"623f8448370b9aaeb71701b509f83954b1d63115c2e739cb4387ec020094e308"} Nov 25 10:34:39 crc kubenswrapper[4821]: I1125 10:34:39.270811 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zrl8b" event={"ID":"d1d6804a-27b6-4ef8-90ce-32606c93869a","Type":"ContainerStarted","Data":"4811a3c8d2e18aa424fa1d9934482633b4268620e85b281a326c72984ea29e55"} Nov 25 10:34:39 crc kubenswrapper[4821]: I1125 10:34:39.270916 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zrl8b" event={"ID":"d1d6804a-27b6-4ef8-90ce-32606c93869a","Type":"ContainerStarted","Data":"ff85d226a79311dc79303698b35f33a95a69497d27785c9a493d4a529ed46f8b"} Nov 25 10:34:39 crc kubenswrapper[4821]: I1125 10:34:39.273058 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"289357ccc0ca7b65668c6c2defb6ffdec3e80f1e7608bd0ca3918389cb2fe757"} Nov 25 10:34:39 crc kubenswrapper[4821]: I1125 10:34:39.273444 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:34:39 crc kubenswrapper[4821]: I1125 10:34:39.274667 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"67e3a9b8546b90ccbdac6dea6ef2aa28bb1015d0e8a63109f7e5870f41ca4107"} Nov 25 10:34:39 crc kubenswrapper[4821]: I1125 10:34:39.274772 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"05e76d2403a78763aaf27e28e3a51e3bd48c0a6582c756885dbad065d0ea6dee"} Nov 25 10:34:39 crc kubenswrapper[4821]: I1125 10:34:39.276433 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cgdxc" event={"ID":"b6e72ee2-50e4-4643-aee6-8be89d2e4589","Type":"ContainerStarted","Data":"7819efc6ad8867915226084d6f5e57ab5eaf307e7406719991c49c0364438b7e"} Nov 25 10:34:39 crc kubenswrapper[4821]: I1125 10:34:39.277640 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r5nmb" event={"ID":"e1391070-3aad-47c3-a2f1-0c1697b014bf","Type":"ContainerStarted","Data":"00c07aed46d32e6e7ec0b81b24b2b64f7f0828080061143f799ce995c5d37992"} Nov 25 10:34:39 crc kubenswrapper[4821]: I1125 10:34:39.320741 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-2kwcp"] Nov 25 10:34:39 crc kubenswrapper[4821]: I1125 10:34:39.322312 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2kwcp" Nov 25 10:34:39 crc kubenswrapper[4821]: I1125 10:34:39.326914 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 25 10:34:39 crc kubenswrapper[4821]: I1125 10:34:39.341864 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6a84b7fa-c197-4526-9373-dff5e0897293-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"6a84b7fa-c197-4526-9373-dff5e0897293\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 10:34:39 crc kubenswrapper[4821]: I1125 10:34:39.341942 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6a84b7fa-c197-4526-9373-dff5e0897293-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"6a84b7fa-c197-4526-9373-dff5e0897293\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 10:34:39 crc kubenswrapper[4821]: I1125 10:34:39.341990 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:39 crc kubenswrapper[4821]: E1125 10:34:39.342303 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:39.842291788 +0000 UTC m=+150.378611635 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:39 crc kubenswrapper[4821]: I1125 10:34:39.349173 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2kwcp"] Nov 25 10:34:39 crc kubenswrapper[4821]: I1125 10:34:39.443509 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:39 crc kubenswrapper[4821]: I1125 10:34:39.444061 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6a84b7fa-c197-4526-9373-dff5e0897293-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"6a84b7fa-c197-4526-9373-dff5e0897293\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 10:34:39 crc kubenswrapper[4821]: I1125 10:34:39.444147 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6a84b7fa-c197-4526-9373-dff5e0897293-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"6a84b7fa-c197-4526-9373-dff5e0897293\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 10:34:39 crc kubenswrapper[4821]: E1125 10:34:39.444253 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:39.944236845 +0000 UTC m=+150.480556692 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:39 crc kubenswrapper[4821]: I1125 10:34:39.444289 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:39 crc kubenswrapper[4821]: I1125 10:34:39.444313 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/927e7d40-852e-4807-bb22-d4f7c6e9d2ca-utilities\") pod \"redhat-marketplace-2kwcp\" (UID: \"927e7d40-852e-4807-bb22-d4f7c6e9d2ca\") " pod="openshift-marketplace/redhat-marketplace-2kwcp" Nov 25 10:34:39 crc kubenswrapper[4821]: I1125 10:34:39.444350 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/927e7d40-852e-4807-bb22-d4f7c6e9d2ca-catalog-content\") pod \"redhat-marketplace-2kwcp\" (UID: \"927e7d40-852e-4807-bb22-d4f7c6e9d2ca\") " pod="openshift-marketplace/redhat-marketplace-2kwcp" Nov 25 10:34:39 crc kubenswrapper[4821]: I1125 10:34:39.444377 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vlpdz\" (UniqueName: \"kubernetes.io/projected/927e7d40-852e-4807-bb22-d4f7c6e9d2ca-kube-api-access-vlpdz\") pod \"redhat-marketplace-2kwcp\" (UID: \"927e7d40-852e-4807-bb22-d4f7c6e9d2ca\") " pod="openshift-marketplace/redhat-marketplace-2kwcp" Nov 25 10:34:39 crc kubenswrapper[4821]: E1125 10:34:39.445582 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:39.945573228 +0000 UTC m=+150.481893075 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:39 crc kubenswrapper[4821]: I1125 10:34:39.445624 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6a84b7fa-c197-4526-9373-dff5e0897293-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"6a84b7fa-c197-4526-9373-dff5e0897293\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 10:34:39 crc kubenswrapper[4821]: I1125 10:34:39.503012 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6a84b7fa-c197-4526-9373-dff5e0897293-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"6a84b7fa-c197-4526-9373-dff5e0897293\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 10:34:39 crc kubenswrapper[4821]: I1125 10:34:39.552799 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:39 crc kubenswrapper[4821]: I1125 10:34:39.553026 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/927e7d40-852e-4807-bb22-d4f7c6e9d2ca-utilities\") pod \"redhat-marketplace-2kwcp\" (UID: \"927e7d40-852e-4807-bb22-d4f7c6e9d2ca\") " pod="openshift-marketplace/redhat-marketplace-2kwcp" Nov 25 10:34:39 crc kubenswrapper[4821]: I1125 10:34:39.553069 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/927e7d40-852e-4807-bb22-d4f7c6e9d2ca-catalog-content\") pod \"redhat-marketplace-2kwcp\" (UID: \"927e7d40-852e-4807-bb22-d4f7c6e9d2ca\") " pod="openshift-marketplace/redhat-marketplace-2kwcp" Nov 25 10:34:39 crc kubenswrapper[4821]: I1125 10:34:39.553097 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vlpdz\" (UniqueName: \"kubernetes.io/projected/927e7d40-852e-4807-bb22-d4f7c6e9d2ca-kube-api-access-vlpdz\") pod \"redhat-marketplace-2kwcp\" (UID: \"927e7d40-852e-4807-bb22-d4f7c6e9d2ca\") " pod="openshift-marketplace/redhat-marketplace-2kwcp" Nov 25 10:34:39 crc kubenswrapper[4821]: E1125 10:34:39.553530 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:40.053511976 +0000 UTC m=+150.589831823 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:39 crc kubenswrapper[4821]: I1125 10:34:39.554003 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/927e7d40-852e-4807-bb22-d4f7c6e9d2ca-catalog-content\") pod \"redhat-marketplace-2kwcp\" (UID: \"927e7d40-852e-4807-bb22-d4f7c6e9d2ca\") " pod="openshift-marketplace/redhat-marketplace-2kwcp" Nov 25 10:34:39 crc kubenswrapper[4821]: I1125 10:34:39.554295 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/927e7d40-852e-4807-bb22-d4f7c6e9d2ca-utilities\") pod \"redhat-marketplace-2kwcp\" (UID: \"927e7d40-852e-4807-bb22-d4f7c6e9d2ca\") " pod="openshift-marketplace/redhat-marketplace-2kwcp" Nov 25 10:34:39 crc kubenswrapper[4821]: I1125 10:34:39.593909 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vlpdz\" (UniqueName: \"kubernetes.io/projected/927e7d40-852e-4807-bb22-d4f7c6e9d2ca-kube-api-access-vlpdz\") pod \"redhat-marketplace-2kwcp\" (UID: \"927e7d40-852e-4807-bb22-d4f7c6e9d2ca\") " pod="openshift-marketplace/redhat-marketplace-2kwcp" Nov 25 10:34:39 crc kubenswrapper[4821]: I1125 10:34:39.637843 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2kwcp" Nov 25 10:34:39 crc kubenswrapper[4821]: I1125 10:34:39.654871 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:39 crc kubenswrapper[4821]: E1125 10:34:39.655334 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:40.155317699 +0000 UTC m=+150.691637546 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:39 crc kubenswrapper[4821]: I1125 10:34:39.699970 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-8qpmf"] Nov 25 10:34:39 crc kubenswrapper[4821]: I1125 10:34:39.700994 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8qpmf" Nov 25 10:34:39 crc kubenswrapper[4821]: I1125 10:34:39.755543 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:39 crc kubenswrapper[4821]: E1125 10:34:39.755754 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:40.255720959 +0000 UTC m=+150.792040806 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:39 crc kubenswrapper[4821]: I1125 10:34:39.756359 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:39 crc kubenswrapper[4821]: E1125 10:34:39.756696 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:40.256686653 +0000 UTC m=+150.793006500 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:39 crc kubenswrapper[4821]: I1125 10:34:39.759833 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8qpmf"] Nov 25 10:34:39 crc kubenswrapper[4821]: I1125 10:34:39.784663 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 10:34:39 crc kubenswrapper[4821]: I1125 10:34:39.858656 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:39 crc kubenswrapper[4821]: I1125 10:34:39.859276 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8becff5-8017-4073-89ae-c17cc4c32922-catalog-content\") pod \"redhat-marketplace-8qpmf\" (UID: \"d8becff5-8017-4073-89ae-c17cc4c32922\") " pod="openshift-marketplace/redhat-marketplace-8qpmf" Nov 25 10:34:39 crc kubenswrapper[4821]: I1125 10:34:39.859324 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gfw7p\" (UniqueName: \"kubernetes.io/projected/d8becff5-8017-4073-89ae-c17cc4c32922-kube-api-access-gfw7p\") pod \"redhat-marketplace-8qpmf\" (UID: \"d8becff5-8017-4073-89ae-c17cc4c32922\") " pod="openshift-marketplace/redhat-marketplace-8qpmf" Nov 25 10:34:39 crc kubenswrapper[4821]: E1125 10:34:39.859478 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:40.359452231 +0000 UTC m=+150.895772078 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:39 crc kubenswrapper[4821]: I1125 10:34:39.859515 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8becff5-8017-4073-89ae-c17cc4c32922-utilities\") pod \"redhat-marketplace-8qpmf\" (UID: \"d8becff5-8017-4073-89ae-c17cc4c32922\") " pod="openshift-marketplace/redhat-marketplace-8qpmf" Nov 25 10:34:39 crc kubenswrapper[4821]: I1125 10:34:39.960933 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8becff5-8017-4073-89ae-c17cc4c32922-utilities\") pod \"redhat-marketplace-8qpmf\" (UID: \"d8becff5-8017-4073-89ae-c17cc4c32922\") " pod="openshift-marketplace/redhat-marketplace-8qpmf" Nov 25 10:34:39 crc kubenswrapper[4821]: I1125 10:34:39.960993 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8becff5-8017-4073-89ae-c17cc4c32922-catalog-content\") pod \"redhat-marketplace-8qpmf\" (UID: \"d8becff5-8017-4073-89ae-c17cc4c32922\") " pod="openshift-marketplace/redhat-marketplace-8qpmf" Nov 25 10:34:39 crc kubenswrapper[4821]: I1125 10:34:39.961031 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gfw7p\" (UniqueName: \"kubernetes.io/projected/d8becff5-8017-4073-89ae-c17cc4c32922-kube-api-access-gfw7p\") pod \"redhat-marketplace-8qpmf\" (UID: \"d8becff5-8017-4073-89ae-c17cc4c32922\") " pod="openshift-marketplace/redhat-marketplace-8qpmf" Nov 25 10:34:39 crc kubenswrapper[4821]: I1125 10:34:39.961104 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:39 crc kubenswrapper[4821]: I1125 10:34:39.961428 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8becff5-8017-4073-89ae-c17cc4c32922-utilities\") pod \"redhat-marketplace-8qpmf\" (UID: \"d8becff5-8017-4073-89ae-c17cc4c32922\") " pod="openshift-marketplace/redhat-marketplace-8qpmf" Nov 25 10:34:39 crc kubenswrapper[4821]: E1125 10:34:39.961497 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:40.461484011 +0000 UTC m=+150.997803928 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:39 crc kubenswrapper[4821]: I1125 10:34:39.961662 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8becff5-8017-4073-89ae-c17cc4c32922-catalog-content\") pod \"redhat-marketplace-8qpmf\" (UID: \"d8becff5-8017-4073-89ae-c17cc4c32922\") " pod="openshift-marketplace/redhat-marketplace-8qpmf" Nov 25 10:34:39 crc kubenswrapper[4821]: I1125 10:34:39.986344 4821 patch_prober.go:28] interesting pod/router-default-5444994796-p92sp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 10:34:39 crc kubenswrapper[4821]: [-]has-synced failed: reason withheld Nov 25 10:34:39 crc kubenswrapper[4821]: [+]process-running ok Nov 25 10:34:39 crc kubenswrapper[4821]: healthz check failed Nov 25 10:34:39 crc kubenswrapper[4821]: I1125 10:34:39.986422 4821 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p92sp" podUID="66798226-d262-4925-82db-2742d96a33fc" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 10:34:40 crc kubenswrapper[4821]: I1125 10:34:40.008026 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gfw7p\" (UniqueName: \"kubernetes.io/projected/d8becff5-8017-4073-89ae-c17cc4c32922-kube-api-access-gfw7p\") pod \"redhat-marketplace-8qpmf\" (UID: \"d8becff5-8017-4073-89ae-c17cc4c32922\") " pod="openshift-marketplace/redhat-marketplace-8qpmf" Nov 25 10:34:40 crc kubenswrapper[4821]: I1125 10:34:40.019507 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8qpmf" Nov 25 10:34:40 crc kubenswrapper[4821]: I1125 10:34:40.061986 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:40 crc kubenswrapper[4821]: E1125 10:34:40.062510 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:40.562488595 +0000 UTC m=+151.098808442 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:40 crc kubenswrapper[4821]: I1125 10:34:40.077331 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2kwcp"] Nov 25 10:34:40 crc kubenswrapper[4821]: W1125 10:34:40.089455 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod927e7d40_852e_4807_bb22_d4f7c6e9d2ca.slice/crio-23cd67b4ab339669af677f251279be6c3db37de80867eaede62cde905682b7ad WatchSource:0}: Error finding container 23cd67b4ab339669af677f251279be6c3db37de80867eaede62cde905682b7ad: Status 404 returned error can't find the container with id 23cd67b4ab339669af677f251279be6c3db37de80867eaede62cde905682b7ad Nov 25 10:34:40 crc kubenswrapper[4821]: I1125 10:34:40.143130 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-t5hm7" Nov 25 10:34:40 crc kubenswrapper[4821]: I1125 10:34:40.143178 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-hf49d" Nov 25 10:34:40 crc kubenswrapper[4821]: I1125 10:34:40.143193 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-hf49d" Nov 25 10:34:40 crc kubenswrapper[4821]: I1125 10:34:40.150330 4821 patch_prober.go:28] interesting pod/console-f9d7485db-hf49d container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.8:8443/health\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Nov 25 10:34:40 crc kubenswrapper[4821]: I1125 10:34:40.150379 4821 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-hf49d" podUID="88725123-3da9-4127-9506-80c90c9accd5" containerName="console" probeResult="failure" output="Get \"https://10.217.0.8:8443/health\": dial tcp 10.217.0.8:8443: connect: connection refused" Nov 25 10:34:40 crc kubenswrapper[4821]: I1125 10:34:40.164897 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:40 crc kubenswrapper[4821]: I1125 10:34:40.165137 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-58htt" Nov 25 10:34:40 crc kubenswrapper[4821]: E1125 10:34:40.165192 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:40.665180121 +0000 UTC m=+151.201499968 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:40 crc kubenswrapper[4821]: I1125 10:34:40.265742 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:40 crc kubenswrapper[4821]: E1125 10:34:40.267453 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:40.767434587 +0000 UTC m=+151.303754434 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:40 crc kubenswrapper[4821]: I1125 10:34:40.302809 4821 generic.go:334] "Generic (PLEG): container finished" podID="b6e72ee2-50e4-4643-aee6-8be89d2e4589" containerID="d4a3c9bae38e7ab88e8c11dc78a97a4907de72007e8373aded9b9c84c97c41b5" exitCode=0 Nov 25 10:34:40 crc kubenswrapper[4821]: I1125 10:34:40.303044 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cgdxc" event={"ID":"b6e72ee2-50e4-4643-aee6-8be89d2e4589","Type":"ContainerDied","Data":"d4a3c9bae38e7ab88e8c11dc78a97a4907de72007e8373aded9b9c84c97c41b5"} Nov 25 10:34:40 crc kubenswrapper[4821]: I1125 10:34:40.305951 4821 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 10:34:40 crc kubenswrapper[4821]: I1125 10:34:40.320074 4821 generic.go:334] "Generic (PLEG): container finished" podID="e1391070-3aad-47c3-a2f1-0c1697b014bf" containerID="b4a2f4f11681121adddba94e268de602497790e7562df7b11e62b1660e19ebe2" exitCode=0 Nov 25 10:34:40 crc kubenswrapper[4821]: I1125 10:34:40.320232 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r5nmb" event={"ID":"e1391070-3aad-47c3-a2f1-0c1697b014bf","Type":"ContainerDied","Data":"b4a2f4f11681121adddba94e268de602497790e7562df7b11e62b1660e19ebe2"} Nov 25 10:34:40 crc kubenswrapper[4821]: I1125 10:34:40.355852 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2kwcp" event={"ID":"927e7d40-852e-4807-bb22-d4f7c6e9d2ca","Type":"ContainerStarted","Data":"23cd67b4ab339669af677f251279be6c3db37de80867eaede62cde905682b7ad"} Nov 25 10:34:40 crc kubenswrapper[4821]: I1125 10:34:40.375752 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:40 crc kubenswrapper[4821]: E1125 10:34:40.380210 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:40.876237665 +0000 UTC m=+151.412557512 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:40 crc kubenswrapper[4821]: I1125 10:34:40.382253 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 25 10:34:40 crc kubenswrapper[4821]: I1125 10:34:40.393730 4821 generic.go:334] "Generic (PLEG): container finished" podID="645993fd-bdce-435e-a600-34bf1ff5b38b" containerID="10de088200d33e95b9677889b71c6c6aa52bf166742c2c0f86179375641acb43" exitCode=0 Nov 25 10:34:40 crc kubenswrapper[4821]: I1125 10:34:40.393847 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ps9cl" event={"ID":"645993fd-bdce-435e-a600-34bf1ff5b38b","Type":"ContainerDied","Data":"10de088200d33e95b9677889b71c6c6aa52bf166742c2c0f86179375641acb43"} Nov 25 10:34:40 crc kubenswrapper[4821]: I1125 10:34:40.404970 4821 generic.go:334] "Generic (PLEG): container finished" podID="d1d6804a-27b6-4ef8-90ce-32606c93869a" containerID="4811a3c8d2e18aa424fa1d9934482633b4268620e85b281a326c72984ea29e55" exitCode=0 Nov 25 10:34:40 crc kubenswrapper[4821]: I1125 10:34:40.406120 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zrl8b" event={"ID":"d1d6804a-27b6-4ef8-90ce-32606c93869a","Type":"ContainerDied","Data":"4811a3c8d2e18aa424fa1d9934482633b4268620e85b281a326c72984ea29e55"} Nov 25 10:34:40 crc kubenswrapper[4821]: I1125 10:34:40.420767 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-mbn4k" Nov 25 10:34:40 crc kubenswrapper[4821]: I1125 10:34:40.477569 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:40 crc kubenswrapper[4821]: E1125 10:34:40.477721 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:40.97768146 +0000 UTC m=+151.514001307 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:40 crc kubenswrapper[4821]: I1125 10:34:40.478118 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:40 crc kubenswrapper[4821]: E1125 10:34:40.479633 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:40.979622188 +0000 UTC m=+151.515942035 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:40 crc kubenswrapper[4821]: I1125 10:34:40.516297 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-w56ql"] Nov 25 10:34:40 crc kubenswrapper[4821]: I1125 10:34:40.517659 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w56ql" Nov 25 10:34:40 crc kubenswrapper[4821]: I1125 10:34:40.526087 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 25 10:34:40 crc kubenswrapper[4821]: I1125 10:34:40.544849 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-w56ql"] Nov 25 10:34:40 crc kubenswrapper[4821]: I1125 10:34:40.579808 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:40 crc kubenswrapper[4821]: E1125 10:34:40.583005 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:41.082986001 +0000 UTC m=+151.619305848 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:40 crc kubenswrapper[4821]: I1125 10:34:40.693982 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:40 crc kubenswrapper[4821]: I1125 10:34:40.694056 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nstfp\" (UniqueName: \"kubernetes.io/projected/4b4b52a8-67ad-4c51-86ab-503c75ad2024-kube-api-access-nstfp\") pod \"redhat-operators-w56ql\" (UID: \"4b4b52a8-67ad-4c51-86ab-503c75ad2024\") " pod="openshift-marketplace/redhat-operators-w56ql" Nov 25 10:34:40 crc kubenswrapper[4821]: I1125 10:34:40.694104 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b4b52a8-67ad-4c51-86ab-503c75ad2024-catalog-content\") pod \"redhat-operators-w56ql\" (UID: \"4b4b52a8-67ad-4c51-86ab-503c75ad2024\") " pod="openshift-marketplace/redhat-operators-w56ql" Nov 25 10:34:40 crc kubenswrapper[4821]: I1125 10:34:40.694207 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b4b52a8-67ad-4c51-86ab-503c75ad2024-utilities\") pod \"redhat-operators-w56ql\" (UID: \"4b4b52a8-67ad-4c51-86ab-503c75ad2024\") " pod="openshift-marketplace/redhat-operators-w56ql" Nov 25 10:34:40 crc kubenswrapper[4821]: E1125 10:34:40.694469 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:41.194458905 +0000 UTC m=+151.730778752 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:40 crc kubenswrapper[4821]: I1125 10:34:40.798672 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:40 crc kubenswrapper[4821]: I1125 10:34:40.798880 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b4b52a8-67ad-4c51-86ab-503c75ad2024-catalog-content\") pod \"redhat-operators-w56ql\" (UID: \"4b4b52a8-67ad-4c51-86ab-503c75ad2024\") " pod="openshift-marketplace/redhat-operators-w56ql" Nov 25 10:34:40 crc kubenswrapper[4821]: I1125 10:34:40.798944 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b4b52a8-67ad-4c51-86ab-503c75ad2024-utilities\") pod \"redhat-operators-w56ql\" (UID: \"4b4b52a8-67ad-4c51-86ab-503c75ad2024\") " pod="openshift-marketplace/redhat-operators-w56ql" Nov 25 10:34:40 crc kubenswrapper[4821]: E1125 10:34:40.799010 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:41.298986127 +0000 UTC m=+151.835305974 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:40 crc kubenswrapper[4821]: I1125 10:34:40.799138 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:40 crc kubenswrapper[4821]: I1125 10:34:40.799263 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nstfp\" (UniqueName: \"kubernetes.io/projected/4b4b52a8-67ad-4c51-86ab-503c75ad2024-kube-api-access-nstfp\") pod \"redhat-operators-w56ql\" (UID: \"4b4b52a8-67ad-4c51-86ab-503c75ad2024\") " pod="openshift-marketplace/redhat-operators-w56ql" Nov 25 10:34:40 crc kubenswrapper[4821]: I1125 10:34:40.799384 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b4b52a8-67ad-4c51-86ab-503c75ad2024-utilities\") pod \"redhat-operators-w56ql\" (UID: \"4b4b52a8-67ad-4c51-86ab-503c75ad2024\") " pod="openshift-marketplace/redhat-operators-w56ql" Nov 25 10:34:40 crc kubenswrapper[4821]: I1125 10:34:40.799659 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b4b52a8-67ad-4c51-86ab-503c75ad2024-catalog-content\") pod \"redhat-operators-w56ql\" (UID: \"4b4b52a8-67ad-4c51-86ab-503c75ad2024\") " pod="openshift-marketplace/redhat-operators-w56ql" Nov 25 10:34:40 crc kubenswrapper[4821]: E1125 10:34:40.799855 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:41.299844018 +0000 UTC m=+151.836163865 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:40 crc kubenswrapper[4821]: I1125 10:34:40.809300 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8qpmf"] Nov 25 10:34:40 crc kubenswrapper[4821]: I1125 10:34:40.821810 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nstfp\" (UniqueName: \"kubernetes.io/projected/4b4b52a8-67ad-4c51-86ab-503c75ad2024-kube-api-access-nstfp\") pod \"redhat-operators-w56ql\" (UID: \"4b4b52a8-67ad-4c51-86ab-503c75ad2024\") " pod="openshift-marketplace/redhat-operators-w56ql" Nov 25 10:34:40 crc kubenswrapper[4821]: I1125 10:34:40.858106 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w56ql" Nov 25 10:34:40 crc kubenswrapper[4821]: I1125 10:34:40.895294 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-87rxm"] Nov 25 10:34:40 crc kubenswrapper[4821]: I1125 10:34:40.899764 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-87rxm" Nov 25 10:34:40 crc kubenswrapper[4821]: I1125 10:34:40.899884 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:40 crc kubenswrapper[4821]: E1125 10:34:40.900265 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:41.400250267 +0000 UTC m=+151.936570114 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:40 crc kubenswrapper[4821]: I1125 10:34:40.909263 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-87rxm"] Nov 25 10:34:40 crc kubenswrapper[4821]: I1125 10:34:40.981521 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401110-dm86l" Nov 25 10:34:40 crc kubenswrapper[4821]: I1125 10:34:40.986246 4821 patch_prober.go:28] interesting pod/router-default-5444994796-p92sp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 10:34:40 crc kubenswrapper[4821]: [-]has-synced failed: reason withheld Nov 25 10:34:40 crc kubenswrapper[4821]: [+]process-running ok Nov 25 10:34:40 crc kubenswrapper[4821]: healthz check failed Nov 25 10:34:40 crc kubenswrapper[4821]: I1125 10:34:40.986299 4821 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p92sp" podUID="66798226-d262-4925-82db-2742d96a33fc" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 10:34:40 crc kubenswrapper[4821]: I1125 10:34:40.986329 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4kjbr" Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.006666 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38068a8c-fd52-492c-9e1e-6425830bd097-catalog-content\") pod \"redhat-operators-87rxm\" (UID: \"38068a8c-fd52-492c-9e1e-6425830bd097\") " pod="openshift-marketplace/redhat-operators-87rxm" Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.006751 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.006835 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6gpf2\" (UniqueName: \"kubernetes.io/projected/38068a8c-fd52-492c-9e1e-6425830bd097-kube-api-access-6gpf2\") pod \"redhat-operators-87rxm\" (UID: \"38068a8c-fd52-492c-9e1e-6425830bd097\") " pod="openshift-marketplace/redhat-operators-87rxm" Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.006863 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38068a8c-fd52-492c-9e1e-6425830bd097-utilities\") pod \"redhat-operators-87rxm\" (UID: \"38068a8c-fd52-492c-9e1e-6425830bd097\") " pod="openshift-marketplace/redhat-operators-87rxm" Nov 25 10:34:41 crc kubenswrapper[4821]: E1125 10:34:41.007184 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:41.507149228 +0000 UTC m=+152.043469085 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.108058 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.108404 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a5033129-ba8c-488f-982e-78e12b2f1664-config-volume\") pod \"a5033129-ba8c-488f-982e-78e12b2f1664\" (UID: \"a5033129-ba8c-488f-982e-78e12b2f1664\") " Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.108457 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a5033129-ba8c-488f-982e-78e12b2f1664-secret-volume\") pod \"a5033129-ba8c-488f-982e-78e12b2f1664\" (UID: \"a5033129-ba8c-488f-982e-78e12b2f1664\") " Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.108475 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g52j5\" (UniqueName: \"kubernetes.io/projected/a5033129-ba8c-488f-982e-78e12b2f1664-kube-api-access-g52j5\") pod \"a5033129-ba8c-488f-982e-78e12b2f1664\" (UID: \"a5033129-ba8c-488f-982e-78e12b2f1664\") " Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.108612 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38068a8c-fd52-492c-9e1e-6425830bd097-catalog-content\") pod \"redhat-operators-87rxm\" (UID: \"38068a8c-fd52-492c-9e1e-6425830bd097\") " pod="openshift-marketplace/redhat-operators-87rxm" Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.108744 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6gpf2\" (UniqueName: \"kubernetes.io/projected/38068a8c-fd52-492c-9e1e-6425830bd097-kube-api-access-6gpf2\") pod \"redhat-operators-87rxm\" (UID: \"38068a8c-fd52-492c-9e1e-6425830bd097\") " pod="openshift-marketplace/redhat-operators-87rxm" Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.108771 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38068a8c-fd52-492c-9e1e-6425830bd097-utilities\") pod \"redhat-operators-87rxm\" (UID: \"38068a8c-fd52-492c-9e1e-6425830bd097\") " pod="openshift-marketplace/redhat-operators-87rxm" Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.109976 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a5033129-ba8c-488f-982e-78e12b2f1664-config-volume" (OuterVolumeSpecName: "config-volume") pod "a5033129-ba8c-488f-982e-78e12b2f1664" (UID: "a5033129-ba8c-488f-982e-78e12b2f1664"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.110527 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38068a8c-fd52-492c-9e1e-6425830bd097-utilities\") pod \"redhat-operators-87rxm\" (UID: \"38068a8c-fd52-492c-9e1e-6425830bd097\") " pod="openshift-marketplace/redhat-operators-87rxm" Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.110742 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38068a8c-fd52-492c-9e1e-6425830bd097-catalog-content\") pod \"redhat-operators-87rxm\" (UID: \"38068a8c-fd52-492c-9e1e-6425830bd097\") " pod="openshift-marketplace/redhat-operators-87rxm" Nov 25 10:34:41 crc kubenswrapper[4821]: E1125 10:34:41.111367 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:41.611350122 +0000 UTC m=+152.147669969 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.116886 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5033129-ba8c-488f-982e-78e12b2f1664-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "a5033129-ba8c-488f-982e-78e12b2f1664" (UID: "a5033129-ba8c-488f-982e-78e12b2f1664"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.128222 4821 patch_prober.go:28] interesting pod/downloads-7954f5f757-hjvt9 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.128272 4821 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-hjvt9" podUID="f954533e-e8a0-4899-be72-241760feae95" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.128721 4821 patch_prober.go:28] interesting pod/downloads-7954f5f757-hjvt9 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.128738 4821 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-hjvt9" podUID="f954533e-e8a0-4899-be72-241760feae95" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.128869 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5033129-ba8c-488f-982e-78e12b2f1664-kube-api-access-g52j5" (OuterVolumeSpecName: "kube-api-access-g52j5") pod "a5033129-ba8c-488f-982e-78e12b2f1664" (UID: "a5033129-ba8c-488f-982e-78e12b2f1664"). InnerVolumeSpecName "kube-api-access-g52j5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.139493 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6gpf2\" (UniqueName: \"kubernetes.io/projected/38068a8c-fd52-492c-9e1e-6425830bd097-kube-api-access-6gpf2\") pod \"redhat-operators-87rxm\" (UID: \"38068a8c-fd52-492c-9e1e-6425830bd097\") " pod="openshift-marketplace/redhat-operators-87rxm" Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.210484 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.210545 4821 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a5033129-ba8c-488f-982e-78e12b2f1664-config-volume\") on node \"crc\" DevicePath \"\"" Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.210556 4821 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a5033129-ba8c-488f-982e-78e12b2f1664-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.210565 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g52j5\" (UniqueName: \"kubernetes.io/projected/a5033129-ba8c-488f-982e-78e12b2f1664-kube-api-access-g52j5\") on node \"crc\" DevicePath \"\"" Nov 25 10:34:41 crc kubenswrapper[4821]: E1125 10:34:41.210818 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:41.710806428 +0000 UTC m=+152.247126275 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.213328 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-jb5sz" Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.213800 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-jb5sz" Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.224591 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-87rxm" Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.225443 4821 patch_prober.go:28] interesting pod/apiserver-76f77b778f-jb5sz container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="Get \"https://10.217.0.14:8443/livez\": dial tcp 10.217.0.14:8443: connect: connection refused" start-of-body= Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.225475 4821 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-jb5sz" podUID="064de93d-7bb5-47b5-9585-bf32be9a05f6" containerName="openshift-apiserver" probeResult="failure" output="Get \"https://10.217.0.14:8443/livez\": dial tcp 10.217.0.14:8443: connect: connection refused" Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.314090 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-w56ql"] Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.314539 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:41 crc kubenswrapper[4821]: E1125 10:34:41.315927 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:41.815912044 +0000 UTC m=+152.352231891 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.396511 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-dxf4x" Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.412649 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d8x5t" Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.415804 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:41 crc kubenswrapper[4821]: E1125 10:34:41.416125 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:41.916113639 +0000 UTC m=+152.452433486 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.429124 4821 generic.go:334] "Generic (PLEG): container finished" podID="927e7d40-852e-4807-bb22-d4f7c6e9d2ca" containerID="977f93d972d776b41806449e65352f20756f6a5350d48e77262c8f4e3e6a19e2" exitCode=0 Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.429475 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2kwcp" event={"ID":"927e7d40-852e-4807-bb22-d4f7c6e9d2ca","Type":"ContainerDied","Data":"977f93d972d776b41806449e65352f20756f6a5350d48e77262c8f4e3e6a19e2"} Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.438431 4821 generic.go:334] "Generic (PLEG): container finished" podID="d8becff5-8017-4073-89ae-c17cc4c32922" containerID="47a1a83c452d0edae993c649647470792cb7de9dafb21beba9dbd623ec1d6852" exitCode=0 Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.438491 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8qpmf" event={"ID":"d8becff5-8017-4073-89ae-c17cc4c32922","Type":"ContainerDied","Data":"47a1a83c452d0edae993c649647470792cb7de9dafb21beba9dbd623ec1d6852"} Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.438516 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8qpmf" event={"ID":"d8becff5-8017-4073-89ae-c17cc4c32922","Type":"ContainerStarted","Data":"0262a0c2a5dbf2dff7fb7034ccd1c965b0de806bc73a49b7084bd6ac6cae9918"} Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.444711 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d8x5t" Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.450299 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"6a84b7fa-c197-4526-9373-dff5e0897293","Type":"ContainerStarted","Data":"c96ca76bbafcaa1b34e97c60381fdc87336a5604d33380d850c668370e821bae"} Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.450355 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"6a84b7fa-c197-4526-9373-dff5e0897293","Type":"ContainerStarted","Data":"2f7684584a7dac78105c67c3f42c6098a77555820df5873e8bee726ec3c40534"} Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.454860 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w56ql" event={"ID":"4b4b52a8-67ad-4c51-86ab-503c75ad2024","Type":"ContainerStarted","Data":"e34dfb84c2952d3e16b75ad704ace4aa1e336f2eef64d1f275e8442e499a3df2"} Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.461440 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401110-dm86l" event={"ID":"a5033129-ba8c-488f-982e-78e12b2f1664","Type":"ContainerDied","Data":"f647a6676a61dff0159b32ea289553f18ca7336aafb1fec95c26f25c594d6e9e"} Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.461485 4821 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f647a6676a61dff0159b32ea289553f18ca7336aafb1fec95c26f25c594d6e9e" Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.461493 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401110-dm86l" Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.521772 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:41 crc kubenswrapper[4821]: E1125 10:34:41.522477 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:42.022451856 +0000 UTC m=+152.558771713 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.524099 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:41 crc kubenswrapper[4821]: E1125 10:34:41.524439 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:42.024424994 +0000 UTC m=+152.560744841 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.626972 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:41 crc kubenswrapper[4821]: E1125 10:34:41.627214 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:42.127143612 +0000 UTC m=+152.663463459 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.627392 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:41 crc kubenswrapper[4821]: E1125 10:34:41.627721 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:42.127709805 +0000 UTC m=+152.664029712 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.673073 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kps6n" Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.674024 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kps6n" Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.685371 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kps6n" Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.699145 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 25 10:34:41 crc kubenswrapper[4821]: E1125 10:34:41.699491 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5033129-ba8c-488f-982e-78e12b2f1664" containerName="collect-profiles" Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.699514 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5033129-ba8c-488f-982e-78e12b2f1664" containerName="collect-profiles" Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.699654 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5033129-ba8c-488f-982e-78e12b2f1664" containerName="collect-profiles" Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.700054 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.702079 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.702317 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.707480 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.729355 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.729671 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f5536d76-8475-416d-ac52-649819925b7c-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"f5536d76-8475-416d-ac52-649819925b7c\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.729764 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f5536d76-8475-416d-ac52-649819925b7c-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"f5536d76-8475-416d-ac52-649819925b7c\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 10:34:41 crc kubenswrapper[4821]: E1125 10:34:41.730506 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:42.230486193 +0000 UTC m=+152.766806040 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.784868 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-87rxm"] Nov 25 10:34:41 crc kubenswrapper[4821]: W1125 10:34:41.799052 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod38068a8c_fd52_492c_9e1e_6425830bd097.slice/crio-8ce0c031e2afe5eb9c02d7a8bf7a49a4a7470c86707d1879863de9afea8a4e43 WatchSource:0}: Error finding container 8ce0c031e2afe5eb9c02d7a8bf7a49a4a7470c86707d1879863de9afea8a4e43: Status 404 returned error can't find the container with id 8ce0c031e2afe5eb9c02d7a8bf7a49a4a7470c86707d1879863de9afea8a4e43 Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.831060 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.831125 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f5536d76-8475-416d-ac52-649819925b7c-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"f5536d76-8475-416d-ac52-649819925b7c\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.831156 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f5536d76-8475-416d-ac52-649819925b7c-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"f5536d76-8475-416d-ac52-649819925b7c\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.831264 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f5536d76-8475-416d-ac52-649819925b7c-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"f5536d76-8475-416d-ac52-649819925b7c\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 10:34:41 crc kubenswrapper[4821]: E1125 10:34:41.831411 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:42.331393035 +0000 UTC m=+152.867712952 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.854556 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f5536d76-8475-416d-ac52-649819925b7c-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"f5536d76-8475-416d-ac52-649819925b7c\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.935671 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:41 crc kubenswrapper[4821]: E1125 10:34:41.936295 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:42.436275466 +0000 UTC m=+152.972595313 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.981681 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-p92sp" Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.985594 4821 patch_prober.go:28] interesting pod/router-default-5444994796-p92sp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 10:34:41 crc kubenswrapper[4821]: [-]has-synced failed: reason withheld Nov 25 10:34:41 crc kubenswrapper[4821]: [+]process-running ok Nov 25 10:34:41 crc kubenswrapper[4821]: healthz check failed Nov 25 10:34:41 crc kubenswrapper[4821]: I1125 10:34:41.985639 4821 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p92sp" podUID="66798226-d262-4925-82db-2742d96a33fc" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 10:34:42 crc kubenswrapper[4821]: I1125 10:34:42.038086 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:42 crc kubenswrapper[4821]: E1125 10:34:42.038468 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:42.538455629 +0000 UTC m=+153.074775476 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:42 crc kubenswrapper[4821]: I1125 10:34:42.053216 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 10:34:42 crc kubenswrapper[4821]: I1125 10:34:42.139090 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:42 crc kubenswrapper[4821]: E1125 10:34:42.139317 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:42.639284469 +0000 UTC m=+153.175604326 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:42 crc kubenswrapper[4821]: I1125 10:34:42.139505 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:42 crc kubenswrapper[4821]: E1125 10:34:42.140698 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:42.640687234 +0000 UTC m=+153.177007181 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:42 crc kubenswrapper[4821]: I1125 10:34:42.241827 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:42 crc kubenswrapper[4821]: E1125 10:34:42.242065 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:42.742025297 +0000 UTC m=+153.278345144 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:42 crc kubenswrapper[4821]: I1125 10:34:42.242540 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:42 crc kubenswrapper[4821]: E1125 10:34:42.243318 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:42.743308839 +0000 UTC m=+153.279628686 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:42 crc kubenswrapper[4821]: I1125 10:34:42.275974 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 25 10:34:42 crc kubenswrapper[4821]: I1125 10:34:42.343646 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:42 crc kubenswrapper[4821]: E1125 10:34:42.343835 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:42.843809651 +0000 UTC m=+153.380129498 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:42 crc kubenswrapper[4821]: I1125 10:34:42.343930 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:42 crc kubenswrapper[4821]: E1125 10:34:42.344327 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:42.844311654 +0000 UTC m=+153.380631501 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:42 crc kubenswrapper[4821]: I1125 10:34:42.403825 4821 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Nov 25 10:34:42 crc kubenswrapper[4821]: I1125 10:34:42.445653 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:42 crc kubenswrapper[4821]: E1125 10:34:42.445960 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:42.945941013 +0000 UTC m=+153.482260860 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:42 crc kubenswrapper[4821]: I1125 10:34:42.482988 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-tsb45" event={"ID":"f97c276b-2b69-49a0-96d1-b1c5171a542e","Type":"ContainerStarted","Data":"ba5a8557581aa81bda36236ed0d3a0f2edd644999c9ff0b12765b6966dc395f4"} Nov 25 10:34:42 crc kubenswrapper[4821]: I1125 10:34:42.483040 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-tsb45" event={"ID":"f97c276b-2b69-49a0-96d1-b1c5171a542e","Type":"ContainerStarted","Data":"679bd1f65e33b43f52fe555ae0bbf5617c0f7509d571a4790095ebbd623373af"} Nov 25 10:34:42 crc kubenswrapper[4821]: I1125 10:34:42.485012 4821 generic.go:334] "Generic (PLEG): container finished" podID="6a84b7fa-c197-4526-9373-dff5e0897293" containerID="c96ca76bbafcaa1b34e97c60381fdc87336a5604d33380d850c668370e821bae" exitCode=0 Nov 25 10:34:42 crc kubenswrapper[4821]: I1125 10:34:42.485093 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"6a84b7fa-c197-4526-9373-dff5e0897293","Type":"ContainerDied","Data":"c96ca76bbafcaa1b34e97c60381fdc87336a5604d33380d850c668370e821bae"} Nov 25 10:34:42 crc kubenswrapper[4821]: I1125 10:34:42.487387 4821 generic.go:334] "Generic (PLEG): container finished" podID="38068a8c-fd52-492c-9e1e-6425830bd097" containerID="9566c37aaf942dcc55362ff747f161856ea86449e19e5c458bda884d3e4afae5" exitCode=0 Nov 25 10:34:42 crc kubenswrapper[4821]: I1125 10:34:42.487426 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-87rxm" event={"ID":"38068a8c-fd52-492c-9e1e-6425830bd097","Type":"ContainerDied","Data":"9566c37aaf942dcc55362ff747f161856ea86449e19e5c458bda884d3e4afae5"} Nov 25 10:34:42 crc kubenswrapper[4821]: I1125 10:34:42.487463 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-87rxm" event={"ID":"38068a8c-fd52-492c-9e1e-6425830bd097","Type":"ContainerStarted","Data":"8ce0c031e2afe5eb9c02d7a8bf7a49a4a7470c86707d1879863de9afea8a4e43"} Nov 25 10:34:42 crc kubenswrapper[4821]: I1125 10:34:42.488636 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"f5536d76-8475-416d-ac52-649819925b7c","Type":"ContainerStarted","Data":"8fcf85b7d351fef14720f52b69276519b1c8072f11a07ae296948b556c4f8263"} Nov 25 10:34:42 crc kubenswrapper[4821]: I1125 10:34:42.491318 4821 generic.go:334] "Generic (PLEG): container finished" podID="4b4b52a8-67ad-4c51-86ab-503c75ad2024" containerID="13ba53a1936435bd0bad9ff1979f3e142b0e1305c98c05123dacf5cd2ece9889" exitCode=0 Nov 25 10:34:42 crc kubenswrapper[4821]: I1125 10:34:42.492614 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w56ql" event={"ID":"4b4b52a8-67ad-4c51-86ab-503c75ad2024","Type":"ContainerDied","Data":"13ba53a1936435bd0bad9ff1979f3e142b0e1305c98c05123dacf5cd2ece9889"} Nov 25 10:34:42 crc kubenswrapper[4821]: I1125 10:34:42.503570 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kps6n" Nov 25 10:34:42 crc kubenswrapper[4821]: I1125 10:34:42.547493 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:42 crc kubenswrapper[4821]: E1125 10:34:42.547920 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:43.047905271 +0000 UTC m=+153.584225118 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:42 crc kubenswrapper[4821]: I1125 10:34:42.651873 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:42 crc kubenswrapper[4821]: E1125 10:34:42.653242 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:43.153223292 +0000 UTC m=+153.689543139 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:42 crc kubenswrapper[4821]: I1125 10:34:42.754941 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:42 crc kubenswrapper[4821]: E1125 10:34:42.755328 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:43.255312304 +0000 UTC m=+153.791632151 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:42 crc kubenswrapper[4821]: I1125 10:34:42.856265 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:42 crc kubenswrapper[4821]: E1125 10:34:42.856749 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:43.356732679 +0000 UTC m=+153.893052526 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:42 crc kubenswrapper[4821]: I1125 10:34:42.958302 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:42 crc kubenswrapper[4821]: E1125 10:34:42.958621 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:43.458610754 +0000 UTC m=+153.994930601 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:42 crc kubenswrapper[4821]: I1125 10:34:42.986318 4821 patch_prober.go:28] interesting pod/router-default-5444994796-p92sp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 10:34:42 crc kubenswrapper[4821]: [-]has-synced failed: reason withheld Nov 25 10:34:42 crc kubenswrapper[4821]: [+]process-running ok Nov 25 10:34:42 crc kubenswrapper[4821]: healthz check failed Nov 25 10:34:42 crc kubenswrapper[4821]: I1125 10:34:42.986377 4821 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p92sp" podUID="66798226-d262-4925-82db-2742d96a33fc" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 10:34:43 crc kubenswrapper[4821]: I1125 10:34:43.059206 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:43 crc kubenswrapper[4821]: E1125 10:34:43.059644 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:43.559621 +0000 UTC m=+154.095940847 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:43 crc kubenswrapper[4821]: I1125 10:34:43.059751 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:43 crc kubenswrapper[4821]: E1125 10:34:43.060011 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:43.560003739 +0000 UTC m=+154.096323586 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:43 crc kubenswrapper[4821]: I1125 10:34:43.166826 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:43 crc kubenswrapper[4821]: E1125 10:34:43.167356 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 10:34:43.667337081 +0000 UTC m=+154.203656928 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:43 crc kubenswrapper[4821]: I1125 10:34:43.268183 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:43 crc kubenswrapper[4821]: E1125 10:34:43.268577 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 10:34:43.76856184 +0000 UTC m=+154.304881687 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xnlvh" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 10:34:43 crc kubenswrapper[4821]: I1125 10:34:43.282408 4821 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-11-25T10:34:42.40385842Z","Handler":null,"Name":""} Nov 25 10:34:43 crc kubenswrapper[4821]: I1125 10:34:43.288803 4821 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Nov 25 10:34:43 crc kubenswrapper[4821]: I1125 10:34:43.288830 4821 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Nov 25 10:34:43 crc kubenswrapper[4821]: I1125 10:34:43.368810 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 10:34:43 crc kubenswrapper[4821]: I1125 10:34:43.379464 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 25 10:34:43 crc kubenswrapper[4821]: I1125 10:34:43.470593 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:43 crc kubenswrapper[4821]: I1125 10:34:43.475403 4821 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 25 10:34:43 crc kubenswrapper[4821]: I1125 10:34:43.475876 4821 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:43 crc kubenswrapper[4821]: I1125 10:34:43.499962 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"f5536d76-8475-416d-ac52-649819925b7c","Type":"ContainerStarted","Data":"e960ce457c71896c0920578cf6456ba518d8bbf1c496e16acb51b6cccef99813"} Nov 25 10:34:43 crc kubenswrapper[4821]: I1125 10:34:43.516029 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=2.516010736 podStartE2EDuration="2.516010736s" podCreationTimestamp="2025-11-25 10:34:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:34:43.515998696 +0000 UTC m=+154.052318543" watchObservedRunningTime="2025-11-25 10:34:43.516010736 +0000 UTC m=+154.052330583" Nov 25 10:34:43 crc kubenswrapper[4821]: I1125 10:34:43.518044 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-tsb45" event={"ID":"f97c276b-2b69-49a0-96d1-b1c5171a542e","Type":"ContainerStarted","Data":"8c251de1bda9a59821c8c2dc5b41fcfdfbeb3a1bc8e595cc6e187b00e7aa3414"} Nov 25 10:34:43 crc kubenswrapper[4821]: I1125 10:34:43.544390 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-tsb45" podStartSLOduration=15.544369879 podStartE2EDuration="15.544369879s" podCreationTimestamp="2025-11-25 10:34:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:34:43.542754999 +0000 UTC m=+154.079074856" watchObservedRunningTime="2025-11-25 10:34:43.544369879 +0000 UTC m=+154.080689726" Nov 25 10:34:43 crc kubenswrapper[4821]: I1125 10:34:43.561594 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xnlvh\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:43 crc kubenswrapper[4821]: I1125 10:34:43.696973 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:43 crc kubenswrapper[4821]: I1125 10:34:43.854863 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 10:34:43 crc kubenswrapper[4821]: I1125 10:34:43.983599 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6a84b7fa-c197-4526-9373-dff5e0897293-kube-api-access\") pod \"6a84b7fa-c197-4526-9373-dff5e0897293\" (UID: \"6a84b7fa-c197-4526-9373-dff5e0897293\") " Nov 25 10:34:43 crc kubenswrapper[4821]: I1125 10:34:43.983861 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6a84b7fa-c197-4526-9373-dff5e0897293-kubelet-dir\") pod \"6a84b7fa-c197-4526-9373-dff5e0897293\" (UID: \"6a84b7fa-c197-4526-9373-dff5e0897293\") " Nov 25 10:34:43 crc kubenswrapper[4821]: I1125 10:34:43.984942 4821 patch_prober.go:28] interesting pod/router-default-5444994796-p92sp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 10:34:43 crc kubenswrapper[4821]: [-]has-synced failed: reason withheld Nov 25 10:34:43 crc kubenswrapper[4821]: [+]process-running ok Nov 25 10:34:43 crc kubenswrapper[4821]: healthz check failed Nov 25 10:34:43 crc kubenswrapper[4821]: I1125 10:34:43.984986 4821 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p92sp" podUID="66798226-d262-4925-82db-2742d96a33fc" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 10:34:43 crc kubenswrapper[4821]: I1125 10:34:43.985362 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6a84b7fa-c197-4526-9373-dff5e0897293-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "6a84b7fa-c197-4526-9373-dff5e0897293" (UID: "6a84b7fa-c197-4526-9373-dff5e0897293"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:34:43 crc kubenswrapper[4821]: I1125 10:34:43.991355 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a84b7fa-c197-4526-9373-dff5e0897293-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "6a84b7fa-c197-4526-9373-dff5e0897293" (UID: "6a84b7fa-c197-4526-9373-dff5e0897293"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:34:44 crc kubenswrapper[4821]: I1125 10:34:44.000295 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-xnlvh"] Nov 25 10:34:44 crc kubenswrapper[4821]: W1125 10:34:44.008723 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3fb2355d_5ec2_44f1_8665_a3d73c24f8a6.slice/crio-02e2ec598c0e3b1f1aaf65b4f15ea64649f966c51706c45004ed2b0fed21fef6 WatchSource:0}: Error finding container 02e2ec598c0e3b1f1aaf65b4f15ea64649f966c51706c45004ed2b0fed21fef6: Status 404 returned error can't find the container with id 02e2ec598c0e3b1f1aaf65b4f15ea64649f966c51706c45004ed2b0fed21fef6 Nov 25 10:34:44 crc kubenswrapper[4821]: I1125 10:34:44.085746 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6a84b7fa-c197-4526-9373-dff5e0897293-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 10:34:44 crc kubenswrapper[4821]: I1125 10:34:44.085802 4821 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6a84b7fa-c197-4526-9373-dff5e0897293-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 25 10:34:44 crc kubenswrapper[4821]: I1125 10:34:44.126410 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Nov 25 10:34:44 crc kubenswrapper[4821]: I1125 10:34:44.529014 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" event={"ID":"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6","Type":"ContainerStarted","Data":"dfeb5a5d613f3284ed07af1e11e0bc565beb1e5303175ab7e151caf082203aed"} Nov 25 10:34:44 crc kubenswrapper[4821]: I1125 10:34:44.529544 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:34:44 crc kubenswrapper[4821]: I1125 10:34:44.529588 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" event={"ID":"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6","Type":"ContainerStarted","Data":"02e2ec598c0e3b1f1aaf65b4f15ea64649f966c51706c45004ed2b0fed21fef6"} Nov 25 10:34:44 crc kubenswrapper[4821]: I1125 10:34:44.533567 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"6a84b7fa-c197-4526-9373-dff5e0897293","Type":"ContainerDied","Data":"2f7684584a7dac78105c67c3f42c6098a77555820df5873e8bee726ec3c40534"} Nov 25 10:34:44 crc kubenswrapper[4821]: I1125 10:34:44.533614 4821 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2f7684584a7dac78105c67c3f42c6098a77555820df5873e8bee726ec3c40534" Nov 25 10:34:44 crc kubenswrapper[4821]: I1125 10:34:44.533589 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 10:34:44 crc kubenswrapper[4821]: I1125 10:34:44.536902 4821 generic.go:334] "Generic (PLEG): container finished" podID="f5536d76-8475-416d-ac52-649819925b7c" containerID="e960ce457c71896c0920578cf6456ba518d8bbf1c496e16acb51b6cccef99813" exitCode=0 Nov 25 10:34:44 crc kubenswrapper[4821]: I1125 10:34:44.537458 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"f5536d76-8475-416d-ac52-649819925b7c","Type":"ContainerDied","Data":"e960ce457c71896c0920578cf6456ba518d8bbf1c496e16acb51b6cccef99813"} Nov 25 10:34:44 crc kubenswrapper[4821]: I1125 10:34:44.568990 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" podStartSLOduration=134.568970574 podStartE2EDuration="2m14.568970574s" podCreationTimestamp="2025-11-25 10:32:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:34:44.549927932 +0000 UTC m=+155.086247789" watchObservedRunningTime="2025-11-25 10:34:44.568970574 +0000 UTC m=+155.105290421" Nov 25 10:34:44 crc kubenswrapper[4821]: I1125 10:34:44.983478 4821 patch_prober.go:28] interesting pod/router-default-5444994796-p92sp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 10:34:44 crc kubenswrapper[4821]: [-]has-synced failed: reason withheld Nov 25 10:34:44 crc kubenswrapper[4821]: [+]process-running ok Nov 25 10:34:44 crc kubenswrapper[4821]: healthz check failed Nov 25 10:34:44 crc kubenswrapper[4821]: I1125 10:34:44.983547 4821 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p92sp" podUID="66798226-d262-4925-82db-2742d96a33fc" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 10:34:45 crc kubenswrapper[4821]: I1125 10:34:45.781820 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 10:34:45 crc kubenswrapper[4821]: I1125 10:34:45.908944 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f5536d76-8475-416d-ac52-649819925b7c-kube-api-access\") pod \"f5536d76-8475-416d-ac52-649819925b7c\" (UID: \"f5536d76-8475-416d-ac52-649819925b7c\") " Nov 25 10:34:45 crc kubenswrapper[4821]: I1125 10:34:45.909092 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f5536d76-8475-416d-ac52-649819925b7c-kubelet-dir\") pod \"f5536d76-8475-416d-ac52-649819925b7c\" (UID: \"f5536d76-8475-416d-ac52-649819925b7c\") " Nov 25 10:34:45 crc kubenswrapper[4821]: I1125 10:34:45.909185 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f5536d76-8475-416d-ac52-649819925b7c-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "f5536d76-8475-416d-ac52-649819925b7c" (UID: "f5536d76-8475-416d-ac52-649819925b7c"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:34:45 crc kubenswrapper[4821]: I1125 10:34:45.909482 4821 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f5536d76-8475-416d-ac52-649819925b7c-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 25 10:34:45 crc kubenswrapper[4821]: I1125 10:34:45.913977 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5536d76-8475-416d-ac52-649819925b7c-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "f5536d76-8475-416d-ac52-649819925b7c" (UID: "f5536d76-8475-416d-ac52-649819925b7c"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:34:45 crc kubenswrapper[4821]: I1125 10:34:45.983634 4821 patch_prober.go:28] interesting pod/router-default-5444994796-p92sp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 10:34:45 crc kubenswrapper[4821]: [-]has-synced failed: reason withheld Nov 25 10:34:45 crc kubenswrapper[4821]: [+]process-running ok Nov 25 10:34:45 crc kubenswrapper[4821]: healthz check failed Nov 25 10:34:45 crc kubenswrapper[4821]: I1125 10:34:45.983708 4821 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p92sp" podUID="66798226-d262-4925-82db-2742d96a33fc" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 10:34:46 crc kubenswrapper[4821]: I1125 10:34:46.010219 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f5536d76-8475-416d-ac52-649819925b7c-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 10:34:46 crc kubenswrapper[4821]: I1125 10:34:46.218937 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-jb5sz" Nov 25 10:34:46 crc kubenswrapper[4821]: I1125 10:34:46.223080 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-jb5sz" Nov 25 10:34:46 crc kubenswrapper[4821]: I1125 10:34:46.550141 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 10:34:46 crc kubenswrapper[4821]: I1125 10:34:46.550465 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"f5536d76-8475-416d-ac52-649819925b7c","Type":"ContainerDied","Data":"8fcf85b7d351fef14720f52b69276519b1c8072f11a07ae296948b556c4f8263"} Nov 25 10:34:46 crc kubenswrapper[4821]: I1125 10:34:46.550496 4821 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8fcf85b7d351fef14720f52b69276519b1c8072f11a07ae296948b556c4f8263" Nov 25 10:34:46 crc kubenswrapper[4821]: I1125 10:34:46.785876 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-rx6sh" Nov 25 10:34:46 crc kubenswrapper[4821]: I1125 10:34:46.983037 4821 patch_prober.go:28] interesting pod/router-default-5444994796-p92sp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 10:34:46 crc kubenswrapper[4821]: [-]has-synced failed: reason withheld Nov 25 10:34:46 crc kubenswrapper[4821]: [+]process-running ok Nov 25 10:34:46 crc kubenswrapper[4821]: healthz check failed Nov 25 10:34:46 crc kubenswrapper[4821]: I1125 10:34:46.983092 4821 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p92sp" podUID="66798226-d262-4925-82db-2742d96a33fc" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 10:34:47 crc kubenswrapper[4821]: I1125 10:34:47.984517 4821 patch_prober.go:28] interesting pod/router-default-5444994796-p92sp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 10:34:47 crc kubenswrapper[4821]: [-]has-synced failed: reason withheld Nov 25 10:34:47 crc kubenswrapper[4821]: [+]process-running ok Nov 25 10:34:47 crc kubenswrapper[4821]: healthz check failed Nov 25 10:34:47 crc kubenswrapper[4821]: I1125 10:34:47.985774 4821 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p92sp" podUID="66798226-d262-4925-82db-2742d96a33fc" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 10:34:48 crc kubenswrapper[4821]: I1125 10:34:48.985146 4821 patch_prober.go:28] interesting pod/router-default-5444994796-p92sp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 10:34:48 crc kubenswrapper[4821]: [-]has-synced failed: reason withheld Nov 25 10:34:48 crc kubenswrapper[4821]: [+]process-running ok Nov 25 10:34:48 crc kubenswrapper[4821]: healthz check failed Nov 25 10:34:48 crc kubenswrapper[4821]: I1125 10:34:48.985227 4821 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p92sp" podUID="66798226-d262-4925-82db-2742d96a33fc" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 10:34:49 crc kubenswrapper[4821]: I1125 10:34:49.983841 4821 patch_prober.go:28] interesting pod/router-default-5444994796-p92sp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 10:34:49 crc kubenswrapper[4821]: [-]has-synced failed: reason withheld Nov 25 10:34:49 crc kubenswrapper[4821]: [+]process-running ok Nov 25 10:34:49 crc kubenswrapper[4821]: healthz check failed Nov 25 10:34:49 crc kubenswrapper[4821]: I1125 10:34:49.984238 4821 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p92sp" podUID="66798226-d262-4925-82db-2742d96a33fc" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 10:34:50 crc kubenswrapper[4821]: I1125 10:34:50.126003 4821 patch_prober.go:28] interesting pod/console-f9d7485db-hf49d container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.8:8443/health\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Nov 25 10:34:50 crc kubenswrapper[4821]: I1125 10:34:50.126946 4821 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-hf49d" podUID="88725123-3da9-4127-9506-80c90c9accd5" containerName="console" probeResult="failure" output="Get \"https://10.217.0.8:8443/health\": dial tcp 10.217.0.8:8443: connect: connection refused" Nov 25 10:34:50 crc kubenswrapper[4821]: I1125 10:34:50.983672 4821 patch_prober.go:28] interesting pod/router-default-5444994796-p92sp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 10:34:50 crc kubenswrapper[4821]: [-]has-synced failed: reason withheld Nov 25 10:34:50 crc kubenswrapper[4821]: [+]process-running ok Nov 25 10:34:50 crc kubenswrapper[4821]: healthz check failed Nov 25 10:34:50 crc kubenswrapper[4821]: I1125 10:34:50.983776 4821 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p92sp" podUID="66798226-d262-4925-82db-2742d96a33fc" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 10:34:51 crc kubenswrapper[4821]: I1125 10:34:51.124048 4821 patch_prober.go:28] interesting pod/downloads-7954f5f757-hjvt9 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Nov 25 10:34:51 crc kubenswrapper[4821]: I1125 10:34:51.124096 4821 patch_prober.go:28] interesting pod/downloads-7954f5f757-hjvt9 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Nov 25 10:34:51 crc kubenswrapper[4821]: I1125 10:34:51.124147 4821 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-hjvt9" podUID="f954533e-e8a0-4899-be72-241760feae95" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Nov 25 10:34:51 crc kubenswrapper[4821]: I1125 10:34:51.124100 4821 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-hjvt9" podUID="f954533e-e8a0-4899-be72-241760feae95" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Nov 25 10:34:51 crc kubenswrapper[4821]: I1125 10:34:51.983889 4821 patch_prober.go:28] interesting pod/router-default-5444994796-p92sp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 10:34:51 crc kubenswrapper[4821]: [-]has-synced failed: reason withheld Nov 25 10:34:51 crc kubenswrapper[4821]: [+]process-running ok Nov 25 10:34:51 crc kubenswrapper[4821]: healthz check failed Nov 25 10:34:51 crc kubenswrapper[4821]: I1125 10:34:51.984455 4821 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p92sp" podUID="66798226-d262-4925-82db-2742d96a33fc" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 10:34:52 crc kubenswrapper[4821]: I1125 10:34:52.523316 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1f12cef3-7e40-4f61-836e-23cd0e578e71-metrics-certs\") pod \"network-metrics-daemon-d4msn\" (UID: \"1f12cef3-7e40-4f61-836e-23cd0e578e71\") " pod="openshift-multus/network-metrics-daemon-d4msn" Nov 25 10:34:52 crc kubenswrapper[4821]: I1125 10:34:52.531190 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1f12cef3-7e40-4f61-836e-23cd0e578e71-metrics-certs\") pod \"network-metrics-daemon-d4msn\" (UID: \"1f12cef3-7e40-4f61-836e-23cd0e578e71\") " pod="openshift-multus/network-metrics-daemon-d4msn" Nov 25 10:34:52 crc kubenswrapper[4821]: I1125 10:34:52.826026 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-d4msn" Nov 25 10:34:52 crc kubenswrapper[4821]: I1125 10:34:52.983542 4821 patch_prober.go:28] interesting pod/router-default-5444994796-p92sp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 10:34:52 crc kubenswrapper[4821]: [+]has-synced ok Nov 25 10:34:52 crc kubenswrapper[4821]: [+]process-running ok Nov 25 10:34:52 crc kubenswrapper[4821]: healthz check failed Nov 25 10:34:52 crc kubenswrapper[4821]: I1125 10:34:52.983599 4821 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p92sp" podUID="66798226-d262-4925-82db-2742d96a33fc" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 10:34:53 crc kubenswrapper[4821]: I1125 10:34:53.984180 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-p92sp" Nov 25 10:34:53 crc kubenswrapper[4821]: I1125 10:34:53.986587 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-p92sp" Nov 25 10:35:00 crc kubenswrapper[4821]: I1125 10:35:00.160545 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-hf49d" Nov 25 10:35:00 crc kubenswrapper[4821]: I1125 10:35:00.165706 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-hf49d" Nov 25 10:35:01 crc kubenswrapper[4821]: I1125 10:35:01.125731 4821 patch_prober.go:28] interesting pod/downloads-7954f5f757-hjvt9 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Nov 25 10:35:01 crc kubenswrapper[4821]: I1125 10:35:01.125784 4821 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-hjvt9" podUID="f954533e-e8a0-4899-be72-241760feae95" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Nov 25 10:35:01 crc kubenswrapper[4821]: I1125 10:35:01.125823 4821 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-console/downloads-7954f5f757-hjvt9" Nov 25 10:35:01 crc kubenswrapper[4821]: I1125 10:35:01.125870 4821 patch_prober.go:28] interesting pod/downloads-7954f5f757-hjvt9 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Nov 25 10:35:01 crc kubenswrapper[4821]: I1125 10:35:01.125895 4821 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-hjvt9" podUID="f954533e-e8a0-4899-be72-241760feae95" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Nov 25 10:35:01 crc kubenswrapper[4821]: I1125 10:35:01.126082 4821 patch_prober.go:28] interesting pod/downloads-7954f5f757-hjvt9 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Nov 25 10:35:01 crc kubenswrapper[4821]: I1125 10:35:01.126099 4821 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-hjvt9" podUID="f954533e-e8a0-4899-be72-241760feae95" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Nov 25 10:35:01 crc kubenswrapper[4821]: I1125 10:35:01.126303 4821 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="download-server" containerStatusID={"Type":"cri-o","ID":"85476df01cdd233aa043cdab8538c2fa562637811fc58da7df8de836e9316399"} pod="openshift-console/downloads-7954f5f757-hjvt9" containerMessage="Container download-server failed liveness probe, will be restarted" Nov 25 10:35:01 crc kubenswrapper[4821]: I1125 10:35:01.126377 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/downloads-7954f5f757-hjvt9" podUID="f954533e-e8a0-4899-be72-241760feae95" containerName="download-server" containerID="cri-o://85476df01cdd233aa043cdab8538c2fa562637811fc58da7df8de836e9316399" gracePeriod=2 Nov 25 10:35:02 crc kubenswrapper[4821]: I1125 10:35:02.640480 4821 generic.go:334] "Generic (PLEG): container finished" podID="f954533e-e8a0-4899-be72-241760feae95" containerID="85476df01cdd233aa043cdab8538c2fa562637811fc58da7df8de836e9316399" exitCode=0 Nov 25 10:35:02 crc kubenswrapper[4821]: I1125 10:35:02.640795 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-hjvt9" event={"ID":"f954533e-e8a0-4899-be72-241760feae95","Type":"ContainerDied","Data":"85476df01cdd233aa043cdab8538c2fa562637811fc58da7df8de836e9316399"} Nov 25 10:35:03 crc kubenswrapper[4821]: I1125 10:35:03.702597 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:35:04 crc kubenswrapper[4821]: I1125 10:35:04.714619 4821 patch_prober.go:28] interesting pod/machine-config-daemon-2krbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:35:04 crc kubenswrapper[4821]: I1125 10:35:04.714967 4821 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:35:11 crc kubenswrapper[4821]: I1125 10:35:11.125600 4821 patch_prober.go:28] interesting pod/downloads-7954f5f757-hjvt9 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Nov 25 10:35:11 crc kubenswrapper[4821]: I1125 10:35:11.126135 4821 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-hjvt9" podUID="f954533e-e8a0-4899-be72-241760feae95" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Nov 25 10:35:11 crc kubenswrapper[4821]: I1125 10:35:11.134643 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-f2tjh" Nov 25 10:35:17 crc kubenswrapper[4821]: I1125 10:35:17.705517 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 10:35:17 crc kubenswrapper[4821]: E1125 10:35:17.981633 4821 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Nov 25 10:35:17 crc kubenswrapper[4821]: E1125 10:35:17.982153 4821 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-f6dp4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-r5nmb_openshift-marketplace(e1391070-3aad-47c3-a2f1-0c1697b014bf): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 25 10:35:17 crc kubenswrapper[4821]: E1125 10:35:17.983471 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-r5nmb" podUID="e1391070-3aad-47c3-a2f1-0c1697b014bf" Nov 25 10:35:21 crc kubenswrapper[4821]: I1125 10:35:21.124352 4821 patch_prober.go:28] interesting pod/downloads-7954f5f757-hjvt9 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Nov 25 10:35:21 crc kubenswrapper[4821]: I1125 10:35:21.124413 4821 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-hjvt9" podUID="f954533e-e8a0-4899-be72-241760feae95" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Nov 25 10:35:27 crc kubenswrapper[4821]: E1125 10:35:27.207588 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-r5nmb" podUID="e1391070-3aad-47c3-a2f1-0c1697b014bf" Nov 25 10:35:27 crc kubenswrapper[4821]: E1125 10:35:27.361106 4821 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Nov 25 10:35:27 crc kubenswrapper[4821]: E1125 10:35:27.361302 4821 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nstfp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-w56ql_openshift-marketplace(4b4b52a8-67ad-4c51-86ab-503c75ad2024): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 25 10:35:27 crc kubenswrapper[4821]: E1125 10:35:27.362472 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-w56ql" podUID="4b4b52a8-67ad-4c51-86ab-503c75ad2024" Nov 25 10:35:28 crc kubenswrapper[4821]: E1125 10:35:28.325225 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-w56ql" podUID="4b4b52a8-67ad-4c51-86ab-503c75ad2024" Nov 25 10:35:28 crc kubenswrapper[4821]: E1125 10:35:28.380690 4821 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Nov 25 10:35:28 crc kubenswrapper[4821]: E1125 10:35:28.381075 4821 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vlpdz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-2kwcp_openshift-marketplace(927e7d40-852e-4807-bb22-d4f7c6e9d2ca): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 25 10:35:28 crc kubenswrapper[4821]: E1125 10:35:28.382246 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-2kwcp" podUID="927e7d40-852e-4807-bb22-d4f7c6e9d2ca" Nov 25 10:35:28 crc kubenswrapper[4821]: E1125 10:35:28.479558 4821 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Nov 25 10:35:28 crc kubenswrapper[4821]: E1125 10:35:28.480010 4821 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gfw7p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-8qpmf_openshift-marketplace(d8becff5-8017-4073-89ae-c17cc4c32922): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 25 10:35:28 crc kubenswrapper[4821]: E1125 10:35:28.481130 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-8qpmf" podUID="d8becff5-8017-4073-89ae-c17cc4c32922" Nov 25 10:35:28 crc kubenswrapper[4821]: E1125 10:35:28.533130 4821 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Nov 25 10:35:28 crc kubenswrapper[4821]: E1125 10:35:28.533397 4821 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xn8qr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-ps9cl_openshift-marketplace(645993fd-bdce-435e-a600-34bf1ff5b38b): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 25 10:35:28 crc kubenswrapper[4821]: E1125 10:35:28.534582 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-ps9cl" podUID="645993fd-bdce-435e-a600-34bf1ff5b38b" Nov 25 10:35:28 crc kubenswrapper[4821]: E1125 10:35:28.582963 4821 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Nov 25 10:35:28 crc kubenswrapper[4821]: E1125 10:35:28.583093 4821 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-njv9h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-cgdxc_openshift-marketplace(b6e72ee2-50e4-4643-aee6-8be89d2e4589): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 25 10:35:28 crc kubenswrapper[4821]: E1125 10:35:28.584233 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-cgdxc" podUID="b6e72ee2-50e4-4643-aee6-8be89d2e4589" Nov 25 10:35:28 crc kubenswrapper[4821]: E1125 10:35:28.661430 4821 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Nov 25 10:35:28 crc kubenswrapper[4821]: E1125 10:35:28.661599 4821 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6gpf2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-87rxm_openshift-marketplace(38068a8c-fd52-492c-9e1e-6425830bd097): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 25 10:35:28 crc kubenswrapper[4821]: E1125 10:35:28.662828 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-87rxm" podUID="38068a8c-fd52-492c-9e1e-6425830bd097" Nov 25 10:35:28 crc kubenswrapper[4821]: I1125 10:35:28.755715 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-d4msn"] Nov 25 10:35:28 crc kubenswrapper[4821]: W1125 10:35:28.765839 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1f12cef3_7e40_4f61_836e_23cd0e578e71.slice/crio-019967a760ae408be812dcdd2807c32f8d4a10419dc50e34fe10789272a6395d WatchSource:0}: Error finding container 019967a760ae408be812dcdd2807c32f8d4a10419dc50e34fe10789272a6395d: Status 404 returned error can't find the container with id 019967a760ae408be812dcdd2807c32f8d4a10419dc50e34fe10789272a6395d Nov 25 10:35:28 crc kubenswrapper[4821]: I1125 10:35:28.769818 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zrl8b" event={"ID":"d1d6804a-27b6-4ef8-90ce-32606c93869a","Type":"ContainerStarted","Data":"da0688502ecc52dd6fc2b2bcb4dedc3ab4e27c36774a7eed2696ea4a0dca81e9"} Nov 25 10:35:28 crc kubenswrapper[4821]: I1125 10:35:28.773224 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-hjvt9" event={"ID":"f954533e-e8a0-4899-be72-241760feae95","Type":"ContainerStarted","Data":"57d1e91dac0ff4c6b056876e811dadc88db65008253f0e7f92a1790ada1e049d"} Nov 25 10:35:28 crc kubenswrapper[4821]: I1125 10:35:28.773769 4821 patch_prober.go:28] interesting pod/downloads-7954f5f757-hjvt9 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Nov 25 10:35:28 crc kubenswrapper[4821]: I1125 10:35:28.783365 4821 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-hjvt9" podUID="f954533e-e8a0-4899-be72-241760feae95" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Nov 25 10:35:28 crc kubenswrapper[4821]: E1125 10:35:28.811068 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-ps9cl" podUID="645993fd-bdce-435e-a600-34bf1ff5b38b" Nov 25 10:35:28 crc kubenswrapper[4821]: E1125 10:35:28.811380 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-2kwcp" podUID="927e7d40-852e-4807-bb22-d4f7c6e9d2ca" Nov 25 10:35:28 crc kubenswrapper[4821]: E1125 10:35:28.811486 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-87rxm" podUID="38068a8c-fd52-492c-9e1e-6425830bd097" Nov 25 10:35:28 crc kubenswrapper[4821]: E1125 10:35:28.813931 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-cgdxc" podUID="b6e72ee2-50e4-4643-aee6-8be89d2e4589" Nov 25 10:35:29 crc kubenswrapper[4821]: I1125 10:35:29.787873 4821 generic.go:334] "Generic (PLEG): container finished" podID="d1d6804a-27b6-4ef8-90ce-32606c93869a" containerID="da0688502ecc52dd6fc2b2bcb4dedc3ab4e27c36774a7eed2696ea4a0dca81e9" exitCode=0 Nov 25 10:35:29 crc kubenswrapper[4821]: I1125 10:35:29.788353 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zrl8b" event={"ID":"d1d6804a-27b6-4ef8-90ce-32606c93869a","Type":"ContainerDied","Data":"da0688502ecc52dd6fc2b2bcb4dedc3ab4e27c36774a7eed2696ea4a0dca81e9"} Nov 25 10:35:29 crc kubenswrapper[4821]: I1125 10:35:29.793079 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-d4msn" event={"ID":"1f12cef3-7e40-4f61-836e-23cd0e578e71","Type":"ContainerStarted","Data":"8898bc1b71a46d5658785e3fa0ccf71de4f2f5b6ac2b5b98a5a01f6356785f41"} Nov 25 10:35:29 crc kubenswrapper[4821]: I1125 10:35:29.793124 4821 patch_prober.go:28] interesting pod/downloads-7954f5f757-hjvt9 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Nov 25 10:35:29 crc kubenswrapper[4821]: I1125 10:35:29.793188 4821 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-hjvt9" podUID="f954533e-e8a0-4899-be72-241760feae95" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Nov 25 10:35:29 crc kubenswrapper[4821]: I1125 10:35:29.793133 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-hjvt9" Nov 25 10:35:29 crc kubenswrapper[4821]: I1125 10:35:29.793242 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-d4msn" event={"ID":"1f12cef3-7e40-4f61-836e-23cd0e578e71","Type":"ContainerStarted","Data":"7806954974f63382c2cc784fd584aafb608862e389f1c72fe931234f4992f121"} Nov 25 10:35:29 crc kubenswrapper[4821]: I1125 10:35:29.793256 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-d4msn" event={"ID":"1f12cef3-7e40-4f61-836e-23cd0e578e71","Type":"ContainerStarted","Data":"019967a760ae408be812dcdd2807c32f8d4a10419dc50e34fe10789272a6395d"} Nov 25 10:35:29 crc kubenswrapper[4821]: I1125 10:35:29.819715 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-d4msn" podStartSLOduration=179.819697241 podStartE2EDuration="2m59.819697241s" podCreationTimestamp="2025-11-25 10:32:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:35:29.818515969 +0000 UTC m=+200.354835816" watchObservedRunningTime="2025-11-25 10:35:29.819697241 +0000 UTC m=+200.356017098" Nov 25 10:35:30 crc kubenswrapper[4821]: I1125 10:35:30.797629 4821 patch_prober.go:28] interesting pod/downloads-7954f5f757-hjvt9 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Nov 25 10:35:30 crc kubenswrapper[4821]: I1125 10:35:30.798212 4821 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-hjvt9" podUID="f954533e-e8a0-4899-be72-241760feae95" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Nov 25 10:35:31 crc kubenswrapper[4821]: I1125 10:35:31.124643 4821 patch_prober.go:28] interesting pod/downloads-7954f5f757-hjvt9 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Nov 25 10:35:31 crc kubenswrapper[4821]: I1125 10:35:31.124684 4821 patch_prober.go:28] interesting pod/downloads-7954f5f757-hjvt9 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Nov 25 10:35:31 crc kubenswrapper[4821]: I1125 10:35:31.124953 4821 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-hjvt9" podUID="f954533e-e8a0-4899-be72-241760feae95" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Nov 25 10:35:31 crc kubenswrapper[4821]: I1125 10:35:31.125019 4821 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-hjvt9" podUID="f954533e-e8a0-4899-be72-241760feae95" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Nov 25 10:35:31 crc kubenswrapper[4821]: I1125 10:35:31.803634 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zrl8b" event={"ID":"d1d6804a-27b6-4ef8-90ce-32606c93869a","Type":"ContainerStarted","Data":"f2aba3affd94c7e2bdb62665d9d38d71f158474e556cb47e4248b9556a96b832"} Nov 25 10:35:31 crc kubenswrapper[4821]: I1125 10:35:31.821280 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zrl8b" podStartSLOduration=4.687707287 podStartE2EDuration="54.82125969s" podCreationTimestamp="2025-11-25 10:34:37 +0000 UTC" firstStartedPulling="2025-11-25 10:34:40.420990854 +0000 UTC m=+150.957310701" lastFinishedPulling="2025-11-25 10:35:30.554543257 +0000 UTC m=+201.090863104" observedRunningTime="2025-11-25 10:35:31.818248368 +0000 UTC m=+202.354568235" watchObservedRunningTime="2025-11-25 10:35:31.82125969 +0000 UTC m=+202.357579537" Nov 25 10:35:34 crc kubenswrapper[4821]: I1125 10:35:34.715116 4821 patch_prober.go:28] interesting pod/machine-config-daemon-2krbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:35:34 crc kubenswrapper[4821]: I1125 10:35:34.715792 4821 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:35:34 crc kubenswrapper[4821]: I1125 10:35:34.715848 4821 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" Nov 25 10:35:34 crc kubenswrapper[4821]: I1125 10:35:34.716493 4821 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9ba68bd093be94d0d40f3851638db5d26bdfc52d175b1ac78afb91dda7affcfe"} pod="openshift-machine-config-operator/machine-config-daemon-2krbf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 10:35:34 crc kubenswrapper[4821]: I1125 10:35:34.716573 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerName="machine-config-daemon" containerID="cri-o://9ba68bd093be94d0d40f3851638db5d26bdfc52d175b1ac78afb91dda7affcfe" gracePeriod=600 Nov 25 10:35:35 crc kubenswrapper[4821]: I1125 10:35:35.823101 4821 generic.go:334] "Generic (PLEG): container finished" podID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerID="9ba68bd093be94d0d40f3851638db5d26bdfc52d175b1ac78afb91dda7affcfe" exitCode=0 Nov 25 10:35:35 crc kubenswrapper[4821]: I1125 10:35:35.823220 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" event={"ID":"5f948b87-ac86-4de6-ad64-c2ef947f84d4","Type":"ContainerDied","Data":"9ba68bd093be94d0d40f3851638db5d26bdfc52d175b1ac78afb91dda7affcfe"} Nov 25 10:35:37 crc kubenswrapper[4821]: I1125 10:35:37.834615 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" event={"ID":"5f948b87-ac86-4de6-ad64-c2ef947f84d4","Type":"ContainerStarted","Data":"131eaad8c97dd44aa2f047258a3256b958757d6189e9d69d307bd69c27cbcd08"} Nov 25 10:35:37 crc kubenswrapper[4821]: I1125 10:35:37.922499 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zrl8b" Nov 25 10:35:37 crc kubenswrapper[4821]: I1125 10:35:37.922542 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zrl8b" Nov 25 10:35:38 crc kubenswrapper[4821]: I1125 10:35:38.760240 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zrl8b" Nov 25 10:35:38 crc kubenswrapper[4821]: I1125 10:35:38.874608 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zrl8b" Nov 25 10:35:41 crc kubenswrapper[4821]: I1125 10:35:41.139708 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-hjvt9" Nov 25 10:35:43 crc kubenswrapper[4821]: I1125 10:35:43.867186 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w56ql" event={"ID":"4b4b52a8-67ad-4c51-86ab-503c75ad2024","Type":"ContainerStarted","Data":"9d16b72aaa2431617a2e647a9a8fb1561bf3d618b0823b487f339531f5284952"} Nov 25 10:35:43 crc kubenswrapper[4821]: I1125 10:35:43.869722 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cgdxc" event={"ID":"b6e72ee2-50e4-4643-aee6-8be89d2e4589","Type":"ContainerStarted","Data":"7b46349b24b5354cf00d29385827cea9614ea973f6654937bc0aabb6c2b529f2"} Nov 25 10:35:43 crc kubenswrapper[4821]: I1125 10:35:43.871284 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-87rxm" event={"ID":"38068a8c-fd52-492c-9e1e-6425830bd097","Type":"ContainerStarted","Data":"205a1bb3a9091c6fdd3137e986e9d6b47b7a4b2fd20e0c92aeb31fccc6aea4a3"} Nov 25 10:35:44 crc kubenswrapper[4821]: I1125 10:35:44.878243 4821 generic.go:334] "Generic (PLEG): container finished" podID="4b4b52a8-67ad-4c51-86ab-503c75ad2024" containerID="9d16b72aaa2431617a2e647a9a8fb1561bf3d618b0823b487f339531f5284952" exitCode=0 Nov 25 10:35:44 crc kubenswrapper[4821]: I1125 10:35:44.878310 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w56ql" event={"ID":"4b4b52a8-67ad-4c51-86ab-503c75ad2024","Type":"ContainerDied","Data":"9d16b72aaa2431617a2e647a9a8fb1561bf3d618b0823b487f339531f5284952"} Nov 25 10:35:44 crc kubenswrapper[4821]: I1125 10:35:44.880313 4821 generic.go:334] "Generic (PLEG): container finished" podID="b6e72ee2-50e4-4643-aee6-8be89d2e4589" containerID="7b46349b24b5354cf00d29385827cea9614ea973f6654937bc0aabb6c2b529f2" exitCode=0 Nov 25 10:35:44 crc kubenswrapper[4821]: I1125 10:35:44.880358 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cgdxc" event={"ID":"b6e72ee2-50e4-4643-aee6-8be89d2e4589","Type":"ContainerDied","Data":"7b46349b24b5354cf00d29385827cea9614ea973f6654937bc0aabb6c2b529f2"} Nov 25 10:35:44 crc kubenswrapper[4821]: I1125 10:35:44.882525 4821 generic.go:334] "Generic (PLEG): container finished" podID="38068a8c-fd52-492c-9e1e-6425830bd097" containerID="205a1bb3a9091c6fdd3137e986e9d6b47b7a4b2fd20e0c92aeb31fccc6aea4a3" exitCode=0 Nov 25 10:35:44 crc kubenswrapper[4821]: I1125 10:35:44.882562 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-87rxm" event={"ID":"38068a8c-fd52-492c-9e1e-6425830bd097","Type":"ContainerDied","Data":"205a1bb3a9091c6fdd3137e986e9d6b47b7a4b2fd20e0c92aeb31fccc6aea4a3"} Nov 25 10:36:11 crc kubenswrapper[4821]: I1125 10:36:11.066779 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-mbn4k"] Nov 25 10:36:11 crc kubenswrapper[4821]: I1125 10:36:11.144765 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-87rxm" event={"ID":"38068a8c-fd52-492c-9e1e-6425830bd097","Type":"ContainerStarted","Data":"c69c3a931a80bb0669ee855880d4f2c246003cbfae84bcf0a7a74900e2a9c840"} Nov 25 10:36:11 crc kubenswrapper[4821]: I1125 10:36:11.147144 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2kwcp" event={"ID":"927e7d40-852e-4807-bb22-d4f7c6e9d2ca","Type":"ContainerStarted","Data":"39e4fa98d1d04e9e87bcd354abd8aeaaecdf77816239d00c555fdc9aee8c652c"} Nov 25 10:36:11 crc kubenswrapper[4821]: I1125 10:36:11.149347 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8qpmf" event={"ID":"d8becff5-8017-4073-89ae-c17cc4c32922","Type":"ContainerStarted","Data":"7588e12b27ca0da7eeb1409085d529ba99aeb108a9c5a525074375a0ed04e76c"} Nov 25 10:36:11 crc kubenswrapper[4821]: I1125 10:36:11.151286 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r5nmb" event={"ID":"e1391070-3aad-47c3-a2f1-0c1697b014bf","Type":"ContainerStarted","Data":"bea8c804c37d876eeb33513f8d24ec7d2523e8a305c58c337aca7e88247d7ea9"} Nov 25 10:36:12 crc kubenswrapper[4821]: I1125 10:36:12.158362 4821 generic.go:334] "Generic (PLEG): container finished" podID="927e7d40-852e-4807-bb22-d4f7c6e9d2ca" containerID="39e4fa98d1d04e9e87bcd354abd8aeaaecdf77816239d00c555fdc9aee8c652c" exitCode=0 Nov 25 10:36:12 crc kubenswrapper[4821]: I1125 10:36:12.158529 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2kwcp" event={"ID":"927e7d40-852e-4807-bb22-d4f7c6e9d2ca","Type":"ContainerDied","Data":"39e4fa98d1d04e9e87bcd354abd8aeaaecdf77816239d00c555fdc9aee8c652c"} Nov 25 10:36:12 crc kubenswrapper[4821]: I1125 10:36:12.164604 4821 generic.go:334] "Generic (PLEG): container finished" podID="d8becff5-8017-4073-89ae-c17cc4c32922" containerID="7588e12b27ca0da7eeb1409085d529ba99aeb108a9c5a525074375a0ed04e76c" exitCode=0 Nov 25 10:36:12 crc kubenswrapper[4821]: I1125 10:36:12.164653 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8qpmf" event={"ID":"d8becff5-8017-4073-89ae-c17cc4c32922","Type":"ContainerDied","Data":"7588e12b27ca0da7eeb1409085d529ba99aeb108a9c5a525074375a0ed04e76c"} Nov 25 10:36:12 crc kubenswrapper[4821]: I1125 10:36:12.166759 4821 generic.go:334] "Generic (PLEG): container finished" podID="e1391070-3aad-47c3-a2f1-0c1697b014bf" containerID="bea8c804c37d876eeb33513f8d24ec7d2523e8a305c58c337aca7e88247d7ea9" exitCode=0 Nov 25 10:36:12 crc kubenswrapper[4821]: I1125 10:36:12.167559 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r5nmb" event={"ID":"e1391070-3aad-47c3-a2f1-0c1697b014bf","Type":"ContainerDied","Data":"bea8c804c37d876eeb33513f8d24ec7d2523e8a305c58c337aca7e88247d7ea9"} Nov 25 10:36:12 crc kubenswrapper[4821]: I1125 10:36:12.223144 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-87rxm" podStartSLOduration=6.093413823 podStartE2EDuration="1m32.223125696s" podCreationTimestamp="2025-11-25 10:34:40 +0000 UTC" firstStartedPulling="2025-11-25 10:34:42.489147274 +0000 UTC m=+153.025467121" lastFinishedPulling="2025-11-25 10:36:08.618859147 +0000 UTC m=+239.155178994" observedRunningTime="2025-11-25 10:36:12.219489278 +0000 UTC m=+242.755809125" watchObservedRunningTime="2025-11-25 10:36:12.223125696 +0000 UTC m=+242.759445553" Nov 25 10:36:13 crc kubenswrapper[4821]: I1125 10:36:13.173761 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w56ql" event={"ID":"4b4b52a8-67ad-4c51-86ab-503c75ad2024","Type":"ContainerStarted","Data":"51fe6850e14ffc694a9db8f49ec8649f913025a36e30f3269bd3e251d416726a"} Nov 25 10:36:13 crc kubenswrapper[4821]: I1125 10:36:13.745653 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-w56ql" podStartSLOduration=4.228922452 podStartE2EDuration="1m33.745633352s" podCreationTimestamp="2025-11-25 10:34:40 +0000 UTC" firstStartedPulling="2025-11-25 10:34:42.49542009 +0000 UTC m=+153.031739937" lastFinishedPulling="2025-11-25 10:36:12.01213099 +0000 UTC m=+242.548450837" observedRunningTime="2025-11-25 10:36:13.192297483 +0000 UTC m=+243.728617330" watchObservedRunningTime="2025-11-25 10:36:13.745633352 +0000 UTC m=+244.281953199" Nov 25 10:36:13 crc kubenswrapper[4821]: I1125 10:36:13.747242 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ps9cl"] Nov 25 10:36:13 crc kubenswrapper[4821]: I1125 10:36:13.761642 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-r5nmb"] Nov 25 10:36:13 crc kubenswrapper[4821]: I1125 10:36:13.771791 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cgdxc"] Nov 25 10:36:13 crc kubenswrapper[4821]: I1125 10:36:13.785964 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zrl8b"] Nov 25 10:36:13 crc kubenswrapper[4821]: I1125 10:36:13.786415 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zrl8b" podUID="d1d6804a-27b6-4ef8-90ce-32606c93869a" containerName="registry-server" containerID="cri-o://f2aba3affd94c7e2bdb62665d9d38d71f158474e556cb47e4248b9556a96b832" gracePeriod=30 Nov 25 10:36:13 crc kubenswrapper[4821]: I1125 10:36:13.805672 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-dxf4x"] Nov 25 10:36:13 crc kubenswrapper[4821]: I1125 10:36:13.807823 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-dxf4x" podUID="590f742c-e0d3-4394-8ef3-08bde5d57d01" containerName="marketplace-operator" containerID="cri-o://41b4dac6291ebdf0ed0a316060e54df84fbc3b2fc75af2abd1e432330754933d" gracePeriod=30 Nov 25 10:36:13 crc kubenswrapper[4821]: I1125 10:36:13.810731 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2kwcp"] Nov 25 10:36:13 crc kubenswrapper[4821]: I1125 10:36:13.819655 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8qpmf"] Nov 25 10:36:13 crc kubenswrapper[4821]: I1125 10:36:13.827184 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-tdctv"] Nov 25 10:36:13 crc kubenswrapper[4821]: E1125 10:36:13.827385 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a84b7fa-c197-4526-9373-dff5e0897293" containerName="pruner" Nov 25 10:36:13 crc kubenswrapper[4821]: I1125 10:36:13.827397 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a84b7fa-c197-4526-9373-dff5e0897293" containerName="pruner" Nov 25 10:36:13 crc kubenswrapper[4821]: E1125 10:36:13.827411 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5536d76-8475-416d-ac52-649819925b7c" containerName="pruner" Nov 25 10:36:13 crc kubenswrapper[4821]: I1125 10:36:13.827417 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5536d76-8475-416d-ac52-649819925b7c" containerName="pruner" Nov 25 10:36:13 crc kubenswrapper[4821]: I1125 10:36:13.827512 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5536d76-8475-416d-ac52-649819925b7c" containerName="pruner" Nov 25 10:36:13 crc kubenswrapper[4821]: I1125 10:36:13.827521 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a84b7fa-c197-4526-9373-dff5e0897293" containerName="pruner" Nov 25 10:36:13 crc kubenswrapper[4821]: I1125 10:36:13.827865 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-tdctv" Nov 25 10:36:13 crc kubenswrapper[4821]: I1125 10:36:13.832025 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-87rxm"] Nov 25 10:36:13 crc kubenswrapper[4821]: I1125 10:36:13.832249 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-87rxm" podUID="38068a8c-fd52-492c-9e1e-6425830bd097" containerName="registry-server" containerID="cri-o://c69c3a931a80bb0669ee855880d4f2c246003cbfae84bcf0a7a74900e2a9c840" gracePeriod=30 Nov 25 10:36:13 crc kubenswrapper[4821]: I1125 10:36:13.840771 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-w56ql"] Nov 25 10:36:13 crc kubenswrapper[4821]: I1125 10:36:13.844873 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-tdctv"] Nov 25 10:36:13 crc kubenswrapper[4821]: I1125 10:36:13.990213 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/05106897-6fd0-462c-9d26-a832f1385e04-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-tdctv\" (UID: \"05106897-6fd0-462c-9d26-a832f1385e04\") " pod="openshift-marketplace/marketplace-operator-79b997595-tdctv" Nov 25 10:36:13 crc kubenswrapper[4821]: I1125 10:36:13.990278 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xb9l4\" (UniqueName: \"kubernetes.io/projected/05106897-6fd0-462c-9d26-a832f1385e04-kube-api-access-xb9l4\") pod \"marketplace-operator-79b997595-tdctv\" (UID: \"05106897-6fd0-462c-9d26-a832f1385e04\") " pod="openshift-marketplace/marketplace-operator-79b997595-tdctv" Nov 25 10:36:13 crc kubenswrapper[4821]: I1125 10:36:13.990314 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/05106897-6fd0-462c-9d26-a832f1385e04-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-tdctv\" (UID: \"05106897-6fd0-462c-9d26-a832f1385e04\") " pod="openshift-marketplace/marketplace-operator-79b997595-tdctv" Nov 25 10:36:14 crc kubenswrapper[4821]: I1125 10:36:14.091980 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xb9l4\" (UniqueName: \"kubernetes.io/projected/05106897-6fd0-462c-9d26-a832f1385e04-kube-api-access-xb9l4\") pod \"marketplace-operator-79b997595-tdctv\" (UID: \"05106897-6fd0-462c-9d26-a832f1385e04\") " pod="openshift-marketplace/marketplace-operator-79b997595-tdctv" Nov 25 10:36:14 crc kubenswrapper[4821]: I1125 10:36:14.092045 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/05106897-6fd0-462c-9d26-a832f1385e04-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-tdctv\" (UID: \"05106897-6fd0-462c-9d26-a832f1385e04\") " pod="openshift-marketplace/marketplace-operator-79b997595-tdctv" Nov 25 10:36:14 crc kubenswrapper[4821]: I1125 10:36:14.092109 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/05106897-6fd0-462c-9d26-a832f1385e04-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-tdctv\" (UID: \"05106897-6fd0-462c-9d26-a832f1385e04\") " pod="openshift-marketplace/marketplace-operator-79b997595-tdctv" Nov 25 10:36:14 crc kubenswrapper[4821]: I1125 10:36:14.093436 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/05106897-6fd0-462c-9d26-a832f1385e04-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-tdctv\" (UID: \"05106897-6fd0-462c-9d26-a832f1385e04\") " pod="openshift-marketplace/marketplace-operator-79b997595-tdctv" Nov 25 10:36:14 crc kubenswrapper[4821]: I1125 10:36:14.099812 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/05106897-6fd0-462c-9d26-a832f1385e04-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-tdctv\" (UID: \"05106897-6fd0-462c-9d26-a832f1385e04\") " pod="openshift-marketplace/marketplace-operator-79b997595-tdctv" Nov 25 10:36:14 crc kubenswrapper[4821]: I1125 10:36:14.108865 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xb9l4\" (UniqueName: \"kubernetes.io/projected/05106897-6fd0-462c-9d26-a832f1385e04-kube-api-access-xb9l4\") pod \"marketplace-operator-79b997595-tdctv\" (UID: \"05106897-6fd0-462c-9d26-a832f1385e04\") " pod="openshift-marketplace/marketplace-operator-79b997595-tdctv" Nov 25 10:36:14 crc kubenswrapper[4821]: I1125 10:36:14.146979 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-tdctv" Nov 25 10:36:14 crc kubenswrapper[4821]: I1125 10:36:14.182653 4821 generic.go:334] "Generic (PLEG): container finished" podID="590f742c-e0d3-4394-8ef3-08bde5d57d01" containerID="41b4dac6291ebdf0ed0a316060e54df84fbc3b2fc75af2abd1e432330754933d" exitCode=0 Nov 25 10:36:14 crc kubenswrapper[4821]: I1125 10:36:14.182724 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-dxf4x" event={"ID":"590f742c-e0d3-4394-8ef3-08bde5d57d01","Type":"ContainerDied","Data":"41b4dac6291ebdf0ed0a316060e54df84fbc3b2fc75af2abd1e432330754933d"} Nov 25 10:36:14 crc kubenswrapper[4821]: I1125 10:36:14.530480 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-dxf4x" Nov 25 10:36:14 crc kubenswrapper[4821]: I1125 10:36:14.536761 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zrl8b" Nov 25 10:36:14 crc kubenswrapper[4821]: I1125 10:36:14.539649 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-87rxm_38068a8c-fd52-492c-9e1e-6425830bd097/registry-server/0.log" Nov 25 10:36:14 crc kubenswrapper[4821]: I1125 10:36:14.540296 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-87rxm" Nov 25 10:36:14 crc kubenswrapper[4821]: I1125 10:36:14.703556 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fswpt\" (UniqueName: \"kubernetes.io/projected/d1d6804a-27b6-4ef8-90ce-32606c93869a-kube-api-access-fswpt\") pod \"d1d6804a-27b6-4ef8-90ce-32606c93869a\" (UID: \"d1d6804a-27b6-4ef8-90ce-32606c93869a\") " Nov 25 10:36:14 crc kubenswrapper[4821]: I1125 10:36:14.703884 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9pxr7\" (UniqueName: \"kubernetes.io/projected/590f742c-e0d3-4394-8ef3-08bde5d57d01-kube-api-access-9pxr7\") pod \"590f742c-e0d3-4394-8ef3-08bde5d57d01\" (UID: \"590f742c-e0d3-4394-8ef3-08bde5d57d01\") " Nov 25 10:36:14 crc kubenswrapper[4821]: I1125 10:36:14.703929 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38068a8c-fd52-492c-9e1e-6425830bd097-utilities\") pod \"38068a8c-fd52-492c-9e1e-6425830bd097\" (UID: \"38068a8c-fd52-492c-9e1e-6425830bd097\") " Nov 25 10:36:14 crc kubenswrapper[4821]: I1125 10:36:14.703956 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1d6804a-27b6-4ef8-90ce-32606c93869a-utilities\") pod \"d1d6804a-27b6-4ef8-90ce-32606c93869a\" (UID: \"d1d6804a-27b6-4ef8-90ce-32606c93869a\") " Nov 25 10:36:14 crc kubenswrapper[4821]: I1125 10:36:14.703973 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/590f742c-e0d3-4394-8ef3-08bde5d57d01-marketplace-trusted-ca\") pod \"590f742c-e0d3-4394-8ef3-08bde5d57d01\" (UID: \"590f742c-e0d3-4394-8ef3-08bde5d57d01\") " Nov 25 10:36:14 crc kubenswrapper[4821]: I1125 10:36:14.703993 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1d6804a-27b6-4ef8-90ce-32606c93869a-catalog-content\") pod \"d1d6804a-27b6-4ef8-90ce-32606c93869a\" (UID: \"d1d6804a-27b6-4ef8-90ce-32606c93869a\") " Nov 25 10:36:14 crc kubenswrapper[4821]: I1125 10:36:14.704017 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6gpf2\" (UniqueName: \"kubernetes.io/projected/38068a8c-fd52-492c-9e1e-6425830bd097-kube-api-access-6gpf2\") pod \"38068a8c-fd52-492c-9e1e-6425830bd097\" (UID: \"38068a8c-fd52-492c-9e1e-6425830bd097\") " Nov 25 10:36:14 crc kubenswrapper[4821]: I1125 10:36:14.704048 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/590f742c-e0d3-4394-8ef3-08bde5d57d01-marketplace-operator-metrics\") pod \"590f742c-e0d3-4394-8ef3-08bde5d57d01\" (UID: \"590f742c-e0d3-4394-8ef3-08bde5d57d01\") " Nov 25 10:36:14 crc kubenswrapper[4821]: I1125 10:36:14.704073 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38068a8c-fd52-492c-9e1e-6425830bd097-catalog-content\") pod \"38068a8c-fd52-492c-9e1e-6425830bd097\" (UID: \"38068a8c-fd52-492c-9e1e-6425830bd097\") " Nov 25 10:36:14 crc kubenswrapper[4821]: I1125 10:36:14.707856 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38068a8c-fd52-492c-9e1e-6425830bd097-utilities" (OuterVolumeSpecName: "utilities") pod "38068a8c-fd52-492c-9e1e-6425830bd097" (UID: "38068a8c-fd52-492c-9e1e-6425830bd097"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:36:14 crc kubenswrapper[4821]: I1125 10:36:14.709196 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/590f742c-e0d3-4394-8ef3-08bde5d57d01-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "590f742c-e0d3-4394-8ef3-08bde5d57d01" (UID: "590f742c-e0d3-4394-8ef3-08bde5d57d01"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:36:14 crc kubenswrapper[4821]: I1125 10:36:14.710633 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d1d6804a-27b6-4ef8-90ce-32606c93869a-utilities" (OuterVolumeSpecName: "utilities") pod "d1d6804a-27b6-4ef8-90ce-32606c93869a" (UID: "d1d6804a-27b6-4ef8-90ce-32606c93869a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:36:14 crc kubenswrapper[4821]: I1125 10:36:14.712122 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/590f742c-e0d3-4394-8ef3-08bde5d57d01-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "590f742c-e0d3-4394-8ef3-08bde5d57d01" (UID: "590f742c-e0d3-4394-8ef3-08bde5d57d01"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:36:14 crc kubenswrapper[4821]: I1125 10:36:14.713632 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38068a8c-fd52-492c-9e1e-6425830bd097-kube-api-access-6gpf2" (OuterVolumeSpecName: "kube-api-access-6gpf2") pod "38068a8c-fd52-492c-9e1e-6425830bd097" (UID: "38068a8c-fd52-492c-9e1e-6425830bd097"). InnerVolumeSpecName "kube-api-access-6gpf2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:36:14 crc kubenswrapper[4821]: I1125 10:36:14.717797 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/590f742c-e0d3-4394-8ef3-08bde5d57d01-kube-api-access-9pxr7" (OuterVolumeSpecName: "kube-api-access-9pxr7") pod "590f742c-e0d3-4394-8ef3-08bde5d57d01" (UID: "590f742c-e0d3-4394-8ef3-08bde5d57d01"). InnerVolumeSpecName "kube-api-access-9pxr7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:36:14 crc kubenswrapper[4821]: I1125 10:36:14.719019 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1d6804a-27b6-4ef8-90ce-32606c93869a-kube-api-access-fswpt" (OuterVolumeSpecName: "kube-api-access-fswpt") pod "d1d6804a-27b6-4ef8-90ce-32606c93869a" (UID: "d1d6804a-27b6-4ef8-90ce-32606c93869a"). InnerVolumeSpecName "kube-api-access-fswpt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:36:14 crc kubenswrapper[4821]: I1125 10:36:14.737302 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-tdctv"] Nov 25 10:36:14 crc kubenswrapper[4821]: I1125 10:36:14.805094 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d1d6804a-27b6-4ef8-90ce-32606c93869a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d1d6804a-27b6-4ef8-90ce-32606c93869a" (UID: "d1d6804a-27b6-4ef8-90ce-32606c93869a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:36:14 crc kubenswrapper[4821]: I1125 10:36:14.805362 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1d6804a-27b6-4ef8-90ce-32606c93869a-catalog-content\") pod \"d1d6804a-27b6-4ef8-90ce-32606c93869a\" (UID: \"d1d6804a-27b6-4ef8-90ce-32606c93869a\") " Nov 25 10:36:14 crc kubenswrapper[4821]: W1125 10:36:14.805535 4821 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/d1d6804a-27b6-4ef8-90ce-32606c93869a/volumes/kubernetes.io~empty-dir/catalog-content Nov 25 10:36:14 crc kubenswrapper[4821]: I1125 10:36:14.805565 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d1d6804a-27b6-4ef8-90ce-32606c93869a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d1d6804a-27b6-4ef8-90ce-32606c93869a" (UID: "d1d6804a-27b6-4ef8-90ce-32606c93869a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:36:14 crc kubenswrapper[4821]: I1125 10:36:14.805856 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9pxr7\" (UniqueName: \"kubernetes.io/projected/590f742c-e0d3-4394-8ef3-08bde5d57d01-kube-api-access-9pxr7\") on node \"crc\" DevicePath \"\"" Nov 25 10:36:14 crc kubenswrapper[4821]: I1125 10:36:14.805872 4821 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38068a8c-fd52-492c-9e1e-6425830bd097-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:36:14 crc kubenswrapper[4821]: I1125 10:36:14.805882 4821 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1d6804a-27b6-4ef8-90ce-32606c93869a-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:36:14 crc kubenswrapper[4821]: I1125 10:36:14.805891 4821 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/590f742c-e0d3-4394-8ef3-08bde5d57d01-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 10:36:14 crc kubenswrapper[4821]: I1125 10:36:14.805899 4821 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1d6804a-27b6-4ef8-90ce-32606c93869a-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:36:14 crc kubenswrapper[4821]: I1125 10:36:14.805925 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6gpf2\" (UniqueName: \"kubernetes.io/projected/38068a8c-fd52-492c-9e1e-6425830bd097-kube-api-access-6gpf2\") on node \"crc\" DevicePath \"\"" Nov 25 10:36:14 crc kubenswrapper[4821]: I1125 10:36:14.805933 4821 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/590f742c-e0d3-4394-8ef3-08bde5d57d01-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 25 10:36:14 crc kubenswrapper[4821]: I1125 10:36:14.805942 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fswpt\" (UniqueName: \"kubernetes.io/projected/d1d6804a-27b6-4ef8-90ce-32606c93869a-kube-api-access-fswpt\") on node \"crc\" DevicePath \"\"" Nov 25 10:36:14 crc kubenswrapper[4821]: I1125 10:36:14.826112 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38068a8c-fd52-492c-9e1e-6425830bd097-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "38068a8c-fd52-492c-9e1e-6425830bd097" (UID: "38068a8c-fd52-492c-9e1e-6425830bd097"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:36:14 crc kubenswrapper[4821]: I1125 10:36:14.907335 4821 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38068a8c-fd52-492c-9e1e-6425830bd097-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:36:15 crc kubenswrapper[4821]: I1125 10:36:15.211563 4821 generic.go:334] "Generic (PLEG): container finished" podID="d1d6804a-27b6-4ef8-90ce-32606c93869a" containerID="f2aba3affd94c7e2bdb62665d9d38d71f158474e556cb47e4248b9556a96b832" exitCode=0 Nov 25 10:36:15 crc kubenswrapper[4821]: I1125 10:36:15.211651 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zrl8b" Nov 25 10:36:15 crc kubenswrapper[4821]: I1125 10:36:15.211672 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zrl8b" event={"ID":"d1d6804a-27b6-4ef8-90ce-32606c93869a","Type":"ContainerDied","Data":"f2aba3affd94c7e2bdb62665d9d38d71f158474e556cb47e4248b9556a96b832"} Nov 25 10:36:15 crc kubenswrapper[4821]: I1125 10:36:15.212200 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zrl8b" event={"ID":"d1d6804a-27b6-4ef8-90ce-32606c93869a","Type":"ContainerDied","Data":"ff85d226a79311dc79303698b35f33a95a69497d27785c9a493d4a529ed46f8b"} Nov 25 10:36:15 crc kubenswrapper[4821]: I1125 10:36:15.212237 4821 scope.go:117] "RemoveContainer" containerID="f2aba3affd94c7e2bdb62665d9d38d71f158474e556cb47e4248b9556a96b832" Nov 25 10:36:15 crc kubenswrapper[4821]: I1125 10:36:15.222362 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cgdxc" event={"ID":"b6e72ee2-50e4-4643-aee6-8be89d2e4589","Type":"ContainerStarted","Data":"07827fbd586d8e165370290e88d9ac62faf240ebe455d8fc2a232225c32e14e0"} Nov 25 10:36:15 crc kubenswrapper[4821]: I1125 10:36:15.222552 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-cgdxc" podUID="b6e72ee2-50e4-4643-aee6-8be89d2e4589" containerName="registry-server" containerID="cri-o://07827fbd586d8e165370290e88d9ac62faf240ebe455d8fc2a232225c32e14e0" gracePeriod=30 Nov 25 10:36:15 crc kubenswrapper[4821]: I1125 10:36:15.232373 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-87rxm_38068a8c-fd52-492c-9e1e-6425830bd097/registry-server/0.log" Nov 25 10:36:15 crc kubenswrapper[4821]: I1125 10:36:15.236398 4821 generic.go:334] "Generic (PLEG): container finished" podID="38068a8c-fd52-492c-9e1e-6425830bd097" containerID="c69c3a931a80bb0669ee855880d4f2c246003cbfae84bcf0a7a74900e2a9c840" exitCode=1 Nov 25 10:36:15 crc kubenswrapper[4821]: I1125 10:36:15.236526 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-87rxm" Nov 25 10:36:15 crc kubenswrapper[4821]: I1125 10:36:15.236527 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-87rxm" event={"ID":"38068a8c-fd52-492c-9e1e-6425830bd097","Type":"ContainerDied","Data":"c69c3a931a80bb0669ee855880d4f2c246003cbfae84bcf0a7a74900e2a9c840"} Nov 25 10:36:15 crc kubenswrapper[4821]: I1125 10:36:15.237877 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-87rxm" event={"ID":"38068a8c-fd52-492c-9e1e-6425830bd097","Type":"ContainerDied","Data":"8ce0c031e2afe5eb9c02d7a8bf7a49a4a7470c86707d1879863de9afea8a4e43"} Nov 25 10:36:15 crc kubenswrapper[4821]: I1125 10:36:15.245857 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-dxf4x" event={"ID":"590f742c-e0d3-4394-8ef3-08bde5d57d01","Type":"ContainerDied","Data":"d5fa1194e5ea743ba24948633dced0251487682c43e560805055518f64d3cc40"} Nov 25 10:36:15 crc kubenswrapper[4821]: I1125 10:36:15.245947 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-dxf4x" Nov 25 10:36:15 crc kubenswrapper[4821]: I1125 10:36:15.251116 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-cgdxc" podStartSLOduration=4.104878377 podStartE2EDuration="1m38.251100462s" podCreationTimestamp="2025-11-25 10:34:37 +0000 UTC" firstStartedPulling="2025-11-25 10:34:40.305631454 +0000 UTC m=+150.841951301" lastFinishedPulling="2025-11-25 10:36:14.451853539 +0000 UTC m=+244.988173386" observedRunningTime="2025-11-25 10:36:15.245440519 +0000 UTC m=+245.781760366" watchObservedRunningTime="2025-11-25 10:36:15.251100462 +0000 UTC m=+245.787420309" Nov 25 10:36:15 crc kubenswrapper[4821]: I1125 10:36:15.255221 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-tdctv" event={"ID":"05106897-6fd0-462c-9d26-a832f1385e04","Type":"ContainerStarted","Data":"2979c9a59cb1f8ed4952562e6979a286ae4b1d7c01884f4fcb82676d62f4cec5"} Nov 25 10:36:15 crc kubenswrapper[4821]: I1125 10:36:15.266552 4821 generic.go:334] "Generic (PLEG): container finished" podID="645993fd-bdce-435e-a600-34bf1ff5b38b" containerID="fb8bc6bac5119a281c39502ff9c012e4339e380bc9b8345e89e8674d044bc435" exitCode=0 Nov 25 10:36:15 crc kubenswrapper[4821]: I1125 10:36:15.266823 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-w56ql" podUID="4b4b52a8-67ad-4c51-86ab-503c75ad2024" containerName="registry-server" containerID="cri-o://51fe6850e14ffc694a9db8f49ec8649f913025a36e30f3269bd3e251d416726a" gracePeriod=30 Nov 25 10:36:15 crc kubenswrapper[4821]: I1125 10:36:15.267047 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ps9cl" event={"ID":"645993fd-bdce-435e-a600-34bf1ff5b38b","Type":"ContainerDied","Data":"fb8bc6bac5119a281c39502ff9c012e4339e380bc9b8345e89e8674d044bc435"} Nov 25 10:36:15 crc kubenswrapper[4821]: I1125 10:36:15.317441 4821 scope.go:117] "RemoveContainer" containerID="da0688502ecc52dd6fc2b2bcb4dedc3ab4e27c36774a7eed2696ea4a0dca81e9" Nov 25 10:36:15 crc kubenswrapper[4821]: I1125 10:36:15.322383 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zrl8b"] Nov 25 10:36:15 crc kubenswrapper[4821]: I1125 10:36:15.328457 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zrl8b"] Nov 25 10:36:15 crc kubenswrapper[4821]: I1125 10:36:15.410395 4821 scope.go:117] "RemoveContainer" containerID="4811a3c8d2e18aa424fa1d9934482633b4268620e85b281a326c72984ea29e55" Nov 25 10:36:15 crc kubenswrapper[4821]: I1125 10:36:15.413460 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-dxf4x"] Nov 25 10:36:15 crc kubenswrapper[4821]: I1125 10:36:15.421018 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-dxf4x"] Nov 25 10:36:15 crc kubenswrapper[4821]: I1125 10:36:15.452140 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-87rxm"] Nov 25 10:36:15 crc kubenswrapper[4821]: I1125 10:36:15.462054 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-87rxm"] Nov 25 10:36:15 crc kubenswrapper[4821]: I1125 10:36:15.520955 4821 scope.go:117] "RemoveContainer" containerID="f2aba3affd94c7e2bdb62665d9d38d71f158474e556cb47e4248b9556a96b832" Nov 25 10:36:15 crc kubenswrapper[4821]: E1125 10:36:15.522529 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f2aba3affd94c7e2bdb62665d9d38d71f158474e556cb47e4248b9556a96b832\": container with ID starting with f2aba3affd94c7e2bdb62665d9d38d71f158474e556cb47e4248b9556a96b832 not found: ID does not exist" containerID="f2aba3affd94c7e2bdb62665d9d38d71f158474e556cb47e4248b9556a96b832" Nov 25 10:36:15 crc kubenswrapper[4821]: I1125 10:36:15.522588 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2aba3affd94c7e2bdb62665d9d38d71f158474e556cb47e4248b9556a96b832"} err="failed to get container status \"f2aba3affd94c7e2bdb62665d9d38d71f158474e556cb47e4248b9556a96b832\": rpc error: code = NotFound desc = could not find container \"f2aba3affd94c7e2bdb62665d9d38d71f158474e556cb47e4248b9556a96b832\": container with ID starting with f2aba3affd94c7e2bdb62665d9d38d71f158474e556cb47e4248b9556a96b832 not found: ID does not exist" Nov 25 10:36:15 crc kubenswrapper[4821]: I1125 10:36:15.522615 4821 scope.go:117] "RemoveContainer" containerID="da0688502ecc52dd6fc2b2bcb4dedc3ab4e27c36774a7eed2696ea4a0dca81e9" Nov 25 10:36:15 crc kubenswrapper[4821]: E1125 10:36:15.524413 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da0688502ecc52dd6fc2b2bcb4dedc3ab4e27c36774a7eed2696ea4a0dca81e9\": container with ID starting with da0688502ecc52dd6fc2b2bcb4dedc3ab4e27c36774a7eed2696ea4a0dca81e9 not found: ID does not exist" containerID="da0688502ecc52dd6fc2b2bcb4dedc3ab4e27c36774a7eed2696ea4a0dca81e9" Nov 25 10:36:15 crc kubenswrapper[4821]: I1125 10:36:15.524455 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da0688502ecc52dd6fc2b2bcb4dedc3ab4e27c36774a7eed2696ea4a0dca81e9"} err="failed to get container status \"da0688502ecc52dd6fc2b2bcb4dedc3ab4e27c36774a7eed2696ea4a0dca81e9\": rpc error: code = NotFound desc = could not find container \"da0688502ecc52dd6fc2b2bcb4dedc3ab4e27c36774a7eed2696ea4a0dca81e9\": container with ID starting with da0688502ecc52dd6fc2b2bcb4dedc3ab4e27c36774a7eed2696ea4a0dca81e9 not found: ID does not exist" Nov 25 10:36:15 crc kubenswrapper[4821]: I1125 10:36:15.524483 4821 scope.go:117] "RemoveContainer" containerID="4811a3c8d2e18aa424fa1d9934482633b4268620e85b281a326c72984ea29e55" Nov 25 10:36:15 crc kubenswrapper[4821]: E1125 10:36:15.528176 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4811a3c8d2e18aa424fa1d9934482633b4268620e85b281a326c72984ea29e55\": container with ID starting with 4811a3c8d2e18aa424fa1d9934482633b4268620e85b281a326c72984ea29e55 not found: ID does not exist" containerID="4811a3c8d2e18aa424fa1d9934482633b4268620e85b281a326c72984ea29e55" Nov 25 10:36:15 crc kubenswrapper[4821]: I1125 10:36:15.528203 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4811a3c8d2e18aa424fa1d9934482633b4268620e85b281a326c72984ea29e55"} err="failed to get container status \"4811a3c8d2e18aa424fa1d9934482633b4268620e85b281a326c72984ea29e55\": rpc error: code = NotFound desc = could not find container \"4811a3c8d2e18aa424fa1d9934482633b4268620e85b281a326c72984ea29e55\": container with ID starting with 4811a3c8d2e18aa424fa1d9934482633b4268620e85b281a326c72984ea29e55 not found: ID does not exist" Nov 25 10:36:15 crc kubenswrapper[4821]: I1125 10:36:15.528220 4821 scope.go:117] "RemoveContainer" containerID="c69c3a931a80bb0669ee855880d4f2c246003cbfae84bcf0a7a74900e2a9c840" Nov 25 10:36:15 crc kubenswrapper[4821]: I1125 10:36:15.553835 4821 scope.go:117] "RemoveContainer" containerID="205a1bb3a9091c6fdd3137e986e9d6b47b7a4b2fd20e0c92aeb31fccc6aea4a3" Nov 25 10:36:15 crc kubenswrapper[4821]: I1125 10:36:15.578989 4821 scope.go:117] "RemoveContainer" containerID="9566c37aaf942dcc55362ff747f161856ea86449e19e5c458bda884d3e4afae5" Nov 25 10:36:15 crc kubenswrapper[4821]: I1125 10:36:15.591535 4821 scope.go:117] "RemoveContainer" containerID="c69c3a931a80bb0669ee855880d4f2c246003cbfae84bcf0a7a74900e2a9c840" Nov 25 10:36:15 crc kubenswrapper[4821]: E1125 10:36:15.594561 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c69c3a931a80bb0669ee855880d4f2c246003cbfae84bcf0a7a74900e2a9c840\": container with ID starting with c69c3a931a80bb0669ee855880d4f2c246003cbfae84bcf0a7a74900e2a9c840 not found: ID does not exist" containerID="c69c3a931a80bb0669ee855880d4f2c246003cbfae84bcf0a7a74900e2a9c840" Nov 25 10:36:15 crc kubenswrapper[4821]: I1125 10:36:15.594602 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c69c3a931a80bb0669ee855880d4f2c246003cbfae84bcf0a7a74900e2a9c840"} err="failed to get container status \"c69c3a931a80bb0669ee855880d4f2c246003cbfae84bcf0a7a74900e2a9c840\": rpc error: code = NotFound desc = could not find container \"c69c3a931a80bb0669ee855880d4f2c246003cbfae84bcf0a7a74900e2a9c840\": container with ID starting with c69c3a931a80bb0669ee855880d4f2c246003cbfae84bcf0a7a74900e2a9c840 not found: ID does not exist" Nov 25 10:36:15 crc kubenswrapper[4821]: I1125 10:36:15.594631 4821 scope.go:117] "RemoveContainer" containerID="205a1bb3a9091c6fdd3137e986e9d6b47b7a4b2fd20e0c92aeb31fccc6aea4a3" Nov 25 10:36:15 crc kubenswrapper[4821]: E1125 10:36:15.595403 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"205a1bb3a9091c6fdd3137e986e9d6b47b7a4b2fd20e0c92aeb31fccc6aea4a3\": container with ID starting with 205a1bb3a9091c6fdd3137e986e9d6b47b7a4b2fd20e0c92aeb31fccc6aea4a3 not found: ID does not exist" containerID="205a1bb3a9091c6fdd3137e986e9d6b47b7a4b2fd20e0c92aeb31fccc6aea4a3" Nov 25 10:36:15 crc kubenswrapper[4821]: I1125 10:36:15.595433 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"205a1bb3a9091c6fdd3137e986e9d6b47b7a4b2fd20e0c92aeb31fccc6aea4a3"} err="failed to get container status \"205a1bb3a9091c6fdd3137e986e9d6b47b7a4b2fd20e0c92aeb31fccc6aea4a3\": rpc error: code = NotFound desc = could not find container \"205a1bb3a9091c6fdd3137e986e9d6b47b7a4b2fd20e0c92aeb31fccc6aea4a3\": container with ID starting with 205a1bb3a9091c6fdd3137e986e9d6b47b7a4b2fd20e0c92aeb31fccc6aea4a3 not found: ID does not exist" Nov 25 10:36:15 crc kubenswrapper[4821]: I1125 10:36:15.595453 4821 scope.go:117] "RemoveContainer" containerID="9566c37aaf942dcc55362ff747f161856ea86449e19e5c458bda884d3e4afae5" Nov 25 10:36:15 crc kubenswrapper[4821]: E1125 10:36:15.595731 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9566c37aaf942dcc55362ff747f161856ea86449e19e5c458bda884d3e4afae5\": container with ID starting with 9566c37aaf942dcc55362ff747f161856ea86449e19e5c458bda884d3e4afae5 not found: ID does not exist" containerID="9566c37aaf942dcc55362ff747f161856ea86449e19e5c458bda884d3e4afae5" Nov 25 10:36:15 crc kubenswrapper[4821]: I1125 10:36:15.595778 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9566c37aaf942dcc55362ff747f161856ea86449e19e5c458bda884d3e4afae5"} err="failed to get container status \"9566c37aaf942dcc55362ff747f161856ea86449e19e5c458bda884d3e4afae5\": rpc error: code = NotFound desc = could not find container \"9566c37aaf942dcc55362ff747f161856ea86449e19e5c458bda884d3e4afae5\": container with ID starting with 9566c37aaf942dcc55362ff747f161856ea86449e19e5c458bda884d3e4afae5 not found: ID does not exist" Nov 25 10:36:15 crc kubenswrapper[4821]: I1125 10:36:15.595817 4821 scope.go:117] "RemoveContainer" containerID="41b4dac6291ebdf0ed0a316060e54df84fbc3b2fc75af2abd1e432330754933d" Nov 25 10:36:15 crc kubenswrapper[4821]: I1125 10:36:15.613207 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ps9cl" Nov 25 10:36:15 crc kubenswrapper[4821]: I1125 10:36:15.716592 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/645993fd-bdce-435e-a600-34bf1ff5b38b-utilities\") pod \"645993fd-bdce-435e-a600-34bf1ff5b38b\" (UID: \"645993fd-bdce-435e-a600-34bf1ff5b38b\") " Nov 25 10:36:15 crc kubenswrapper[4821]: I1125 10:36:15.716668 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/645993fd-bdce-435e-a600-34bf1ff5b38b-catalog-content\") pod \"645993fd-bdce-435e-a600-34bf1ff5b38b\" (UID: \"645993fd-bdce-435e-a600-34bf1ff5b38b\") " Nov 25 10:36:15 crc kubenswrapper[4821]: I1125 10:36:15.716694 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xn8qr\" (UniqueName: \"kubernetes.io/projected/645993fd-bdce-435e-a600-34bf1ff5b38b-kube-api-access-xn8qr\") pod \"645993fd-bdce-435e-a600-34bf1ff5b38b\" (UID: \"645993fd-bdce-435e-a600-34bf1ff5b38b\") " Nov 25 10:36:15 crc kubenswrapper[4821]: I1125 10:36:15.717667 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/645993fd-bdce-435e-a600-34bf1ff5b38b-utilities" (OuterVolumeSpecName: "utilities") pod "645993fd-bdce-435e-a600-34bf1ff5b38b" (UID: "645993fd-bdce-435e-a600-34bf1ff5b38b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:36:15 crc kubenswrapper[4821]: I1125 10:36:15.721827 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/645993fd-bdce-435e-a600-34bf1ff5b38b-kube-api-access-xn8qr" (OuterVolumeSpecName: "kube-api-access-xn8qr") pod "645993fd-bdce-435e-a600-34bf1ff5b38b" (UID: "645993fd-bdce-435e-a600-34bf1ff5b38b"). InnerVolumeSpecName "kube-api-access-xn8qr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:36:15 crc kubenswrapper[4821]: I1125 10:36:15.772742 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/645993fd-bdce-435e-a600-34bf1ff5b38b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "645993fd-bdce-435e-a600-34bf1ff5b38b" (UID: "645993fd-bdce-435e-a600-34bf1ff5b38b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:36:15 crc kubenswrapper[4821]: I1125 10:36:15.818155 4821 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/645993fd-bdce-435e-a600-34bf1ff5b38b-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:36:15 crc kubenswrapper[4821]: I1125 10:36:15.818210 4821 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/645993fd-bdce-435e-a600-34bf1ff5b38b-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:36:15 crc kubenswrapper[4821]: I1125 10:36:15.818221 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xn8qr\" (UniqueName: \"kubernetes.io/projected/645993fd-bdce-435e-a600-34bf1ff5b38b-kube-api-access-xn8qr\") on node \"crc\" DevicePath \"\"" Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.122491 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="38068a8c-fd52-492c-9e1e-6425830bd097" path="/var/lib/kubelet/pods/38068a8c-fd52-492c-9e1e-6425830bd097/volumes" Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.123285 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="590f742c-e0d3-4394-8ef3-08bde5d57d01" path="/var/lib/kubelet/pods/590f742c-e0d3-4394-8ef3-08bde5d57d01/volumes" Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.123799 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1d6804a-27b6-4ef8-90ce-32606c93869a" path="/var/lib/kubelet/pods/d1d6804a-27b6-4ef8-90ce-32606c93869a/volumes" Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.276302 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8qpmf" event={"ID":"d8becff5-8017-4073-89ae-c17cc4c32922","Type":"ContainerStarted","Data":"d9cebe2065a7afab486b6363583a5992907042ff24a825458f0d0a35a8da9eeb"} Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.276652 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-8qpmf" podUID="d8becff5-8017-4073-89ae-c17cc4c32922" containerName="registry-server" containerID="cri-o://d9cebe2065a7afab486b6363583a5992907042ff24a825458f0d0a35a8da9eeb" gracePeriod=30 Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.285364 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-cgdxc_b6e72ee2-50e4-4643-aee6-8be89d2e4589/registry-server/0.log" Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.290758 4821 generic.go:334] "Generic (PLEG): container finished" podID="b6e72ee2-50e4-4643-aee6-8be89d2e4589" containerID="07827fbd586d8e165370290e88d9ac62faf240ebe455d8fc2a232225c32e14e0" exitCode=1 Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.290870 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cgdxc" event={"ID":"b6e72ee2-50e4-4643-aee6-8be89d2e4589","Type":"ContainerDied","Data":"07827fbd586d8e165370290e88d9ac62faf240ebe455d8fc2a232225c32e14e0"} Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.302663 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r5nmb" event={"ID":"e1391070-3aad-47c3-a2f1-0c1697b014bf","Type":"ContainerStarted","Data":"0ab5237249ce6715dff6a126e5d1583e88387a901b75476727dc44baddeef332"} Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.302701 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-r5nmb" podUID="e1391070-3aad-47c3-a2f1-0c1697b014bf" containerName="registry-server" containerID="cri-o://0ab5237249ce6715dff6a126e5d1583e88387a901b75476727dc44baddeef332" gracePeriod=30 Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.308594 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-w56ql_4b4b52a8-67ad-4c51-86ab-503c75ad2024/registry-server/0.log" Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.309480 4821 generic.go:334] "Generic (PLEG): container finished" podID="4b4b52a8-67ad-4c51-86ab-503c75ad2024" containerID="51fe6850e14ffc694a9db8f49ec8649f913025a36e30f3269bd3e251d416726a" exitCode=1 Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.309542 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w56ql" event={"ID":"4b4b52a8-67ad-4c51-86ab-503c75ad2024","Type":"ContainerDied","Data":"51fe6850e14ffc694a9db8f49ec8649f913025a36e30f3269bd3e251d416726a"} Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.312721 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ps9cl" event={"ID":"645993fd-bdce-435e-a600-34bf1ff5b38b","Type":"ContainerDied","Data":"623f8448370b9aaeb71701b509f83954b1d63115c2e739cb4387ec020094e308"} Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.312761 4821 scope.go:117] "RemoveContainer" containerID="fb8bc6bac5119a281c39502ff9c012e4339e380bc9b8345e89e8674d044bc435" Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.312798 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ps9cl" Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.320720 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-8qpmf" podStartSLOduration=5.2514367029999995 podStartE2EDuration="1m37.320705554s" podCreationTimestamp="2025-11-25 10:34:39 +0000 UTC" firstStartedPulling="2025-11-25 10:34:42.495089001 +0000 UTC m=+153.031408848" lastFinishedPulling="2025-11-25 10:36:14.564357852 +0000 UTC m=+245.100677699" observedRunningTime="2025-11-25 10:36:16.302063928 +0000 UTC m=+246.838383785" watchObservedRunningTime="2025-11-25 10:36:16.320705554 +0000 UTC m=+246.857025391" Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.322134 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-r5nmb" podStartSLOduration=5.048232932 podStartE2EDuration="1m39.322128993s" podCreationTimestamp="2025-11-25 10:34:37 +0000 UTC" firstStartedPulling="2025-11-25 10:34:40.335564516 +0000 UTC m=+150.871884363" lastFinishedPulling="2025-11-25 10:36:14.609460577 +0000 UTC m=+245.145780424" observedRunningTime="2025-11-25 10:36:16.318983178 +0000 UTC m=+246.855303025" watchObservedRunningTime="2025-11-25 10:36:16.322128993 +0000 UTC m=+246.858448840" Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.325415 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2kwcp" event={"ID":"927e7d40-852e-4807-bb22-d4f7c6e9d2ca","Type":"ContainerStarted","Data":"42feaa125fff6b12b55e21ca82b443e338cf19fd8d839e42cd6ddb18de30a91a"} Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.325528 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-2kwcp" podUID="927e7d40-852e-4807-bb22-d4f7c6e9d2ca" containerName="registry-server" containerID="cri-o://42feaa125fff6b12b55e21ca82b443e338cf19fd8d839e42cd6ddb18de30a91a" gracePeriod=30 Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.328601 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-tdctv" event={"ID":"05106897-6fd0-462c-9d26-a832f1385e04","Type":"ContainerStarted","Data":"7e77d58e759be1880f16e7395d34215cbb28ad66779e0ea19215778fa8bbeb61"} Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.329026 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-tdctv" Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.334466 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-tdctv" Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.342084 4821 scope.go:117] "RemoveContainer" containerID="10de088200d33e95b9677889b71c6c6aa52bf166742c2c0f86179375641acb43" Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.364474 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-2kwcp" podStartSLOduration=3.673977417 podStartE2EDuration="1m37.364327958s" podCreationTimestamp="2025-11-25 10:34:39 +0000 UTC" firstStartedPulling="2025-11-25 10:34:41.434676289 +0000 UTC m=+151.970996136" lastFinishedPulling="2025-11-25 10:36:15.12502683 +0000 UTC m=+245.661346677" observedRunningTime="2025-11-25 10:36:16.351432428 +0000 UTC m=+246.887752295" watchObservedRunningTime="2025-11-25 10:36:16.364327958 +0000 UTC m=+246.900647805" Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.389358 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ps9cl"] Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.395857 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-ps9cl"] Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.402602 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-cgdxc_b6e72ee2-50e4-4643-aee6-8be89d2e4589/registry-server/0.log" Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.403367 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cgdxc" Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.415467 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-tdctv" podStartSLOduration=3.415449516 podStartE2EDuration="3.415449516s" podCreationTimestamp="2025-11-25 10:36:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:36:16.411906509 +0000 UTC m=+246.948226356" watchObservedRunningTime="2025-11-25 10:36:16.415449516 +0000 UTC m=+246.951769363" Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.522865 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-4j8hk"] Nov 25 10:36:16 crc kubenswrapper[4821]: E1125 10:36:16.523083 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1d6804a-27b6-4ef8-90ce-32606c93869a" containerName="registry-server" Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.523098 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1d6804a-27b6-4ef8-90ce-32606c93869a" containerName="registry-server" Nov 25 10:36:16 crc kubenswrapper[4821]: E1125 10:36:16.523110 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="590f742c-e0d3-4394-8ef3-08bde5d57d01" containerName="marketplace-operator" Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.523118 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="590f742c-e0d3-4394-8ef3-08bde5d57d01" containerName="marketplace-operator" Nov 25 10:36:16 crc kubenswrapper[4821]: E1125 10:36:16.523131 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6e72ee2-50e4-4643-aee6-8be89d2e4589" containerName="extract-content" Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.523142 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6e72ee2-50e4-4643-aee6-8be89d2e4589" containerName="extract-content" Nov 25 10:36:16 crc kubenswrapper[4821]: E1125 10:36:16.523152 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6e72ee2-50e4-4643-aee6-8be89d2e4589" containerName="extract-utilities" Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.523177 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6e72ee2-50e4-4643-aee6-8be89d2e4589" containerName="extract-utilities" Nov 25 10:36:16 crc kubenswrapper[4821]: E1125 10:36:16.523189 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38068a8c-fd52-492c-9e1e-6425830bd097" containerName="extract-content" Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.523196 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="38068a8c-fd52-492c-9e1e-6425830bd097" containerName="extract-content" Nov 25 10:36:16 crc kubenswrapper[4821]: E1125 10:36:16.523208 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1d6804a-27b6-4ef8-90ce-32606c93869a" containerName="extract-utilities" Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.523215 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1d6804a-27b6-4ef8-90ce-32606c93869a" containerName="extract-utilities" Nov 25 10:36:16 crc kubenswrapper[4821]: E1125 10:36:16.523223 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38068a8c-fd52-492c-9e1e-6425830bd097" containerName="registry-server" Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.523230 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="38068a8c-fd52-492c-9e1e-6425830bd097" containerName="registry-server" Nov 25 10:36:16 crc kubenswrapper[4821]: E1125 10:36:16.523240 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6e72ee2-50e4-4643-aee6-8be89d2e4589" containerName="registry-server" Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.523248 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6e72ee2-50e4-4643-aee6-8be89d2e4589" containerName="registry-server" Nov 25 10:36:16 crc kubenswrapper[4821]: E1125 10:36:16.523257 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38068a8c-fd52-492c-9e1e-6425830bd097" containerName="extract-utilities" Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.523264 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="38068a8c-fd52-492c-9e1e-6425830bd097" containerName="extract-utilities" Nov 25 10:36:16 crc kubenswrapper[4821]: E1125 10:36:16.523274 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="645993fd-bdce-435e-a600-34bf1ff5b38b" containerName="extract-content" Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.523283 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="645993fd-bdce-435e-a600-34bf1ff5b38b" containerName="extract-content" Nov 25 10:36:16 crc kubenswrapper[4821]: E1125 10:36:16.523294 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="645993fd-bdce-435e-a600-34bf1ff5b38b" containerName="extract-utilities" Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.523303 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="645993fd-bdce-435e-a600-34bf1ff5b38b" containerName="extract-utilities" Nov 25 10:36:16 crc kubenswrapper[4821]: E1125 10:36:16.523315 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1d6804a-27b6-4ef8-90ce-32606c93869a" containerName="extract-content" Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.523323 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1d6804a-27b6-4ef8-90ce-32606c93869a" containerName="extract-content" Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.523429 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="38068a8c-fd52-492c-9e1e-6425830bd097" containerName="registry-server" Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.523447 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="590f742c-e0d3-4394-8ef3-08bde5d57d01" containerName="marketplace-operator" Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.523459 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="645993fd-bdce-435e-a600-34bf1ff5b38b" containerName="extract-content" Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.523468 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1d6804a-27b6-4ef8-90ce-32606c93869a" containerName="registry-server" Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.523476 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6e72ee2-50e4-4643-aee6-8be89d2e4589" containerName="registry-server" Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.524505 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4j8hk" Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.524726 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-njv9h\" (UniqueName: \"kubernetes.io/projected/b6e72ee2-50e4-4643-aee6-8be89d2e4589-kube-api-access-njv9h\") pod \"b6e72ee2-50e4-4643-aee6-8be89d2e4589\" (UID: \"b6e72ee2-50e4-4643-aee6-8be89d2e4589\") " Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.524873 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6e72ee2-50e4-4643-aee6-8be89d2e4589-utilities\") pod \"b6e72ee2-50e4-4643-aee6-8be89d2e4589\" (UID: \"b6e72ee2-50e4-4643-aee6-8be89d2e4589\") " Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.524912 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6e72ee2-50e4-4643-aee6-8be89d2e4589-catalog-content\") pod \"b6e72ee2-50e4-4643-aee6-8be89d2e4589\" (UID: \"b6e72ee2-50e4-4643-aee6-8be89d2e4589\") " Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.525843 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b6e72ee2-50e4-4643-aee6-8be89d2e4589-utilities" (OuterVolumeSpecName: "utilities") pod "b6e72ee2-50e4-4643-aee6-8be89d2e4589" (UID: "b6e72ee2-50e4-4643-aee6-8be89d2e4589"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.533066 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4j8hk"] Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.533261 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6e72ee2-50e4-4643-aee6-8be89d2e4589-kube-api-access-njv9h" (OuterVolumeSpecName: "kube-api-access-njv9h") pod "b6e72ee2-50e4-4643-aee6-8be89d2e4589" (UID: "b6e72ee2-50e4-4643-aee6-8be89d2e4589"). InnerVolumeSpecName "kube-api-access-njv9h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.586577 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b6e72ee2-50e4-4643-aee6-8be89d2e4589-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b6e72ee2-50e4-4643-aee6-8be89d2e4589" (UID: "b6e72ee2-50e4-4643-aee6-8be89d2e4589"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.626486 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7swjp\" (UniqueName: \"kubernetes.io/projected/a43c28e0-8883-4567-a86f-bcd311098876-kube-api-access-7swjp\") pod \"community-operators-4j8hk\" (UID: \"a43c28e0-8883-4567-a86f-bcd311098876\") " pod="openshift-marketplace/community-operators-4j8hk" Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.626546 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a43c28e0-8883-4567-a86f-bcd311098876-utilities\") pod \"community-operators-4j8hk\" (UID: \"a43c28e0-8883-4567-a86f-bcd311098876\") " pod="openshift-marketplace/community-operators-4j8hk" Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.626586 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a43c28e0-8883-4567-a86f-bcd311098876-catalog-content\") pod \"community-operators-4j8hk\" (UID: \"a43c28e0-8883-4567-a86f-bcd311098876\") " pod="openshift-marketplace/community-operators-4j8hk" Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.626629 4821 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6e72ee2-50e4-4643-aee6-8be89d2e4589-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.626640 4821 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6e72ee2-50e4-4643-aee6-8be89d2e4589-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.626649 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-njv9h\" (UniqueName: \"kubernetes.io/projected/b6e72ee2-50e4-4643-aee6-8be89d2e4589-kube-api-access-njv9h\") on node \"crc\" DevicePath \"\"" Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.630401 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-w56ql_4b4b52a8-67ad-4c51-86ab-503c75ad2024/registry-server/0.log" Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.630977 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w56ql" Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.727265 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nstfp\" (UniqueName: \"kubernetes.io/projected/4b4b52a8-67ad-4c51-86ab-503c75ad2024-kube-api-access-nstfp\") pod \"4b4b52a8-67ad-4c51-86ab-503c75ad2024\" (UID: \"4b4b52a8-67ad-4c51-86ab-503c75ad2024\") " Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.727388 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b4b52a8-67ad-4c51-86ab-503c75ad2024-utilities\") pod \"4b4b52a8-67ad-4c51-86ab-503c75ad2024\" (UID: \"4b4b52a8-67ad-4c51-86ab-503c75ad2024\") " Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.727423 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b4b52a8-67ad-4c51-86ab-503c75ad2024-catalog-content\") pod \"4b4b52a8-67ad-4c51-86ab-503c75ad2024\" (UID: \"4b4b52a8-67ad-4c51-86ab-503c75ad2024\") " Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.727577 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7swjp\" (UniqueName: \"kubernetes.io/projected/a43c28e0-8883-4567-a86f-bcd311098876-kube-api-access-7swjp\") pod \"community-operators-4j8hk\" (UID: \"a43c28e0-8883-4567-a86f-bcd311098876\") " pod="openshift-marketplace/community-operators-4j8hk" Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.727618 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a43c28e0-8883-4567-a86f-bcd311098876-utilities\") pod \"community-operators-4j8hk\" (UID: \"a43c28e0-8883-4567-a86f-bcd311098876\") " pod="openshift-marketplace/community-operators-4j8hk" Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.727638 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a43c28e0-8883-4567-a86f-bcd311098876-catalog-content\") pod \"community-operators-4j8hk\" (UID: \"a43c28e0-8883-4567-a86f-bcd311098876\") " pod="openshift-marketplace/community-operators-4j8hk" Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.728028 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a43c28e0-8883-4567-a86f-bcd311098876-catalog-content\") pod \"community-operators-4j8hk\" (UID: \"a43c28e0-8883-4567-a86f-bcd311098876\") " pod="openshift-marketplace/community-operators-4j8hk" Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.728182 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4b4b52a8-67ad-4c51-86ab-503c75ad2024-utilities" (OuterVolumeSpecName: "utilities") pod "4b4b52a8-67ad-4c51-86ab-503c75ad2024" (UID: "4b4b52a8-67ad-4c51-86ab-503c75ad2024"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.728263 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a43c28e0-8883-4567-a86f-bcd311098876-utilities\") pod \"community-operators-4j8hk\" (UID: \"a43c28e0-8883-4567-a86f-bcd311098876\") " pod="openshift-marketplace/community-operators-4j8hk" Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.732653 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b4b52a8-67ad-4c51-86ab-503c75ad2024-kube-api-access-nstfp" (OuterVolumeSpecName: "kube-api-access-nstfp") pod "4b4b52a8-67ad-4c51-86ab-503c75ad2024" (UID: "4b4b52a8-67ad-4c51-86ab-503c75ad2024"). InnerVolumeSpecName "kube-api-access-nstfp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.747088 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7swjp\" (UniqueName: \"kubernetes.io/projected/a43c28e0-8883-4567-a86f-bcd311098876-kube-api-access-7swjp\") pod \"community-operators-4j8hk\" (UID: \"a43c28e0-8883-4567-a86f-bcd311098876\") " pod="openshift-marketplace/community-operators-4j8hk" Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.817289 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4b4b52a8-67ad-4c51-86ab-503c75ad2024-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4b4b52a8-67ad-4c51-86ab-503c75ad2024" (UID: "4b4b52a8-67ad-4c51-86ab-503c75ad2024"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.829193 4821 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b4b52a8-67ad-4c51-86ab-503c75ad2024-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.829230 4821 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b4b52a8-67ad-4c51-86ab-503c75ad2024-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.829244 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nstfp\" (UniqueName: \"kubernetes.io/projected/4b4b52a8-67ad-4c51-86ab-503c75ad2024-kube-api-access-nstfp\") on node \"crc\" DevicePath \"\"" Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.843040 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4j8hk" Nov 25 10:36:16 crc kubenswrapper[4821]: I1125 10:36:16.915086 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cgdxc"] Nov 25 10:36:17 crc kubenswrapper[4821]: I1125 10:36:17.210598 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4j8hk"] Nov 25 10:36:17 crc kubenswrapper[4821]: I1125 10:36:17.335448 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4j8hk" event={"ID":"a43c28e0-8883-4567-a86f-bcd311098876","Type":"ContainerStarted","Data":"6af501032b9069ab2a65a1c2e49c5bd7c4127db3448f32a8fe250ed7cc3f2340"} Nov 25 10:36:17 crc kubenswrapper[4821]: I1125 10:36:17.339024 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-cgdxc_b6e72ee2-50e4-4643-aee6-8be89d2e4589/registry-server/0.log" Nov 25 10:36:17 crc kubenswrapper[4821]: I1125 10:36:17.339944 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cgdxc" event={"ID":"b6e72ee2-50e4-4643-aee6-8be89d2e4589","Type":"ContainerDied","Data":"7819efc6ad8867915226084d6f5e57ab5eaf307e7406719991c49c0364438b7e"} Nov 25 10:36:17 crc kubenswrapper[4821]: I1125 10:36:17.339990 4821 scope.go:117] "RemoveContainer" containerID="07827fbd586d8e165370290e88d9ac62faf240ebe455d8fc2a232225c32e14e0" Nov 25 10:36:17 crc kubenswrapper[4821]: I1125 10:36:17.339957 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cgdxc" Nov 25 10:36:17 crc kubenswrapper[4821]: I1125 10:36:17.341432 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-w56ql_4b4b52a8-67ad-4c51-86ab-503c75ad2024/registry-server/0.log" Nov 25 10:36:17 crc kubenswrapper[4821]: I1125 10:36:17.342257 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w56ql" event={"ID":"4b4b52a8-67ad-4c51-86ab-503c75ad2024","Type":"ContainerDied","Data":"e34dfb84c2952d3e16b75ad704ace4aa1e336f2eef64d1f275e8442e499a3df2"} Nov 25 10:36:17 crc kubenswrapper[4821]: I1125 10:36:17.342304 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w56ql" Nov 25 10:36:17 crc kubenswrapper[4821]: I1125 10:36:17.354670 4821 scope.go:117] "RemoveContainer" containerID="7b46349b24b5354cf00d29385827cea9614ea973f6654937bc0aabb6c2b529f2" Nov 25 10:36:17 crc kubenswrapper[4821]: I1125 10:36:17.371127 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cgdxc"] Nov 25 10:36:17 crc kubenswrapper[4821]: I1125 10:36:17.373678 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-cgdxc"] Nov 25 10:36:17 crc kubenswrapper[4821]: I1125 10:36:17.383568 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-w56ql"] Nov 25 10:36:17 crc kubenswrapper[4821]: I1125 10:36:17.389648 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-w56ql"] Nov 25 10:36:17 crc kubenswrapper[4821]: I1125 10:36:17.391372 4821 scope.go:117] "RemoveContainer" containerID="d4a3c9bae38e7ab88e8c11dc78a97a4907de72007e8373aded9b9c84c97c41b5" Nov 25 10:36:17 crc kubenswrapper[4821]: I1125 10:36:17.404026 4821 scope.go:117] "RemoveContainer" containerID="51fe6850e14ffc694a9db8f49ec8649f913025a36e30f3269bd3e251d416726a" Nov 25 10:36:17 crc kubenswrapper[4821]: I1125 10:36:17.421572 4821 scope.go:117] "RemoveContainer" containerID="9d16b72aaa2431617a2e647a9a8fb1561bf3d618b0823b487f339531f5284952" Nov 25 10:36:17 crc kubenswrapper[4821]: I1125 10:36:17.440332 4821 scope.go:117] "RemoveContainer" containerID="13ba53a1936435bd0bad9ff1979f3e142b0e1305c98c05123dacf5cd2ece9889" Nov 25 10:36:17 crc kubenswrapper[4821]: I1125 10:36:17.750284 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-r5nmb_e1391070-3aad-47c3-a2f1-0c1697b014bf/registry-server/0.log" Nov 25 10:36:17 crc kubenswrapper[4821]: I1125 10:36:17.751220 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r5nmb" Nov 25 10:36:17 crc kubenswrapper[4821]: I1125 10:36:17.844410 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1391070-3aad-47c3-a2f1-0c1697b014bf-catalog-content\") pod \"e1391070-3aad-47c3-a2f1-0c1697b014bf\" (UID: \"e1391070-3aad-47c3-a2f1-0c1697b014bf\") " Nov 25 10:36:17 crc kubenswrapper[4821]: I1125 10:36:17.844516 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f6dp4\" (UniqueName: \"kubernetes.io/projected/e1391070-3aad-47c3-a2f1-0c1697b014bf-kube-api-access-f6dp4\") pod \"e1391070-3aad-47c3-a2f1-0c1697b014bf\" (UID: \"e1391070-3aad-47c3-a2f1-0c1697b014bf\") " Nov 25 10:36:17 crc kubenswrapper[4821]: I1125 10:36:17.844715 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1391070-3aad-47c3-a2f1-0c1697b014bf-utilities\") pod \"e1391070-3aad-47c3-a2f1-0c1697b014bf\" (UID: \"e1391070-3aad-47c3-a2f1-0c1697b014bf\") " Nov 25 10:36:17 crc kubenswrapper[4821]: I1125 10:36:17.845579 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1391070-3aad-47c3-a2f1-0c1697b014bf-utilities" (OuterVolumeSpecName: "utilities") pod "e1391070-3aad-47c3-a2f1-0c1697b014bf" (UID: "e1391070-3aad-47c3-a2f1-0c1697b014bf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:36:17 crc kubenswrapper[4821]: I1125 10:36:17.850075 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1391070-3aad-47c3-a2f1-0c1697b014bf-kube-api-access-f6dp4" (OuterVolumeSpecName: "kube-api-access-f6dp4") pod "e1391070-3aad-47c3-a2f1-0c1697b014bf" (UID: "e1391070-3aad-47c3-a2f1-0c1697b014bf"). InnerVolumeSpecName "kube-api-access-f6dp4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:36:17 crc kubenswrapper[4821]: I1125 10:36:17.889227 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1391070-3aad-47c3-a2f1-0c1697b014bf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e1391070-3aad-47c3-a2f1-0c1697b014bf" (UID: "e1391070-3aad-47c3-a2f1-0c1697b014bf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:36:17 crc kubenswrapper[4821]: I1125 10:36:17.946008 4821 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1391070-3aad-47c3-a2f1-0c1697b014bf-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:36:17 crc kubenswrapper[4821]: I1125 10:36:17.946035 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f6dp4\" (UniqueName: \"kubernetes.io/projected/e1391070-3aad-47c3-a2f1-0c1697b014bf-kube-api-access-f6dp4\") on node \"crc\" DevicePath \"\"" Nov 25 10:36:17 crc kubenswrapper[4821]: I1125 10:36:17.946048 4821 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1391070-3aad-47c3-a2f1-0c1697b014bf-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.043694 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2kwcp_927e7d40-852e-4807-bb22-d4f7c6e9d2ca/registry-server/0.log" Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.044789 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2kwcp" Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.122769 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-8qpmf_d8becff5-8017-4073-89ae-c17cc4c32922/registry-server/0.log" Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.123063 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b4b52a8-67ad-4c51-86ab-503c75ad2024" path="/var/lib/kubelet/pods/4b4b52a8-67ad-4c51-86ab-503c75ad2024/volumes" Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.123652 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8qpmf" Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.123818 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="645993fd-bdce-435e-a600-34bf1ff5b38b" path="/var/lib/kubelet/pods/645993fd-bdce-435e-a600-34bf1ff5b38b/volumes" Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.124366 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6e72ee2-50e4-4643-aee6-8be89d2e4589" path="/var/lib/kubelet/pods/b6e72ee2-50e4-4643-aee6-8be89d2e4589/volumes" Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.148521 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/927e7d40-852e-4807-bb22-d4f7c6e9d2ca-utilities\") pod \"927e7d40-852e-4807-bb22-d4f7c6e9d2ca\" (UID: \"927e7d40-852e-4807-bb22-d4f7c6e9d2ca\") " Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.148621 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/927e7d40-852e-4807-bb22-d4f7c6e9d2ca-catalog-content\") pod \"927e7d40-852e-4807-bb22-d4f7c6e9d2ca\" (UID: \"927e7d40-852e-4807-bb22-d4f7c6e9d2ca\") " Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.148654 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vlpdz\" (UniqueName: \"kubernetes.io/projected/927e7d40-852e-4807-bb22-d4f7c6e9d2ca-kube-api-access-vlpdz\") pod \"927e7d40-852e-4807-bb22-d4f7c6e9d2ca\" (UID: \"927e7d40-852e-4807-bb22-d4f7c6e9d2ca\") " Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.149810 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/927e7d40-852e-4807-bb22-d4f7c6e9d2ca-utilities" (OuterVolumeSpecName: "utilities") pod "927e7d40-852e-4807-bb22-d4f7c6e9d2ca" (UID: "927e7d40-852e-4807-bb22-d4f7c6e9d2ca"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.164736 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/927e7d40-852e-4807-bb22-d4f7c6e9d2ca-kube-api-access-vlpdz" (OuterVolumeSpecName: "kube-api-access-vlpdz") pod "927e7d40-852e-4807-bb22-d4f7c6e9d2ca" (UID: "927e7d40-852e-4807-bb22-d4f7c6e9d2ca"). InnerVolumeSpecName "kube-api-access-vlpdz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.170375 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/927e7d40-852e-4807-bb22-d4f7c6e9d2ca-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "927e7d40-852e-4807-bb22-d4f7c6e9d2ca" (UID: "927e7d40-852e-4807-bb22-d4f7c6e9d2ca"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.249880 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8becff5-8017-4073-89ae-c17cc4c32922-catalog-content\") pod \"d8becff5-8017-4073-89ae-c17cc4c32922\" (UID: \"d8becff5-8017-4073-89ae-c17cc4c32922\") " Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.265672 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8becff5-8017-4073-89ae-c17cc4c32922-utilities\") pod \"d8becff5-8017-4073-89ae-c17cc4c32922\" (UID: \"d8becff5-8017-4073-89ae-c17cc4c32922\") " Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.265727 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gfw7p\" (UniqueName: \"kubernetes.io/projected/d8becff5-8017-4073-89ae-c17cc4c32922-kube-api-access-gfw7p\") pod \"d8becff5-8017-4073-89ae-c17cc4c32922\" (UID: \"d8becff5-8017-4073-89ae-c17cc4c32922\") " Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.266144 4821 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/927e7d40-852e-4807-bb22-d4f7c6e9d2ca-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.266185 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vlpdz\" (UniqueName: \"kubernetes.io/projected/927e7d40-852e-4807-bb22-d4f7c6e9d2ca-kube-api-access-vlpdz\") on node \"crc\" DevicePath \"\"" Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.266198 4821 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/927e7d40-852e-4807-bb22-d4f7c6e9d2ca-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.266472 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d8becff5-8017-4073-89ae-c17cc4c32922-utilities" (OuterVolumeSpecName: "utilities") pod "d8becff5-8017-4073-89ae-c17cc4c32922" (UID: "d8becff5-8017-4073-89ae-c17cc4c32922"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.266629 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d8becff5-8017-4073-89ae-c17cc4c32922-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d8becff5-8017-4073-89ae-c17cc4c32922" (UID: "d8becff5-8017-4073-89ae-c17cc4c32922"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.268135 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8becff5-8017-4073-89ae-c17cc4c32922-kube-api-access-gfw7p" (OuterVolumeSpecName: "kube-api-access-gfw7p") pod "d8becff5-8017-4073-89ae-c17cc4c32922" (UID: "d8becff5-8017-4073-89ae-c17cc4c32922"). InnerVolumeSpecName "kube-api-access-gfw7p". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.349975 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2kwcp_927e7d40-852e-4807-bb22-d4f7c6e9d2ca/registry-server/0.log" Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.350708 4821 generic.go:334] "Generic (PLEG): container finished" podID="927e7d40-852e-4807-bb22-d4f7c6e9d2ca" containerID="42feaa125fff6b12b55e21ca82b443e338cf19fd8d839e42cd6ddb18de30a91a" exitCode=1 Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.350777 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2kwcp" Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.350792 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2kwcp" event={"ID":"927e7d40-852e-4807-bb22-d4f7c6e9d2ca","Type":"ContainerDied","Data":"42feaa125fff6b12b55e21ca82b443e338cf19fd8d839e42cd6ddb18de30a91a"} Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.350834 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2kwcp" event={"ID":"927e7d40-852e-4807-bb22-d4f7c6e9d2ca","Type":"ContainerDied","Data":"23cd67b4ab339669af677f251279be6c3db37de80867eaede62cde905682b7ad"} Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.350856 4821 scope.go:117] "RemoveContainer" containerID="42feaa125fff6b12b55e21ca82b443e338cf19fd8d839e42cd6ddb18de30a91a" Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.351951 4821 generic.go:334] "Generic (PLEG): container finished" podID="a43c28e0-8883-4567-a86f-bcd311098876" containerID="7d771e32b7069f177d67c0d886c7fa23b7a91926741fbc3df48e94233d4246f9" exitCode=0 Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.352018 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4j8hk" event={"ID":"a43c28e0-8883-4567-a86f-bcd311098876","Type":"ContainerDied","Data":"7d771e32b7069f177d67c0d886c7fa23b7a91926741fbc3df48e94233d4246f9"} Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.356639 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-8qpmf_d8becff5-8017-4073-89ae-c17cc4c32922/registry-server/0.log" Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.359018 4821 generic.go:334] "Generic (PLEG): container finished" podID="d8becff5-8017-4073-89ae-c17cc4c32922" containerID="d9cebe2065a7afab486b6363583a5992907042ff24a825458f0d0a35a8da9eeb" exitCode=1 Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.359083 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8qpmf" event={"ID":"d8becff5-8017-4073-89ae-c17cc4c32922","Type":"ContainerDied","Data":"d9cebe2065a7afab486b6363583a5992907042ff24a825458f0d0a35a8da9eeb"} Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.359111 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8qpmf" event={"ID":"d8becff5-8017-4073-89ae-c17cc4c32922","Type":"ContainerDied","Data":"0262a0c2a5dbf2dff7fb7034ccd1c965b0de806bc73a49b7084bd6ac6cae9918"} Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.359218 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8qpmf" Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.367044 4821 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8becff5-8017-4073-89ae-c17cc4c32922-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.367086 4821 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8becff5-8017-4073-89ae-c17cc4c32922-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.367100 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gfw7p\" (UniqueName: \"kubernetes.io/projected/d8becff5-8017-4073-89ae-c17cc4c32922-kube-api-access-gfw7p\") on node \"crc\" DevicePath \"\"" Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.370899 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-r5nmb_e1391070-3aad-47c3-a2f1-0c1697b014bf/registry-server/0.log" Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.375537 4821 generic.go:334] "Generic (PLEG): container finished" podID="e1391070-3aad-47c3-a2f1-0c1697b014bf" containerID="0ab5237249ce6715dff6a126e5d1583e88387a901b75476727dc44baddeef332" exitCode=1 Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.376840 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r5nmb" Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.377175 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r5nmb" event={"ID":"e1391070-3aad-47c3-a2f1-0c1697b014bf","Type":"ContainerDied","Data":"0ab5237249ce6715dff6a126e5d1583e88387a901b75476727dc44baddeef332"} Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.377211 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r5nmb" event={"ID":"e1391070-3aad-47c3-a2f1-0c1697b014bf","Type":"ContainerDied","Data":"00c07aed46d32e6e7ec0b81b24b2b64f7f0828080061143f799ce995c5d37992"} Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.377385 4821 scope.go:117] "RemoveContainer" containerID="39e4fa98d1d04e9e87bcd354abd8aeaaecdf77816239d00c555fdc9aee8c652c" Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.399628 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2kwcp"] Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.404722 4821 scope.go:117] "RemoveContainer" containerID="977f93d972d776b41806449e65352f20756f6a5350d48e77262c8f4e3e6a19e2" Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.415113 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-2kwcp"] Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.423029 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-r5nmb"] Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.437990 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-r5nmb"] Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.439138 4821 scope.go:117] "RemoveContainer" containerID="42feaa125fff6b12b55e21ca82b443e338cf19fd8d839e42cd6ddb18de30a91a" Nov 25 10:36:18 crc kubenswrapper[4821]: E1125 10:36:18.439948 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42feaa125fff6b12b55e21ca82b443e338cf19fd8d839e42cd6ddb18de30a91a\": container with ID starting with 42feaa125fff6b12b55e21ca82b443e338cf19fd8d839e42cd6ddb18de30a91a not found: ID does not exist" containerID="42feaa125fff6b12b55e21ca82b443e338cf19fd8d839e42cd6ddb18de30a91a" Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.440001 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42feaa125fff6b12b55e21ca82b443e338cf19fd8d839e42cd6ddb18de30a91a"} err="failed to get container status \"42feaa125fff6b12b55e21ca82b443e338cf19fd8d839e42cd6ddb18de30a91a\": rpc error: code = NotFound desc = could not find container \"42feaa125fff6b12b55e21ca82b443e338cf19fd8d839e42cd6ddb18de30a91a\": container with ID starting with 42feaa125fff6b12b55e21ca82b443e338cf19fd8d839e42cd6ddb18de30a91a not found: ID does not exist" Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.440034 4821 scope.go:117] "RemoveContainer" containerID="39e4fa98d1d04e9e87bcd354abd8aeaaecdf77816239d00c555fdc9aee8c652c" Nov 25 10:36:18 crc kubenswrapper[4821]: E1125 10:36:18.440730 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39e4fa98d1d04e9e87bcd354abd8aeaaecdf77816239d00c555fdc9aee8c652c\": container with ID starting with 39e4fa98d1d04e9e87bcd354abd8aeaaecdf77816239d00c555fdc9aee8c652c not found: ID does not exist" containerID="39e4fa98d1d04e9e87bcd354abd8aeaaecdf77816239d00c555fdc9aee8c652c" Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.440760 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39e4fa98d1d04e9e87bcd354abd8aeaaecdf77816239d00c555fdc9aee8c652c"} err="failed to get container status \"39e4fa98d1d04e9e87bcd354abd8aeaaecdf77816239d00c555fdc9aee8c652c\": rpc error: code = NotFound desc = could not find container \"39e4fa98d1d04e9e87bcd354abd8aeaaecdf77816239d00c555fdc9aee8c652c\": container with ID starting with 39e4fa98d1d04e9e87bcd354abd8aeaaecdf77816239d00c555fdc9aee8c652c not found: ID does not exist" Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.440797 4821 scope.go:117] "RemoveContainer" containerID="977f93d972d776b41806449e65352f20756f6a5350d48e77262c8f4e3e6a19e2" Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.441286 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8qpmf"] Nov 25 10:36:18 crc kubenswrapper[4821]: E1125 10:36:18.441286 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"977f93d972d776b41806449e65352f20756f6a5350d48e77262c8f4e3e6a19e2\": container with ID starting with 977f93d972d776b41806449e65352f20756f6a5350d48e77262c8f4e3e6a19e2 not found: ID does not exist" containerID="977f93d972d776b41806449e65352f20756f6a5350d48e77262c8f4e3e6a19e2" Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.441322 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"977f93d972d776b41806449e65352f20756f6a5350d48e77262c8f4e3e6a19e2"} err="failed to get container status \"977f93d972d776b41806449e65352f20756f6a5350d48e77262c8f4e3e6a19e2\": rpc error: code = NotFound desc = could not find container \"977f93d972d776b41806449e65352f20756f6a5350d48e77262c8f4e3e6a19e2\": container with ID starting with 977f93d972d776b41806449e65352f20756f6a5350d48e77262c8f4e3e6a19e2 not found: ID does not exist" Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.441353 4821 scope.go:117] "RemoveContainer" containerID="d9cebe2065a7afab486b6363583a5992907042ff24a825458f0d0a35a8da9eeb" Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.445625 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-8qpmf"] Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.464735 4821 scope.go:117] "RemoveContainer" containerID="7588e12b27ca0da7eeb1409085d529ba99aeb108a9c5a525074375a0ed04e76c" Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.476150 4821 scope.go:117] "RemoveContainer" containerID="47a1a83c452d0edae993c649647470792cb7de9dafb21beba9dbd623ec1d6852" Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.490957 4821 scope.go:117] "RemoveContainer" containerID="d9cebe2065a7afab486b6363583a5992907042ff24a825458f0d0a35a8da9eeb" Nov 25 10:36:18 crc kubenswrapper[4821]: E1125 10:36:18.491522 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9cebe2065a7afab486b6363583a5992907042ff24a825458f0d0a35a8da9eeb\": container with ID starting with d9cebe2065a7afab486b6363583a5992907042ff24a825458f0d0a35a8da9eeb not found: ID does not exist" containerID="d9cebe2065a7afab486b6363583a5992907042ff24a825458f0d0a35a8da9eeb" Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.491552 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9cebe2065a7afab486b6363583a5992907042ff24a825458f0d0a35a8da9eeb"} err="failed to get container status \"d9cebe2065a7afab486b6363583a5992907042ff24a825458f0d0a35a8da9eeb\": rpc error: code = NotFound desc = could not find container \"d9cebe2065a7afab486b6363583a5992907042ff24a825458f0d0a35a8da9eeb\": container with ID starting with d9cebe2065a7afab486b6363583a5992907042ff24a825458f0d0a35a8da9eeb not found: ID does not exist" Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.491573 4821 scope.go:117] "RemoveContainer" containerID="7588e12b27ca0da7eeb1409085d529ba99aeb108a9c5a525074375a0ed04e76c" Nov 25 10:36:18 crc kubenswrapper[4821]: E1125 10:36:18.491816 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7588e12b27ca0da7eeb1409085d529ba99aeb108a9c5a525074375a0ed04e76c\": container with ID starting with 7588e12b27ca0da7eeb1409085d529ba99aeb108a9c5a525074375a0ed04e76c not found: ID does not exist" containerID="7588e12b27ca0da7eeb1409085d529ba99aeb108a9c5a525074375a0ed04e76c" Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.491838 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7588e12b27ca0da7eeb1409085d529ba99aeb108a9c5a525074375a0ed04e76c"} err="failed to get container status \"7588e12b27ca0da7eeb1409085d529ba99aeb108a9c5a525074375a0ed04e76c\": rpc error: code = NotFound desc = could not find container \"7588e12b27ca0da7eeb1409085d529ba99aeb108a9c5a525074375a0ed04e76c\": container with ID starting with 7588e12b27ca0da7eeb1409085d529ba99aeb108a9c5a525074375a0ed04e76c not found: ID does not exist" Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.491849 4821 scope.go:117] "RemoveContainer" containerID="47a1a83c452d0edae993c649647470792cb7de9dafb21beba9dbd623ec1d6852" Nov 25 10:36:18 crc kubenswrapper[4821]: E1125 10:36:18.492089 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"47a1a83c452d0edae993c649647470792cb7de9dafb21beba9dbd623ec1d6852\": container with ID starting with 47a1a83c452d0edae993c649647470792cb7de9dafb21beba9dbd623ec1d6852 not found: ID does not exist" containerID="47a1a83c452d0edae993c649647470792cb7de9dafb21beba9dbd623ec1d6852" Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.492212 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47a1a83c452d0edae993c649647470792cb7de9dafb21beba9dbd623ec1d6852"} err="failed to get container status \"47a1a83c452d0edae993c649647470792cb7de9dafb21beba9dbd623ec1d6852\": rpc error: code = NotFound desc = could not find container \"47a1a83c452d0edae993c649647470792cb7de9dafb21beba9dbd623ec1d6852\": container with ID starting with 47a1a83c452d0edae993c649647470792cb7de9dafb21beba9dbd623ec1d6852 not found: ID does not exist" Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.492301 4821 scope.go:117] "RemoveContainer" containerID="0ab5237249ce6715dff6a126e5d1583e88387a901b75476727dc44baddeef332" Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.505028 4821 scope.go:117] "RemoveContainer" containerID="bea8c804c37d876eeb33513f8d24ec7d2523e8a305c58c337aca7e88247d7ea9" Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.519992 4821 scope.go:117] "RemoveContainer" containerID="b4a2f4f11681121adddba94e268de602497790e7562df7b11e62b1660e19ebe2" Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.531512 4821 scope.go:117] "RemoveContainer" containerID="0ab5237249ce6715dff6a126e5d1583e88387a901b75476727dc44baddeef332" Nov 25 10:36:18 crc kubenswrapper[4821]: E1125 10:36:18.531908 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ab5237249ce6715dff6a126e5d1583e88387a901b75476727dc44baddeef332\": container with ID starting with 0ab5237249ce6715dff6a126e5d1583e88387a901b75476727dc44baddeef332 not found: ID does not exist" containerID="0ab5237249ce6715dff6a126e5d1583e88387a901b75476727dc44baddeef332" Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.531949 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ab5237249ce6715dff6a126e5d1583e88387a901b75476727dc44baddeef332"} err="failed to get container status \"0ab5237249ce6715dff6a126e5d1583e88387a901b75476727dc44baddeef332\": rpc error: code = NotFound desc = could not find container \"0ab5237249ce6715dff6a126e5d1583e88387a901b75476727dc44baddeef332\": container with ID starting with 0ab5237249ce6715dff6a126e5d1583e88387a901b75476727dc44baddeef332 not found: ID does not exist" Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.531982 4821 scope.go:117] "RemoveContainer" containerID="bea8c804c37d876eeb33513f8d24ec7d2523e8a305c58c337aca7e88247d7ea9" Nov 25 10:36:18 crc kubenswrapper[4821]: E1125 10:36:18.532220 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bea8c804c37d876eeb33513f8d24ec7d2523e8a305c58c337aca7e88247d7ea9\": container with ID starting with bea8c804c37d876eeb33513f8d24ec7d2523e8a305c58c337aca7e88247d7ea9 not found: ID does not exist" containerID="bea8c804c37d876eeb33513f8d24ec7d2523e8a305c58c337aca7e88247d7ea9" Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.532247 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bea8c804c37d876eeb33513f8d24ec7d2523e8a305c58c337aca7e88247d7ea9"} err="failed to get container status \"bea8c804c37d876eeb33513f8d24ec7d2523e8a305c58c337aca7e88247d7ea9\": rpc error: code = NotFound desc = could not find container \"bea8c804c37d876eeb33513f8d24ec7d2523e8a305c58c337aca7e88247d7ea9\": container with ID starting with bea8c804c37d876eeb33513f8d24ec7d2523e8a305c58c337aca7e88247d7ea9 not found: ID does not exist" Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.532261 4821 scope.go:117] "RemoveContainer" containerID="b4a2f4f11681121adddba94e268de602497790e7562df7b11e62b1660e19ebe2" Nov 25 10:36:18 crc kubenswrapper[4821]: E1125 10:36:18.532428 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b4a2f4f11681121adddba94e268de602497790e7562df7b11e62b1660e19ebe2\": container with ID starting with b4a2f4f11681121adddba94e268de602497790e7562df7b11e62b1660e19ebe2 not found: ID does not exist" containerID="b4a2f4f11681121adddba94e268de602497790e7562df7b11e62b1660e19ebe2" Nov 25 10:36:18 crc kubenswrapper[4821]: I1125 10:36:18.532444 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4a2f4f11681121adddba94e268de602497790e7562df7b11e62b1660e19ebe2"} err="failed to get container status \"b4a2f4f11681121adddba94e268de602497790e7562df7b11e62b1660e19ebe2\": rpc error: code = NotFound desc = could not find container \"b4a2f4f11681121adddba94e268de602497790e7562df7b11e62b1660e19ebe2\": container with ID starting with b4a2f4f11681121adddba94e268de602497790e7562df7b11e62b1660e19ebe2 not found: ID does not exist" Nov 25 10:36:19 crc kubenswrapper[4821]: I1125 10:36:19.319998 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-g4r76"] Nov 25 10:36:19 crc kubenswrapper[4821]: E1125 10:36:19.320197 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b4b52a8-67ad-4c51-86ab-503c75ad2024" containerName="registry-server" Nov 25 10:36:19 crc kubenswrapper[4821]: I1125 10:36:19.320210 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b4b52a8-67ad-4c51-86ab-503c75ad2024" containerName="registry-server" Nov 25 10:36:19 crc kubenswrapper[4821]: E1125 10:36:19.320221 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1391070-3aad-47c3-a2f1-0c1697b014bf" containerName="registry-server" Nov 25 10:36:19 crc kubenswrapper[4821]: I1125 10:36:19.320229 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1391070-3aad-47c3-a2f1-0c1697b014bf" containerName="registry-server" Nov 25 10:36:19 crc kubenswrapper[4821]: E1125 10:36:19.320241 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b4b52a8-67ad-4c51-86ab-503c75ad2024" containerName="extract-utilities" Nov 25 10:36:19 crc kubenswrapper[4821]: I1125 10:36:19.320247 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b4b52a8-67ad-4c51-86ab-503c75ad2024" containerName="extract-utilities" Nov 25 10:36:19 crc kubenswrapper[4821]: E1125 10:36:19.320257 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1391070-3aad-47c3-a2f1-0c1697b014bf" containerName="extract-content" Nov 25 10:36:19 crc kubenswrapper[4821]: I1125 10:36:19.320264 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1391070-3aad-47c3-a2f1-0c1697b014bf" containerName="extract-content" Nov 25 10:36:19 crc kubenswrapper[4821]: E1125 10:36:19.320272 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1391070-3aad-47c3-a2f1-0c1697b014bf" containerName="extract-utilities" Nov 25 10:36:19 crc kubenswrapper[4821]: I1125 10:36:19.320279 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1391070-3aad-47c3-a2f1-0c1697b014bf" containerName="extract-utilities" Nov 25 10:36:19 crc kubenswrapper[4821]: E1125 10:36:19.320285 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="927e7d40-852e-4807-bb22-d4f7c6e9d2ca" containerName="extract-utilities" Nov 25 10:36:19 crc kubenswrapper[4821]: I1125 10:36:19.320326 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="927e7d40-852e-4807-bb22-d4f7c6e9d2ca" containerName="extract-utilities" Nov 25 10:36:19 crc kubenswrapper[4821]: E1125 10:36:19.320333 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="927e7d40-852e-4807-bb22-d4f7c6e9d2ca" containerName="registry-server" Nov 25 10:36:19 crc kubenswrapper[4821]: I1125 10:36:19.320339 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="927e7d40-852e-4807-bb22-d4f7c6e9d2ca" containerName="registry-server" Nov 25 10:36:19 crc kubenswrapper[4821]: E1125 10:36:19.320376 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8becff5-8017-4073-89ae-c17cc4c32922" containerName="extract-utilities" Nov 25 10:36:19 crc kubenswrapper[4821]: I1125 10:36:19.320383 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8becff5-8017-4073-89ae-c17cc4c32922" containerName="extract-utilities" Nov 25 10:36:19 crc kubenswrapper[4821]: E1125 10:36:19.320392 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="927e7d40-852e-4807-bb22-d4f7c6e9d2ca" containerName="extract-content" Nov 25 10:36:19 crc kubenswrapper[4821]: I1125 10:36:19.320398 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="927e7d40-852e-4807-bb22-d4f7c6e9d2ca" containerName="extract-content" Nov 25 10:36:19 crc kubenswrapper[4821]: E1125 10:36:19.320408 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8becff5-8017-4073-89ae-c17cc4c32922" containerName="registry-server" Nov 25 10:36:19 crc kubenswrapper[4821]: I1125 10:36:19.320414 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8becff5-8017-4073-89ae-c17cc4c32922" containerName="registry-server" Nov 25 10:36:19 crc kubenswrapper[4821]: E1125 10:36:19.320426 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8becff5-8017-4073-89ae-c17cc4c32922" containerName="extract-content" Nov 25 10:36:19 crc kubenswrapper[4821]: I1125 10:36:19.320432 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8becff5-8017-4073-89ae-c17cc4c32922" containerName="extract-content" Nov 25 10:36:19 crc kubenswrapper[4821]: E1125 10:36:19.320438 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b4b52a8-67ad-4c51-86ab-503c75ad2024" containerName="extract-content" Nov 25 10:36:19 crc kubenswrapper[4821]: I1125 10:36:19.320444 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b4b52a8-67ad-4c51-86ab-503c75ad2024" containerName="extract-content" Nov 25 10:36:19 crc kubenswrapper[4821]: I1125 10:36:19.320538 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b4b52a8-67ad-4c51-86ab-503c75ad2024" containerName="registry-server" Nov 25 10:36:19 crc kubenswrapper[4821]: I1125 10:36:19.320551 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8becff5-8017-4073-89ae-c17cc4c32922" containerName="registry-server" Nov 25 10:36:19 crc kubenswrapper[4821]: I1125 10:36:19.320559 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="927e7d40-852e-4807-bb22-d4f7c6e9d2ca" containerName="registry-server" Nov 25 10:36:19 crc kubenswrapper[4821]: I1125 10:36:19.320566 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1391070-3aad-47c3-a2f1-0c1697b014bf" containerName="registry-server" Nov 25 10:36:19 crc kubenswrapper[4821]: I1125 10:36:19.321435 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g4r76" Nov 25 10:36:19 crc kubenswrapper[4821]: I1125 10:36:19.325921 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 25 10:36:19 crc kubenswrapper[4821]: I1125 10:36:19.333532 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-g4r76"] Nov 25 10:36:19 crc kubenswrapper[4821]: I1125 10:36:19.392855 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4j8hk" event={"ID":"a43c28e0-8883-4567-a86f-bcd311098876","Type":"ContainerStarted","Data":"05c2329fbef9e93530d42eddf45e76eef02a4548d2ae6a24cfdc13aa61c6a005"} Nov 25 10:36:19 crc kubenswrapper[4821]: I1125 10:36:19.480154 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/268c8582-b9a0-4e22-ac84-e47a6fa3f284-catalog-content\") pod \"certified-operators-g4r76\" (UID: \"268c8582-b9a0-4e22-ac84-e47a6fa3f284\") " pod="openshift-marketplace/certified-operators-g4r76" Nov 25 10:36:19 crc kubenswrapper[4821]: I1125 10:36:19.480276 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/268c8582-b9a0-4e22-ac84-e47a6fa3f284-utilities\") pod \"certified-operators-g4r76\" (UID: \"268c8582-b9a0-4e22-ac84-e47a6fa3f284\") " pod="openshift-marketplace/certified-operators-g4r76" Nov 25 10:36:19 crc kubenswrapper[4821]: I1125 10:36:19.480314 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7lnv\" (UniqueName: \"kubernetes.io/projected/268c8582-b9a0-4e22-ac84-e47a6fa3f284-kube-api-access-g7lnv\") pod \"certified-operators-g4r76\" (UID: \"268c8582-b9a0-4e22-ac84-e47a6fa3f284\") " pod="openshift-marketplace/certified-operators-g4r76" Nov 25 10:36:19 crc kubenswrapper[4821]: I1125 10:36:19.582114 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/268c8582-b9a0-4e22-ac84-e47a6fa3f284-catalog-content\") pod \"certified-operators-g4r76\" (UID: \"268c8582-b9a0-4e22-ac84-e47a6fa3f284\") " pod="openshift-marketplace/certified-operators-g4r76" Nov 25 10:36:19 crc kubenswrapper[4821]: I1125 10:36:19.582277 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/268c8582-b9a0-4e22-ac84-e47a6fa3f284-utilities\") pod \"certified-operators-g4r76\" (UID: \"268c8582-b9a0-4e22-ac84-e47a6fa3f284\") " pod="openshift-marketplace/certified-operators-g4r76" Nov 25 10:36:19 crc kubenswrapper[4821]: I1125 10:36:19.582318 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g7lnv\" (UniqueName: \"kubernetes.io/projected/268c8582-b9a0-4e22-ac84-e47a6fa3f284-kube-api-access-g7lnv\") pod \"certified-operators-g4r76\" (UID: \"268c8582-b9a0-4e22-ac84-e47a6fa3f284\") " pod="openshift-marketplace/certified-operators-g4r76" Nov 25 10:36:19 crc kubenswrapper[4821]: I1125 10:36:19.582626 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/268c8582-b9a0-4e22-ac84-e47a6fa3f284-catalog-content\") pod \"certified-operators-g4r76\" (UID: \"268c8582-b9a0-4e22-ac84-e47a6fa3f284\") " pod="openshift-marketplace/certified-operators-g4r76" Nov 25 10:36:19 crc kubenswrapper[4821]: I1125 10:36:19.582665 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/268c8582-b9a0-4e22-ac84-e47a6fa3f284-utilities\") pod \"certified-operators-g4r76\" (UID: \"268c8582-b9a0-4e22-ac84-e47a6fa3f284\") " pod="openshift-marketplace/certified-operators-g4r76" Nov 25 10:36:19 crc kubenswrapper[4821]: I1125 10:36:19.600106 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7lnv\" (UniqueName: \"kubernetes.io/projected/268c8582-b9a0-4e22-ac84-e47a6fa3f284-kube-api-access-g7lnv\") pod \"certified-operators-g4r76\" (UID: \"268c8582-b9a0-4e22-ac84-e47a6fa3f284\") " pod="openshift-marketplace/certified-operators-g4r76" Nov 25 10:36:19 crc kubenswrapper[4821]: I1125 10:36:19.646507 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g4r76" Nov 25 10:36:20 crc kubenswrapper[4821]: I1125 10:36:20.021847 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-g4r76"] Nov 25 10:36:20 crc kubenswrapper[4821]: I1125 10:36:20.121415 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="927e7d40-852e-4807-bb22-d4f7c6e9d2ca" path="/var/lib/kubelet/pods/927e7d40-852e-4807-bb22-d4f7c6e9d2ca/volumes" Nov 25 10:36:20 crc kubenswrapper[4821]: I1125 10:36:20.122367 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8becff5-8017-4073-89ae-c17cc4c32922" path="/var/lib/kubelet/pods/d8becff5-8017-4073-89ae-c17cc4c32922/volumes" Nov 25 10:36:20 crc kubenswrapper[4821]: I1125 10:36:20.122899 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1391070-3aad-47c3-a2f1-0c1697b014bf" path="/var/lib/kubelet/pods/e1391070-3aad-47c3-a2f1-0c1697b014bf/volumes" Nov 25 10:36:20 crc kubenswrapper[4821]: I1125 10:36:20.326352 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-vrjfk"] Nov 25 10:36:20 crc kubenswrapper[4821]: I1125 10:36:20.327533 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vrjfk" Nov 25 10:36:20 crc kubenswrapper[4821]: I1125 10:36:20.332043 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 25 10:36:20 crc kubenswrapper[4821]: I1125 10:36:20.336234 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vrjfk"] Nov 25 10:36:20 crc kubenswrapper[4821]: I1125 10:36:20.404961 4821 generic.go:334] "Generic (PLEG): container finished" podID="a43c28e0-8883-4567-a86f-bcd311098876" containerID="05c2329fbef9e93530d42eddf45e76eef02a4548d2ae6a24cfdc13aa61c6a005" exitCode=0 Nov 25 10:36:20 crc kubenswrapper[4821]: I1125 10:36:20.405067 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4j8hk" event={"ID":"a43c28e0-8883-4567-a86f-bcd311098876","Type":"ContainerDied","Data":"05c2329fbef9e93530d42eddf45e76eef02a4548d2ae6a24cfdc13aa61c6a005"} Nov 25 10:36:20 crc kubenswrapper[4821]: I1125 10:36:20.409384 4821 generic.go:334] "Generic (PLEG): container finished" podID="268c8582-b9a0-4e22-ac84-e47a6fa3f284" containerID="1cc00a4c480de98e11c9fcbe09399ddb90d9a63af828f33af5ae299c987ddc6a" exitCode=0 Nov 25 10:36:20 crc kubenswrapper[4821]: I1125 10:36:20.409536 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g4r76" event={"ID":"268c8582-b9a0-4e22-ac84-e47a6fa3f284","Type":"ContainerDied","Data":"1cc00a4c480de98e11c9fcbe09399ddb90d9a63af828f33af5ae299c987ddc6a"} Nov 25 10:36:20 crc kubenswrapper[4821]: I1125 10:36:20.409563 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g4r76" event={"ID":"268c8582-b9a0-4e22-ac84-e47a6fa3f284","Type":"ContainerStarted","Data":"d3ce50fc235bad9afb3ca250067a103239ab31c884315d7903686e5831ae6e04"} Nov 25 10:36:20 crc kubenswrapper[4821]: I1125 10:36:20.491044 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b01ce52-a13b-431c-8597-fbd5dc25183f-catalog-content\") pod \"redhat-operators-vrjfk\" (UID: \"7b01ce52-a13b-431c-8597-fbd5dc25183f\") " pod="openshift-marketplace/redhat-operators-vrjfk" Nov 25 10:36:20 crc kubenswrapper[4821]: I1125 10:36:20.491106 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b01ce52-a13b-431c-8597-fbd5dc25183f-utilities\") pod \"redhat-operators-vrjfk\" (UID: \"7b01ce52-a13b-431c-8597-fbd5dc25183f\") " pod="openshift-marketplace/redhat-operators-vrjfk" Nov 25 10:36:20 crc kubenswrapper[4821]: I1125 10:36:20.491138 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8kj9\" (UniqueName: \"kubernetes.io/projected/7b01ce52-a13b-431c-8597-fbd5dc25183f-kube-api-access-s8kj9\") pod \"redhat-operators-vrjfk\" (UID: \"7b01ce52-a13b-431c-8597-fbd5dc25183f\") " pod="openshift-marketplace/redhat-operators-vrjfk" Nov 25 10:36:20 crc kubenswrapper[4821]: I1125 10:36:20.592227 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b01ce52-a13b-431c-8597-fbd5dc25183f-catalog-content\") pod \"redhat-operators-vrjfk\" (UID: \"7b01ce52-a13b-431c-8597-fbd5dc25183f\") " pod="openshift-marketplace/redhat-operators-vrjfk" Nov 25 10:36:20 crc kubenswrapper[4821]: I1125 10:36:20.592347 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b01ce52-a13b-431c-8597-fbd5dc25183f-utilities\") pod \"redhat-operators-vrjfk\" (UID: \"7b01ce52-a13b-431c-8597-fbd5dc25183f\") " pod="openshift-marketplace/redhat-operators-vrjfk" Nov 25 10:36:20 crc kubenswrapper[4821]: I1125 10:36:20.592404 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8kj9\" (UniqueName: \"kubernetes.io/projected/7b01ce52-a13b-431c-8597-fbd5dc25183f-kube-api-access-s8kj9\") pod \"redhat-operators-vrjfk\" (UID: \"7b01ce52-a13b-431c-8597-fbd5dc25183f\") " pod="openshift-marketplace/redhat-operators-vrjfk" Nov 25 10:36:20 crc kubenswrapper[4821]: I1125 10:36:20.592692 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b01ce52-a13b-431c-8597-fbd5dc25183f-catalog-content\") pod \"redhat-operators-vrjfk\" (UID: \"7b01ce52-a13b-431c-8597-fbd5dc25183f\") " pod="openshift-marketplace/redhat-operators-vrjfk" Nov 25 10:36:20 crc kubenswrapper[4821]: I1125 10:36:20.592779 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b01ce52-a13b-431c-8597-fbd5dc25183f-utilities\") pod \"redhat-operators-vrjfk\" (UID: \"7b01ce52-a13b-431c-8597-fbd5dc25183f\") " pod="openshift-marketplace/redhat-operators-vrjfk" Nov 25 10:36:20 crc kubenswrapper[4821]: I1125 10:36:20.610216 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8kj9\" (UniqueName: \"kubernetes.io/projected/7b01ce52-a13b-431c-8597-fbd5dc25183f-kube-api-access-s8kj9\") pod \"redhat-operators-vrjfk\" (UID: \"7b01ce52-a13b-431c-8597-fbd5dc25183f\") " pod="openshift-marketplace/redhat-operators-vrjfk" Nov 25 10:36:20 crc kubenswrapper[4821]: I1125 10:36:20.651703 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vrjfk" Nov 25 10:36:21 crc kubenswrapper[4821]: I1125 10:36:21.091027 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vrjfk"] Nov 25 10:36:21 crc kubenswrapper[4821]: I1125 10:36:21.416565 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g4r76" event={"ID":"268c8582-b9a0-4e22-ac84-e47a6fa3f284","Type":"ContainerStarted","Data":"e9d2dc3ae125b4c3a0a7f0a47570f0351d7988bd73e7f68b65fddb861667f227"} Nov 25 10:36:21 crc kubenswrapper[4821]: I1125 10:36:21.419951 4821 generic.go:334] "Generic (PLEG): container finished" podID="7b01ce52-a13b-431c-8597-fbd5dc25183f" containerID="7f00fac22e0cf3ffaf5716d44747e9ea8239f22a2bae85ae8bef7e9428ce8a0f" exitCode=0 Nov 25 10:36:21 crc kubenswrapper[4821]: I1125 10:36:21.420045 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vrjfk" event={"ID":"7b01ce52-a13b-431c-8597-fbd5dc25183f","Type":"ContainerDied","Data":"7f00fac22e0cf3ffaf5716d44747e9ea8239f22a2bae85ae8bef7e9428ce8a0f"} Nov 25 10:36:21 crc kubenswrapper[4821]: I1125 10:36:21.420072 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vrjfk" event={"ID":"7b01ce52-a13b-431c-8597-fbd5dc25183f","Type":"ContainerStarted","Data":"afda5ad4dc2ed3b17044e732cc977c668a32c8f95a394a1a3ea7d0aec0bef38a"} Nov 25 10:36:21 crc kubenswrapper[4821]: I1125 10:36:21.427146 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4j8hk" event={"ID":"a43c28e0-8883-4567-a86f-bcd311098876","Type":"ContainerStarted","Data":"0a042aad8674d29e8ea187e8987a8cb4795e28da9e5f96dd4e192f5165d544f5"} Nov 25 10:36:21 crc kubenswrapper[4821]: I1125 10:36:21.454620 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-4j8hk" podStartSLOduration=2.881245055 podStartE2EDuration="5.454601463s" podCreationTimestamp="2025-11-25 10:36:16 +0000 UTC" firstStartedPulling="2025-11-25 10:36:18.357236951 +0000 UTC m=+248.893556798" lastFinishedPulling="2025-11-25 10:36:20.930593359 +0000 UTC m=+251.466913206" observedRunningTime="2025-11-25 10:36:21.451772196 +0000 UTC m=+251.988092043" watchObservedRunningTime="2025-11-25 10:36:21.454601463 +0000 UTC m=+251.990921320" Nov 25 10:36:22 crc kubenswrapper[4821]: I1125 10:36:22.434637 4821 generic.go:334] "Generic (PLEG): container finished" podID="268c8582-b9a0-4e22-ac84-e47a6fa3f284" containerID="e9d2dc3ae125b4c3a0a7f0a47570f0351d7988bd73e7f68b65fddb861667f227" exitCode=0 Nov 25 10:36:22 crc kubenswrapper[4821]: I1125 10:36:22.434714 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g4r76" event={"ID":"268c8582-b9a0-4e22-ac84-e47a6fa3f284","Type":"ContainerDied","Data":"e9d2dc3ae125b4c3a0a7f0a47570f0351d7988bd73e7f68b65fddb861667f227"} Nov 25 10:36:22 crc kubenswrapper[4821]: I1125 10:36:22.728711 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-m99sj"] Nov 25 10:36:22 crc kubenswrapper[4821]: I1125 10:36:22.729651 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m99sj" Nov 25 10:36:22 crc kubenswrapper[4821]: I1125 10:36:22.732367 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 25 10:36:22 crc kubenswrapper[4821]: I1125 10:36:22.734510 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-m99sj"] Nov 25 10:36:22 crc kubenswrapper[4821]: I1125 10:36:22.822312 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a0df682-fa69-4ef1-a733-214abe72f1ad-utilities\") pod \"redhat-marketplace-m99sj\" (UID: \"8a0df682-fa69-4ef1-a733-214abe72f1ad\") " pod="openshift-marketplace/redhat-marketplace-m99sj" Nov 25 10:36:22 crc kubenswrapper[4821]: I1125 10:36:22.822741 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a0df682-fa69-4ef1-a733-214abe72f1ad-catalog-content\") pod \"redhat-marketplace-m99sj\" (UID: \"8a0df682-fa69-4ef1-a733-214abe72f1ad\") " pod="openshift-marketplace/redhat-marketplace-m99sj" Nov 25 10:36:22 crc kubenswrapper[4821]: I1125 10:36:22.822839 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f6mzh\" (UniqueName: \"kubernetes.io/projected/8a0df682-fa69-4ef1-a733-214abe72f1ad-kube-api-access-f6mzh\") pod \"redhat-marketplace-m99sj\" (UID: \"8a0df682-fa69-4ef1-a733-214abe72f1ad\") " pod="openshift-marketplace/redhat-marketplace-m99sj" Nov 25 10:36:22 crc kubenswrapper[4821]: I1125 10:36:22.924191 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f6mzh\" (UniqueName: \"kubernetes.io/projected/8a0df682-fa69-4ef1-a733-214abe72f1ad-kube-api-access-f6mzh\") pod \"redhat-marketplace-m99sj\" (UID: \"8a0df682-fa69-4ef1-a733-214abe72f1ad\") " pod="openshift-marketplace/redhat-marketplace-m99sj" Nov 25 10:36:22 crc kubenswrapper[4821]: I1125 10:36:22.924280 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a0df682-fa69-4ef1-a733-214abe72f1ad-utilities\") pod \"redhat-marketplace-m99sj\" (UID: \"8a0df682-fa69-4ef1-a733-214abe72f1ad\") " pod="openshift-marketplace/redhat-marketplace-m99sj" Nov 25 10:36:22 crc kubenswrapper[4821]: I1125 10:36:22.924304 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a0df682-fa69-4ef1-a733-214abe72f1ad-catalog-content\") pod \"redhat-marketplace-m99sj\" (UID: \"8a0df682-fa69-4ef1-a733-214abe72f1ad\") " pod="openshift-marketplace/redhat-marketplace-m99sj" Nov 25 10:36:22 crc kubenswrapper[4821]: I1125 10:36:22.924997 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a0df682-fa69-4ef1-a733-214abe72f1ad-utilities\") pod \"redhat-marketplace-m99sj\" (UID: \"8a0df682-fa69-4ef1-a733-214abe72f1ad\") " pod="openshift-marketplace/redhat-marketplace-m99sj" Nov 25 10:36:22 crc kubenswrapper[4821]: I1125 10:36:22.925155 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a0df682-fa69-4ef1-a733-214abe72f1ad-catalog-content\") pod \"redhat-marketplace-m99sj\" (UID: \"8a0df682-fa69-4ef1-a733-214abe72f1ad\") " pod="openshift-marketplace/redhat-marketplace-m99sj" Nov 25 10:36:22 crc kubenswrapper[4821]: I1125 10:36:22.949748 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f6mzh\" (UniqueName: \"kubernetes.io/projected/8a0df682-fa69-4ef1-a733-214abe72f1ad-kube-api-access-f6mzh\") pod \"redhat-marketplace-m99sj\" (UID: \"8a0df682-fa69-4ef1-a733-214abe72f1ad\") " pod="openshift-marketplace/redhat-marketplace-m99sj" Nov 25 10:36:23 crc kubenswrapper[4821]: I1125 10:36:23.077275 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m99sj" Nov 25 10:36:23 crc kubenswrapper[4821]: I1125 10:36:23.451128 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vrjfk" event={"ID":"7b01ce52-a13b-431c-8597-fbd5dc25183f","Type":"ContainerStarted","Data":"02730237ea8ba401377da3e688f1268c42752d3dc386c2cbd3fd7481401320dd"} Nov 25 10:36:23 crc kubenswrapper[4821]: I1125 10:36:23.463349 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-m99sj"] Nov 25 10:36:23 crc kubenswrapper[4821]: W1125 10:36:23.466998 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8a0df682_fa69_4ef1_a733_214abe72f1ad.slice/crio-3ab84016dd192629e6c763c9614b0cdc4934a3b1071b13120df29feaef6145ee WatchSource:0}: Error finding container 3ab84016dd192629e6c763c9614b0cdc4934a3b1071b13120df29feaef6145ee: Status 404 returned error can't find the container with id 3ab84016dd192629e6c763c9614b0cdc4934a3b1071b13120df29feaef6145ee Nov 25 10:36:24 crc kubenswrapper[4821]: I1125 10:36:24.459229 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g4r76" event={"ID":"268c8582-b9a0-4e22-ac84-e47a6fa3f284","Type":"ContainerStarted","Data":"d62769432b7b327ab2021a111e56df775eb219b3e265c3391fa560d009de6411"} Nov 25 10:36:24 crc kubenswrapper[4821]: I1125 10:36:24.461190 4821 generic.go:334] "Generic (PLEG): container finished" podID="7b01ce52-a13b-431c-8597-fbd5dc25183f" containerID="02730237ea8ba401377da3e688f1268c42752d3dc386c2cbd3fd7481401320dd" exitCode=0 Nov 25 10:36:24 crc kubenswrapper[4821]: I1125 10:36:24.461416 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vrjfk" event={"ID":"7b01ce52-a13b-431c-8597-fbd5dc25183f","Type":"ContainerDied","Data":"02730237ea8ba401377da3e688f1268c42752d3dc386c2cbd3fd7481401320dd"} Nov 25 10:36:24 crc kubenswrapper[4821]: I1125 10:36:24.462852 4821 generic.go:334] "Generic (PLEG): container finished" podID="8a0df682-fa69-4ef1-a733-214abe72f1ad" containerID="2f5751c1295f9fbb3e07a58af2b636c75140378f7d9a4a58fb960a4c435c41d7" exitCode=0 Nov 25 10:36:24 crc kubenswrapper[4821]: I1125 10:36:24.462899 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m99sj" event={"ID":"8a0df682-fa69-4ef1-a733-214abe72f1ad","Type":"ContainerDied","Data":"2f5751c1295f9fbb3e07a58af2b636c75140378f7d9a4a58fb960a4c435c41d7"} Nov 25 10:36:24 crc kubenswrapper[4821]: I1125 10:36:24.462917 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m99sj" event={"ID":"8a0df682-fa69-4ef1-a733-214abe72f1ad","Type":"ContainerStarted","Data":"3ab84016dd192629e6c763c9614b0cdc4934a3b1071b13120df29feaef6145ee"} Nov 25 10:36:24 crc kubenswrapper[4821]: I1125 10:36:24.483637 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-g4r76" podStartSLOduration=2.50090539 podStartE2EDuration="5.483620699s" podCreationTimestamp="2025-11-25 10:36:19 +0000 UTC" firstStartedPulling="2025-11-25 10:36:20.410610196 +0000 UTC m=+250.946930053" lastFinishedPulling="2025-11-25 10:36:23.393325515 +0000 UTC m=+253.929645362" observedRunningTime="2025-11-25 10:36:24.479036204 +0000 UTC m=+255.015356041" watchObservedRunningTime="2025-11-25 10:36:24.483620699 +0000 UTC m=+255.019940546" Nov 25 10:36:25 crc kubenswrapper[4821]: I1125 10:36:25.469469 4821 generic.go:334] "Generic (PLEG): container finished" podID="8a0df682-fa69-4ef1-a733-214abe72f1ad" containerID="c2cf24b60f694ef301f5d131f444a6e422a4f13934fb11f6b72c6bd8a03c173e" exitCode=0 Nov 25 10:36:25 crc kubenswrapper[4821]: I1125 10:36:25.469525 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m99sj" event={"ID":"8a0df682-fa69-4ef1-a733-214abe72f1ad","Type":"ContainerDied","Data":"c2cf24b60f694ef301f5d131f444a6e422a4f13934fb11f6b72c6bd8a03c173e"} Nov 25 10:36:25 crc kubenswrapper[4821]: I1125 10:36:25.473292 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vrjfk" event={"ID":"7b01ce52-a13b-431c-8597-fbd5dc25183f","Type":"ContainerStarted","Data":"ec86cffebbf2fa03d00d13faa4f868071533d25e0bf365675bf6e468d9d829f5"} Nov 25 10:36:25 crc kubenswrapper[4821]: I1125 10:36:25.505614 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-vrjfk" podStartSLOduration=2.04877053 podStartE2EDuration="5.505599578s" podCreationTimestamp="2025-11-25 10:36:20 +0000 UTC" firstStartedPulling="2025-11-25 10:36:21.421774241 +0000 UTC m=+251.958094088" lastFinishedPulling="2025-11-25 10:36:24.878603289 +0000 UTC m=+255.414923136" observedRunningTime="2025-11-25 10:36:25.504831357 +0000 UTC m=+256.041151204" watchObservedRunningTime="2025-11-25 10:36:25.505599578 +0000 UTC m=+256.041919425" Nov 25 10:36:26 crc kubenswrapper[4821]: I1125 10:36:26.844024 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-4j8hk" Nov 25 10:36:26 crc kubenswrapper[4821]: I1125 10:36:26.844401 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-4j8hk" Nov 25 10:36:26 crc kubenswrapper[4821]: I1125 10:36:26.886785 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-4j8hk" Nov 25 10:36:27 crc kubenswrapper[4821]: I1125 10:36:27.484020 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m99sj" event={"ID":"8a0df682-fa69-4ef1-a733-214abe72f1ad","Type":"ContainerStarted","Data":"a68e617b43590a80d3205c73d631be2dee2625b1766034a95aa910ff9aa47c95"} Nov 25 10:36:27 crc kubenswrapper[4821]: I1125 10:36:27.502972 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-m99sj" podStartSLOduration=4.094327267 podStartE2EDuration="5.502952281s" podCreationTimestamp="2025-11-25 10:36:22 +0000 UTC" firstStartedPulling="2025-11-25 10:36:24.464564131 +0000 UTC m=+255.000883978" lastFinishedPulling="2025-11-25 10:36:25.873189145 +0000 UTC m=+256.409508992" observedRunningTime="2025-11-25 10:36:27.502593452 +0000 UTC m=+258.038913319" watchObservedRunningTime="2025-11-25 10:36:27.502952281 +0000 UTC m=+258.039272128" Nov 25 10:36:27 crc kubenswrapper[4821]: I1125 10:36:27.523280 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-4j8hk" Nov 25 10:36:29 crc kubenswrapper[4821]: I1125 10:36:29.647596 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-g4r76" Nov 25 10:36:29 crc kubenswrapper[4821]: I1125 10:36:29.647968 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-g4r76" Nov 25 10:36:29 crc kubenswrapper[4821]: I1125 10:36:29.688757 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-g4r76" Nov 25 10:36:30 crc kubenswrapper[4821]: I1125 10:36:30.539675 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-g4r76" Nov 25 10:36:30 crc kubenswrapper[4821]: I1125 10:36:30.652355 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-vrjfk" Nov 25 10:36:30 crc kubenswrapper[4821]: I1125 10:36:30.652406 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-vrjfk" Nov 25 10:36:30 crc kubenswrapper[4821]: I1125 10:36:30.688028 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-vrjfk" Nov 25 10:36:31 crc kubenswrapper[4821]: I1125 10:36:31.537632 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-vrjfk" Nov 25 10:36:33 crc kubenswrapper[4821]: I1125 10:36:33.078279 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-m99sj" Nov 25 10:36:33 crc kubenswrapper[4821]: I1125 10:36:33.078449 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-m99sj" Nov 25 10:36:33 crc kubenswrapper[4821]: I1125 10:36:33.113510 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-m99sj" Nov 25 10:36:33 crc kubenswrapper[4821]: I1125 10:36:33.545607 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-m99sj" Nov 25 10:36:36 crc kubenswrapper[4821]: I1125 10:36:36.108700 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-mbn4k" podUID="88feae6a-5031-48ed-93ec-34ef8d2eff40" containerName="oauth-openshift" containerID="cri-o://694393c7f8c46b2f1b6aa9777e6755df30be0d3f493bb5ab58f77f1e24aeb086" gracePeriod=15 Nov 25 10:36:36 crc kubenswrapper[4821]: I1125 10:36:36.977642 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-mbn4k" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.006136 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-96d6999f9-5rjl8"] Nov 25 10:36:37 crc kubenswrapper[4821]: E1125 10:36:37.006383 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88feae6a-5031-48ed-93ec-34ef8d2eff40" containerName="oauth-openshift" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.006398 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="88feae6a-5031-48ed-93ec-34ef8d2eff40" containerName="oauth-openshift" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.006516 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="88feae6a-5031-48ed-93ec-34ef8d2eff40" containerName="oauth-openshift" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.006965 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-96d6999f9-5rjl8" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.067482 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-96d6999f9-5rjl8"] Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.113142 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/88feae6a-5031-48ed-93ec-34ef8d2eff40-v4-0-config-system-serving-cert\") pod \"88feae6a-5031-48ed-93ec-34ef8d2eff40\" (UID: \"88feae6a-5031-48ed-93ec-34ef8d2eff40\") " Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.113216 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/88feae6a-5031-48ed-93ec-34ef8d2eff40-v4-0-config-user-template-error\") pod \"88feae6a-5031-48ed-93ec-34ef8d2eff40\" (UID: \"88feae6a-5031-48ed-93ec-34ef8d2eff40\") " Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.113261 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/88feae6a-5031-48ed-93ec-34ef8d2eff40-v4-0-config-system-trusted-ca-bundle\") pod \"88feae6a-5031-48ed-93ec-34ef8d2eff40\" (UID: \"88feae6a-5031-48ed-93ec-34ef8d2eff40\") " Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.113296 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q7b5z\" (UniqueName: \"kubernetes.io/projected/88feae6a-5031-48ed-93ec-34ef8d2eff40-kube-api-access-q7b5z\") pod \"88feae6a-5031-48ed-93ec-34ef8d2eff40\" (UID: \"88feae6a-5031-48ed-93ec-34ef8d2eff40\") " Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.113322 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/88feae6a-5031-48ed-93ec-34ef8d2eff40-v4-0-config-system-cliconfig\") pod \"88feae6a-5031-48ed-93ec-34ef8d2eff40\" (UID: \"88feae6a-5031-48ed-93ec-34ef8d2eff40\") " Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.113348 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/88feae6a-5031-48ed-93ec-34ef8d2eff40-v4-0-config-system-service-ca\") pod \"88feae6a-5031-48ed-93ec-34ef8d2eff40\" (UID: \"88feae6a-5031-48ed-93ec-34ef8d2eff40\") " Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.113368 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/88feae6a-5031-48ed-93ec-34ef8d2eff40-v4-0-config-system-session\") pod \"88feae6a-5031-48ed-93ec-34ef8d2eff40\" (UID: \"88feae6a-5031-48ed-93ec-34ef8d2eff40\") " Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.113394 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/88feae6a-5031-48ed-93ec-34ef8d2eff40-v4-0-config-user-idp-0-file-data\") pod \"88feae6a-5031-48ed-93ec-34ef8d2eff40\" (UID: \"88feae6a-5031-48ed-93ec-34ef8d2eff40\") " Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.113459 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/88feae6a-5031-48ed-93ec-34ef8d2eff40-v4-0-config-system-ocp-branding-template\") pod \"88feae6a-5031-48ed-93ec-34ef8d2eff40\" (UID: \"88feae6a-5031-48ed-93ec-34ef8d2eff40\") " Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.113481 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/88feae6a-5031-48ed-93ec-34ef8d2eff40-audit-dir\") pod \"88feae6a-5031-48ed-93ec-34ef8d2eff40\" (UID: \"88feae6a-5031-48ed-93ec-34ef8d2eff40\") " Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.113506 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/88feae6a-5031-48ed-93ec-34ef8d2eff40-v4-0-config-system-router-certs\") pod \"88feae6a-5031-48ed-93ec-34ef8d2eff40\" (UID: \"88feae6a-5031-48ed-93ec-34ef8d2eff40\") " Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.113544 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/88feae6a-5031-48ed-93ec-34ef8d2eff40-v4-0-config-user-template-provider-selection\") pod \"88feae6a-5031-48ed-93ec-34ef8d2eff40\" (UID: \"88feae6a-5031-48ed-93ec-34ef8d2eff40\") " Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.113579 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/88feae6a-5031-48ed-93ec-34ef8d2eff40-v4-0-config-user-template-login\") pod \"88feae6a-5031-48ed-93ec-34ef8d2eff40\" (UID: \"88feae6a-5031-48ed-93ec-34ef8d2eff40\") " Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.113627 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/88feae6a-5031-48ed-93ec-34ef8d2eff40-audit-policies\") pod \"88feae6a-5031-48ed-93ec-34ef8d2eff40\" (UID: \"88feae6a-5031-48ed-93ec-34ef8d2eff40\") " Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.113775 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/5e449fa7-9b14-4c30-8af8-a94731358054-v4-0-config-user-template-error\") pod \"oauth-openshift-96d6999f9-5rjl8\" (UID: \"5e449fa7-9b14-4c30-8af8-a94731358054\") " pod="openshift-authentication/oauth-openshift-96d6999f9-5rjl8" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.113802 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/5e449fa7-9b14-4c30-8af8-a94731358054-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-96d6999f9-5rjl8\" (UID: \"5e449fa7-9b14-4c30-8af8-a94731358054\") " pod="openshift-authentication/oauth-openshift-96d6999f9-5rjl8" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.113823 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85wpt\" (UniqueName: \"kubernetes.io/projected/5e449fa7-9b14-4c30-8af8-a94731358054-kube-api-access-85wpt\") pod \"oauth-openshift-96d6999f9-5rjl8\" (UID: \"5e449fa7-9b14-4c30-8af8-a94731358054\") " pod="openshift-authentication/oauth-openshift-96d6999f9-5rjl8" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.113842 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/5e449fa7-9b14-4c30-8af8-a94731358054-v4-0-config-system-cliconfig\") pod \"oauth-openshift-96d6999f9-5rjl8\" (UID: \"5e449fa7-9b14-4c30-8af8-a94731358054\") " pod="openshift-authentication/oauth-openshift-96d6999f9-5rjl8" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.113859 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/5e449fa7-9b14-4c30-8af8-a94731358054-v4-0-config-user-template-login\") pod \"oauth-openshift-96d6999f9-5rjl8\" (UID: \"5e449fa7-9b14-4c30-8af8-a94731358054\") " pod="openshift-authentication/oauth-openshift-96d6999f9-5rjl8" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.113887 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/5e449fa7-9b14-4c30-8af8-a94731358054-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-96d6999f9-5rjl8\" (UID: \"5e449fa7-9b14-4c30-8af8-a94731358054\") " pod="openshift-authentication/oauth-openshift-96d6999f9-5rjl8" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.113915 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5e449fa7-9b14-4c30-8af8-a94731358054-audit-policies\") pod \"oauth-openshift-96d6999f9-5rjl8\" (UID: \"5e449fa7-9b14-4c30-8af8-a94731358054\") " pod="openshift-authentication/oauth-openshift-96d6999f9-5rjl8" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.113947 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/5e449fa7-9b14-4c30-8af8-a94731358054-v4-0-config-system-serving-cert\") pod \"oauth-openshift-96d6999f9-5rjl8\" (UID: \"5e449fa7-9b14-4c30-8af8-a94731358054\") " pod="openshift-authentication/oauth-openshift-96d6999f9-5rjl8" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.113973 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/5e449fa7-9b14-4c30-8af8-a94731358054-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-96d6999f9-5rjl8\" (UID: \"5e449fa7-9b14-4c30-8af8-a94731358054\") " pod="openshift-authentication/oauth-openshift-96d6999f9-5rjl8" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.114000 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/5e449fa7-9b14-4c30-8af8-a94731358054-v4-0-config-system-router-certs\") pod \"oauth-openshift-96d6999f9-5rjl8\" (UID: \"5e449fa7-9b14-4c30-8af8-a94731358054\") " pod="openshift-authentication/oauth-openshift-96d6999f9-5rjl8" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.114026 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/5e449fa7-9b14-4c30-8af8-a94731358054-v4-0-config-system-service-ca\") pod \"oauth-openshift-96d6999f9-5rjl8\" (UID: \"5e449fa7-9b14-4c30-8af8-a94731358054\") " pod="openshift-authentication/oauth-openshift-96d6999f9-5rjl8" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.114054 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5e449fa7-9b14-4c30-8af8-a94731358054-audit-dir\") pod \"oauth-openshift-96d6999f9-5rjl8\" (UID: \"5e449fa7-9b14-4c30-8af8-a94731358054\") " pod="openshift-authentication/oauth-openshift-96d6999f9-5rjl8" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.114077 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5e449fa7-9b14-4c30-8af8-a94731358054-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-96d6999f9-5rjl8\" (UID: \"5e449fa7-9b14-4c30-8af8-a94731358054\") " pod="openshift-authentication/oauth-openshift-96d6999f9-5rjl8" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.114117 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/5e449fa7-9b14-4c30-8af8-a94731358054-v4-0-config-system-session\") pod \"oauth-openshift-96d6999f9-5rjl8\" (UID: \"5e449fa7-9b14-4c30-8af8-a94731358054\") " pod="openshift-authentication/oauth-openshift-96d6999f9-5rjl8" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.114049 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/88feae6a-5031-48ed-93ec-34ef8d2eff40-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "88feae6a-5031-48ed-93ec-34ef8d2eff40" (UID: "88feae6a-5031-48ed-93ec-34ef8d2eff40"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.114097 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/88feae6a-5031-48ed-93ec-34ef8d2eff40-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "88feae6a-5031-48ed-93ec-34ef8d2eff40" (UID: "88feae6a-5031-48ed-93ec-34ef8d2eff40"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.114304 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/88feae6a-5031-48ed-93ec-34ef8d2eff40-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "88feae6a-5031-48ed-93ec-34ef8d2eff40" (UID: "88feae6a-5031-48ed-93ec-34ef8d2eff40"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.116605 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/88feae6a-5031-48ed-93ec-34ef8d2eff40-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "88feae6a-5031-48ed-93ec-34ef8d2eff40" (UID: "88feae6a-5031-48ed-93ec-34ef8d2eff40"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.116855 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/88feae6a-5031-48ed-93ec-34ef8d2eff40-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "88feae6a-5031-48ed-93ec-34ef8d2eff40" (UID: "88feae6a-5031-48ed-93ec-34ef8d2eff40"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.119231 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88feae6a-5031-48ed-93ec-34ef8d2eff40-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "88feae6a-5031-48ed-93ec-34ef8d2eff40" (UID: "88feae6a-5031-48ed-93ec-34ef8d2eff40"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.119478 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88feae6a-5031-48ed-93ec-34ef8d2eff40-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "88feae6a-5031-48ed-93ec-34ef8d2eff40" (UID: "88feae6a-5031-48ed-93ec-34ef8d2eff40"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.119703 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88feae6a-5031-48ed-93ec-34ef8d2eff40-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "88feae6a-5031-48ed-93ec-34ef8d2eff40" (UID: "88feae6a-5031-48ed-93ec-34ef8d2eff40"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.119850 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88feae6a-5031-48ed-93ec-34ef8d2eff40-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "88feae6a-5031-48ed-93ec-34ef8d2eff40" (UID: "88feae6a-5031-48ed-93ec-34ef8d2eff40"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.120410 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88feae6a-5031-48ed-93ec-34ef8d2eff40-kube-api-access-q7b5z" (OuterVolumeSpecName: "kube-api-access-q7b5z") pod "88feae6a-5031-48ed-93ec-34ef8d2eff40" (UID: "88feae6a-5031-48ed-93ec-34ef8d2eff40"). InnerVolumeSpecName "kube-api-access-q7b5z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.121691 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88feae6a-5031-48ed-93ec-34ef8d2eff40-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "88feae6a-5031-48ed-93ec-34ef8d2eff40" (UID: "88feae6a-5031-48ed-93ec-34ef8d2eff40"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.122353 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88feae6a-5031-48ed-93ec-34ef8d2eff40-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "88feae6a-5031-48ed-93ec-34ef8d2eff40" (UID: "88feae6a-5031-48ed-93ec-34ef8d2eff40"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.122514 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88feae6a-5031-48ed-93ec-34ef8d2eff40-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "88feae6a-5031-48ed-93ec-34ef8d2eff40" (UID: "88feae6a-5031-48ed-93ec-34ef8d2eff40"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.122983 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88feae6a-5031-48ed-93ec-34ef8d2eff40-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "88feae6a-5031-48ed-93ec-34ef8d2eff40" (UID: "88feae6a-5031-48ed-93ec-34ef8d2eff40"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.218929 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/5e449fa7-9b14-4c30-8af8-a94731358054-v4-0-config-system-session\") pod \"oauth-openshift-96d6999f9-5rjl8\" (UID: \"5e449fa7-9b14-4c30-8af8-a94731358054\") " pod="openshift-authentication/oauth-openshift-96d6999f9-5rjl8" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.219020 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/5e449fa7-9b14-4c30-8af8-a94731358054-v4-0-config-user-template-error\") pod \"oauth-openshift-96d6999f9-5rjl8\" (UID: \"5e449fa7-9b14-4c30-8af8-a94731358054\") " pod="openshift-authentication/oauth-openshift-96d6999f9-5rjl8" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.219058 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85wpt\" (UniqueName: \"kubernetes.io/projected/5e449fa7-9b14-4c30-8af8-a94731358054-kube-api-access-85wpt\") pod \"oauth-openshift-96d6999f9-5rjl8\" (UID: \"5e449fa7-9b14-4c30-8af8-a94731358054\") " pod="openshift-authentication/oauth-openshift-96d6999f9-5rjl8" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.219089 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/5e449fa7-9b14-4c30-8af8-a94731358054-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-96d6999f9-5rjl8\" (UID: \"5e449fa7-9b14-4c30-8af8-a94731358054\") " pod="openshift-authentication/oauth-openshift-96d6999f9-5rjl8" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.219119 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/5e449fa7-9b14-4c30-8af8-a94731358054-v4-0-config-system-cliconfig\") pod \"oauth-openshift-96d6999f9-5rjl8\" (UID: \"5e449fa7-9b14-4c30-8af8-a94731358054\") " pod="openshift-authentication/oauth-openshift-96d6999f9-5rjl8" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.219142 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/5e449fa7-9b14-4c30-8af8-a94731358054-v4-0-config-user-template-login\") pod \"oauth-openshift-96d6999f9-5rjl8\" (UID: \"5e449fa7-9b14-4c30-8af8-a94731358054\") " pod="openshift-authentication/oauth-openshift-96d6999f9-5rjl8" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.219199 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/5e449fa7-9b14-4c30-8af8-a94731358054-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-96d6999f9-5rjl8\" (UID: \"5e449fa7-9b14-4c30-8af8-a94731358054\") " pod="openshift-authentication/oauth-openshift-96d6999f9-5rjl8" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.219241 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5e449fa7-9b14-4c30-8af8-a94731358054-audit-policies\") pod \"oauth-openshift-96d6999f9-5rjl8\" (UID: \"5e449fa7-9b14-4c30-8af8-a94731358054\") " pod="openshift-authentication/oauth-openshift-96d6999f9-5rjl8" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.219283 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/5e449fa7-9b14-4c30-8af8-a94731358054-v4-0-config-system-serving-cert\") pod \"oauth-openshift-96d6999f9-5rjl8\" (UID: \"5e449fa7-9b14-4c30-8af8-a94731358054\") " pod="openshift-authentication/oauth-openshift-96d6999f9-5rjl8" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.219314 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/5e449fa7-9b14-4c30-8af8-a94731358054-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-96d6999f9-5rjl8\" (UID: \"5e449fa7-9b14-4c30-8af8-a94731358054\") " pod="openshift-authentication/oauth-openshift-96d6999f9-5rjl8" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.219352 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/5e449fa7-9b14-4c30-8af8-a94731358054-v4-0-config-system-router-certs\") pod \"oauth-openshift-96d6999f9-5rjl8\" (UID: \"5e449fa7-9b14-4c30-8af8-a94731358054\") " pod="openshift-authentication/oauth-openshift-96d6999f9-5rjl8" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.219383 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/5e449fa7-9b14-4c30-8af8-a94731358054-v4-0-config-system-service-ca\") pod \"oauth-openshift-96d6999f9-5rjl8\" (UID: \"5e449fa7-9b14-4c30-8af8-a94731358054\") " pod="openshift-authentication/oauth-openshift-96d6999f9-5rjl8" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.219418 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5e449fa7-9b14-4c30-8af8-a94731358054-audit-dir\") pod \"oauth-openshift-96d6999f9-5rjl8\" (UID: \"5e449fa7-9b14-4c30-8af8-a94731358054\") " pod="openshift-authentication/oauth-openshift-96d6999f9-5rjl8" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.219449 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5e449fa7-9b14-4c30-8af8-a94731358054-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-96d6999f9-5rjl8\" (UID: \"5e449fa7-9b14-4c30-8af8-a94731358054\") " pod="openshift-authentication/oauth-openshift-96d6999f9-5rjl8" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.219533 4821 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/88feae6a-5031-48ed-93ec-34ef8d2eff40-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.219555 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q7b5z\" (UniqueName: \"kubernetes.io/projected/88feae6a-5031-48ed-93ec-34ef8d2eff40-kube-api-access-q7b5z\") on node \"crc\" DevicePath \"\"" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.219567 4821 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/88feae6a-5031-48ed-93ec-34ef8d2eff40-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.219581 4821 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/88feae6a-5031-48ed-93ec-34ef8d2eff40-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.219593 4821 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/88feae6a-5031-48ed-93ec-34ef8d2eff40-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.219609 4821 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/88feae6a-5031-48ed-93ec-34ef8d2eff40-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.219622 4821 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/88feae6a-5031-48ed-93ec-34ef8d2eff40-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.219635 4821 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/88feae6a-5031-48ed-93ec-34ef8d2eff40-audit-dir\") on node \"crc\" DevicePath \"\"" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.219647 4821 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/88feae6a-5031-48ed-93ec-34ef8d2eff40-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.219664 4821 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/88feae6a-5031-48ed-93ec-34ef8d2eff40-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.219677 4821 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/88feae6a-5031-48ed-93ec-34ef8d2eff40-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.219688 4821 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/88feae6a-5031-48ed-93ec-34ef8d2eff40-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.219707 4821 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/88feae6a-5031-48ed-93ec-34ef8d2eff40-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.219721 4821 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/88feae6a-5031-48ed-93ec-34ef8d2eff40-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.220727 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5e449fa7-9b14-4c30-8af8-a94731358054-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-96d6999f9-5rjl8\" (UID: \"5e449fa7-9b14-4c30-8af8-a94731358054\") " pod="openshift-authentication/oauth-openshift-96d6999f9-5rjl8" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.224345 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/5e449fa7-9b14-4c30-8af8-a94731358054-v4-0-config-system-cliconfig\") pod \"oauth-openshift-96d6999f9-5rjl8\" (UID: \"5e449fa7-9b14-4c30-8af8-a94731358054\") " pod="openshift-authentication/oauth-openshift-96d6999f9-5rjl8" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.224506 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5e449fa7-9b14-4c30-8af8-a94731358054-audit-dir\") pod \"oauth-openshift-96d6999f9-5rjl8\" (UID: \"5e449fa7-9b14-4c30-8af8-a94731358054\") " pod="openshift-authentication/oauth-openshift-96d6999f9-5rjl8" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.224865 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/5e449fa7-9b14-4c30-8af8-a94731358054-v4-0-config-system-service-ca\") pod \"oauth-openshift-96d6999f9-5rjl8\" (UID: \"5e449fa7-9b14-4c30-8af8-a94731358054\") " pod="openshift-authentication/oauth-openshift-96d6999f9-5rjl8" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.225130 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5e449fa7-9b14-4c30-8af8-a94731358054-audit-policies\") pod \"oauth-openshift-96d6999f9-5rjl8\" (UID: \"5e449fa7-9b14-4c30-8af8-a94731358054\") " pod="openshift-authentication/oauth-openshift-96d6999f9-5rjl8" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.225242 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/5e449fa7-9b14-4c30-8af8-a94731358054-v4-0-config-system-session\") pod \"oauth-openshift-96d6999f9-5rjl8\" (UID: \"5e449fa7-9b14-4c30-8af8-a94731358054\") " pod="openshift-authentication/oauth-openshift-96d6999f9-5rjl8" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.226990 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/5e449fa7-9b14-4c30-8af8-a94731358054-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-96d6999f9-5rjl8\" (UID: \"5e449fa7-9b14-4c30-8af8-a94731358054\") " pod="openshift-authentication/oauth-openshift-96d6999f9-5rjl8" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.227211 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/5e449fa7-9b14-4c30-8af8-a94731358054-v4-0-config-user-template-login\") pod \"oauth-openshift-96d6999f9-5rjl8\" (UID: \"5e449fa7-9b14-4c30-8af8-a94731358054\") " pod="openshift-authentication/oauth-openshift-96d6999f9-5rjl8" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.229007 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/5e449fa7-9b14-4c30-8af8-a94731358054-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-96d6999f9-5rjl8\" (UID: \"5e449fa7-9b14-4c30-8af8-a94731358054\") " pod="openshift-authentication/oauth-openshift-96d6999f9-5rjl8" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.230266 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/5e449fa7-9b14-4c30-8af8-a94731358054-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-96d6999f9-5rjl8\" (UID: \"5e449fa7-9b14-4c30-8af8-a94731358054\") " pod="openshift-authentication/oauth-openshift-96d6999f9-5rjl8" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.239693 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/5e449fa7-9b14-4c30-8af8-a94731358054-v4-0-config-system-serving-cert\") pod \"oauth-openshift-96d6999f9-5rjl8\" (UID: \"5e449fa7-9b14-4c30-8af8-a94731358054\") " pod="openshift-authentication/oauth-openshift-96d6999f9-5rjl8" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.240355 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/5e449fa7-9b14-4c30-8af8-a94731358054-v4-0-config-system-router-certs\") pod \"oauth-openshift-96d6999f9-5rjl8\" (UID: \"5e449fa7-9b14-4c30-8af8-a94731358054\") " pod="openshift-authentication/oauth-openshift-96d6999f9-5rjl8" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.240286 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/5e449fa7-9b14-4c30-8af8-a94731358054-v4-0-config-user-template-error\") pod \"oauth-openshift-96d6999f9-5rjl8\" (UID: \"5e449fa7-9b14-4c30-8af8-a94731358054\") " pod="openshift-authentication/oauth-openshift-96d6999f9-5rjl8" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.241586 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85wpt\" (UniqueName: \"kubernetes.io/projected/5e449fa7-9b14-4c30-8af8-a94731358054-kube-api-access-85wpt\") pod \"oauth-openshift-96d6999f9-5rjl8\" (UID: \"5e449fa7-9b14-4c30-8af8-a94731358054\") " pod="openshift-authentication/oauth-openshift-96d6999f9-5rjl8" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.325070 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-96d6999f9-5rjl8" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.533841 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-mbn4k" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.533897 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-mbn4k" event={"ID":"88feae6a-5031-48ed-93ec-34ef8d2eff40","Type":"ContainerDied","Data":"694393c7f8c46b2f1b6aa9777e6755df30be0d3f493bb5ab58f77f1e24aeb086"} Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.534257 4821 scope.go:117] "RemoveContainer" containerID="694393c7f8c46b2f1b6aa9777e6755df30be0d3f493bb5ab58f77f1e24aeb086" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.533718 4821 generic.go:334] "Generic (PLEG): container finished" podID="88feae6a-5031-48ed-93ec-34ef8d2eff40" containerID="694393c7f8c46b2f1b6aa9777e6755df30be0d3f493bb5ab58f77f1e24aeb086" exitCode=0 Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.534444 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-mbn4k" event={"ID":"88feae6a-5031-48ed-93ec-34ef8d2eff40","Type":"ContainerDied","Data":"d9d52c11eb28b942ee927fd912e4bb08f1ffc0b522c3507677b9105434e099d9"} Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.553726 4821 scope.go:117] "RemoveContainer" containerID="694393c7f8c46b2f1b6aa9777e6755df30be0d3f493bb5ab58f77f1e24aeb086" Nov 25 10:36:37 crc kubenswrapper[4821]: E1125 10:36:37.554222 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"694393c7f8c46b2f1b6aa9777e6755df30be0d3f493bb5ab58f77f1e24aeb086\": container with ID starting with 694393c7f8c46b2f1b6aa9777e6755df30be0d3f493bb5ab58f77f1e24aeb086 not found: ID does not exist" containerID="694393c7f8c46b2f1b6aa9777e6755df30be0d3f493bb5ab58f77f1e24aeb086" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.554304 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"694393c7f8c46b2f1b6aa9777e6755df30be0d3f493bb5ab58f77f1e24aeb086"} err="failed to get container status \"694393c7f8c46b2f1b6aa9777e6755df30be0d3f493bb5ab58f77f1e24aeb086\": rpc error: code = NotFound desc = could not find container \"694393c7f8c46b2f1b6aa9777e6755df30be0d3f493bb5ab58f77f1e24aeb086\": container with ID starting with 694393c7f8c46b2f1b6aa9777e6755df30be0d3f493bb5ab58f77f1e24aeb086 not found: ID does not exist" Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.560593 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-mbn4k"] Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.563413 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-mbn4k"] Nov 25 10:36:37 crc kubenswrapper[4821]: I1125 10:36:37.723201 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-96d6999f9-5rjl8"] Nov 25 10:36:37 crc kubenswrapper[4821]: W1125 10:36:37.727188 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5e449fa7_9b14_4c30_8af8_a94731358054.slice/crio-0a9167eae05a980dfc9f631678652741ee111875f166a0bdc257cfab22f8c98c WatchSource:0}: Error finding container 0a9167eae05a980dfc9f631678652741ee111875f166a0bdc257cfab22f8c98c: Status 404 returned error can't find the container with id 0a9167eae05a980dfc9f631678652741ee111875f166a0bdc257cfab22f8c98c Nov 25 10:36:38 crc kubenswrapper[4821]: I1125 10:36:38.122267 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88feae6a-5031-48ed-93ec-34ef8d2eff40" path="/var/lib/kubelet/pods/88feae6a-5031-48ed-93ec-34ef8d2eff40/volumes" Nov 25 10:36:38 crc kubenswrapper[4821]: I1125 10:36:38.541142 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-96d6999f9-5rjl8" event={"ID":"5e449fa7-9b14-4c30-8af8-a94731358054","Type":"ContainerStarted","Data":"7252e93f9482779cdf7e4c5057fd050254c2b374d5863f4d573b0c349ca178bf"} Nov 25 10:36:38 crc kubenswrapper[4821]: I1125 10:36:38.541266 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-96d6999f9-5rjl8" event={"ID":"5e449fa7-9b14-4c30-8af8-a94731358054","Type":"ContainerStarted","Data":"0a9167eae05a980dfc9f631678652741ee111875f166a0bdc257cfab22f8c98c"} Nov 25 10:36:38 crc kubenswrapper[4821]: I1125 10:36:38.541657 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-96d6999f9-5rjl8" Nov 25 10:36:38 crc kubenswrapper[4821]: I1125 10:36:38.545896 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-96d6999f9-5rjl8" Nov 25 10:36:38 crc kubenswrapper[4821]: I1125 10:36:38.562893 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-96d6999f9-5rjl8" podStartSLOduration=27.562875588 podStartE2EDuration="27.562875588s" podCreationTimestamp="2025-11-25 10:36:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:36:38.562122496 +0000 UTC m=+269.098442353" watchObservedRunningTime="2025-11-25 10:36:38.562875588 +0000 UTC m=+269.099195435" Nov 25 10:38:04 crc kubenswrapper[4821]: I1125 10:38:04.714992 4821 patch_prober.go:28] interesting pod/machine-config-daemon-2krbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:38:04 crc kubenswrapper[4821]: I1125 10:38:04.716344 4821 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:38:34 crc kubenswrapper[4821]: I1125 10:38:34.715266 4821 patch_prober.go:28] interesting pod/machine-config-daemon-2krbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:38:34 crc kubenswrapper[4821]: I1125 10:38:34.715746 4821 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:39:04 crc kubenswrapper[4821]: I1125 10:39:04.715455 4821 patch_prober.go:28] interesting pod/machine-config-daemon-2krbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:39:04 crc kubenswrapper[4821]: I1125 10:39:04.716301 4821 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:39:04 crc kubenswrapper[4821]: I1125 10:39:04.716369 4821 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" Nov 25 10:39:04 crc kubenswrapper[4821]: I1125 10:39:04.717054 4821 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"131eaad8c97dd44aa2f047258a3256b958757d6189e9d69d307bd69c27cbcd08"} pod="openshift-machine-config-operator/machine-config-daemon-2krbf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 10:39:04 crc kubenswrapper[4821]: I1125 10:39:04.717116 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerName="machine-config-daemon" containerID="cri-o://131eaad8c97dd44aa2f047258a3256b958757d6189e9d69d307bd69c27cbcd08" gracePeriod=600 Nov 25 10:39:05 crc kubenswrapper[4821]: I1125 10:39:05.260760 4821 generic.go:334] "Generic (PLEG): container finished" podID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerID="131eaad8c97dd44aa2f047258a3256b958757d6189e9d69d307bd69c27cbcd08" exitCode=0 Nov 25 10:39:05 crc kubenswrapper[4821]: I1125 10:39:05.260865 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" event={"ID":"5f948b87-ac86-4de6-ad64-c2ef947f84d4","Type":"ContainerDied","Data":"131eaad8c97dd44aa2f047258a3256b958757d6189e9d69d307bd69c27cbcd08"} Nov 25 10:39:05 crc kubenswrapper[4821]: I1125 10:39:05.262450 4821 scope.go:117] "RemoveContainer" containerID="9ba68bd093be94d0d40f3851638db5d26bdfc52d175b1ac78afb91dda7affcfe" Nov 25 10:39:06 crc kubenswrapper[4821]: I1125 10:39:06.270693 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" event={"ID":"5f948b87-ac86-4de6-ad64-c2ef947f84d4","Type":"ContainerStarted","Data":"808518ddff79187bc281fef8f1f64e070e5abe86921d58acacca289efda8d1d4"} Nov 25 10:39:45 crc kubenswrapper[4821]: I1125 10:39:45.184831 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-xrh2h"] Nov 25 10:39:45 crc kubenswrapper[4821]: I1125 10:39:45.186224 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-xrh2h" Nov 25 10:39:45 crc kubenswrapper[4821]: I1125 10:39:45.200513 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-xrh2h"] Nov 25 10:39:45 crc kubenswrapper[4821]: I1125 10:39:45.251426 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/b7cb462a-d71d-4f64-8c00-178d0fff072d-registry-certificates\") pod \"image-registry-66df7c8f76-xrh2h\" (UID: \"b7cb462a-d71d-4f64-8c00-178d0fff072d\") " pod="openshift-image-registry/image-registry-66df7c8f76-xrh2h" Nov 25 10:39:45 crc kubenswrapper[4821]: I1125 10:39:45.251728 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-xrh2h\" (UID: \"b7cb462a-d71d-4f64-8c00-178d0fff072d\") " pod="openshift-image-registry/image-registry-66df7c8f76-xrh2h" Nov 25 10:39:45 crc kubenswrapper[4821]: I1125 10:39:45.251829 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b7cb462a-d71d-4f64-8c00-178d0fff072d-trusted-ca\") pod \"image-registry-66df7c8f76-xrh2h\" (UID: \"b7cb462a-d71d-4f64-8c00-178d0fff072d\") " pod="openshift-image-registry/image-registry-66df7c8f76-xrh2h" Nov 25 10:39:45 crc kubenswrapper[4821]: I1125 10:39:45.251918 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b7cb462a-d71d-4f64-8c00-178d0fff072d-bound-sa-token\") pod \"image-registry-66df7c8f76-xrh2h\" (UID: \"b7cb462a-d71d-4f64-8c00-178d0fff072d\") " pod="openshift-image-registry/image-registry-66df7c8f76-xrh2h" Nov 25 10:39:45 crc kubenswrapper[4821]: I1125 10:39:45.252133 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/b7cb462a-d71d-4f64-8c00-178d0fff072d-ca-trust-extracted\") pod \"image-registry-66df7c8f76-xrh2h\" (UID: \"b7cb462a-d71d-4f64-8c00-178d0fff072d\") " pod="openshift-image-registry/image-registry-66df7c8f76-xrh2h" Nov 25 10:39:45 crc kubenswrapper[4821]: I1125 10:39:45.252306 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/b7cb462a-d71d-4f64-8c00-178d0fff072d-installation-pull-secrets\") pod \"image-registry-66df7c8f76-xrh2h\" (UID: \"b7cb462a-d71d-4f64-8c00-178d0fff072d\") " pod="openshift-image-registry/image-registry-66df7c8f76-xrh2h" Nov 25 10:39:45 crc kubenswrapper[4821]: I1125 10:39:45.252367 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/b7cb462a-d71d-4f64-8c00-178d0fff072d-registry-tls\") pod \"image-registry-66df7c8f76-xrh2h\" (UID: \"b7cb462a-d71d-4f64-8c00-178d0fff072d\") " pod="openshift-image-registry/image-registry-66df7c8f76-xrh2h" Nov 25 10:39:45 crc kubenswrapper[4821]: I1125 10:39:45.252407 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9nhfz\" (UniqueName: \"kubernetes.io/projected/b7cb462a-d71d-4f64-8c00-178d0fff072d-kube-api-access-9nhfz\") pod \"image-registry-66df7c8f76-xrh2h\" (UID: \"b7cb462a-d71d-4f64-8c00-178d0fff072d\") " pod="openshift-image-registry/image-registry-66df7c8f76-xrh2h" Nov 25 10:39:45 crc kubenswrapper[4821]: I1125 10:39:45.276297 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-xrh2h\" (UID: \"b7cb462a-d71d-4f64-8c00-178d0fff072d\") " pod="openshift-image-registry/image-registry-66df7c8f76-xrh2h" Nov 25 10:39:45 crc kubenswrapper[4821]: I1125 10:39:45.353914 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/b7cb462a-d71d-4f64-8c00-178d0fff072d-ca-trust-extracted\") pod \"image-registry-66df7c8f76-xrh2h\" (UID: \"b7cb462a-d71d-4f64-8c00-178d0fff072d\") " pod="openshift-image-registry/image-registry-66df7c8f76-xrh2h" Nov 25 10:39:45 crc kubenswrapper[4821]: I1125 10:39:45.353977 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/b7cb462a-d71d-4f64-8c00-178d0fff072d-installation-pull-secrets\") pod \"image-registry-66df7c8f76-xrh2h\" (UID: \"b7cb462a-d71d-4f64-8c00-178d0fff072d\") " pod="openshift-image-registry/image-registry-66df7c8f76-xrh2h" Nov 25 10:39:45 crc kubenswrapper[4821]: I1125 10:39:45.354005 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/b7cb462a-d71d-4f64-8c00-178d0fff072d-registry-tls\") pod \"image-registry-66df7c8f76-xrh2h\" (UID: \"b7cb462a-d71d-4f64-8c00-178d0fff072d\") " pod="openshift-image-registry/image-registry-66df7c8f76-xrh2h" Nov 25 10:39:45 crc kubenswrapper[4821]: I1125 10:39:45.354027 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9nhfz\" (UniqueName: \"kubernetes.io/projected/b7cb462a-d71d-4f64-8c00-178d0fff072d-kube-api-access-9nhfz\") pod \"image-registry-66df7c8f76-xrh2h\" (UID: \"b7cb462a-d71d-4f64-8c00-178d0fff072d\") " pod="openshift-image-registry/image-registry-66df7c8f76-xrh2h" Nov 25 10:39:45 crc kubenswrapper[4821]: I1125 10:39:45.355010 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/b7cb462a-d71d-4f64-8c00-178d0fff072d-ca-trust-extracted\") pod \"image-registry-66df7c8f76-xrh2h\" (UID: \"b7cb462a-d71d-4f64-8c00-178d0fff072d\") " pod="openshift-image-registry/image-registry-66df7c8f76-xrh2h" Nov 25 10:39:45 crc kubenswrapper[4821]: I1125 10:39:45.355044 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/b7cb462a-d71d-4f64-8c00-178d0fff072d-registry-certificates\") pod \"image-registry-66df7c8f76-xrh2h\" (UID: \"b7cb462a-d71d-4f64-8c00-178d0fff072d\") " pod="openshift-image-registry/image-registry-66df7c8f76-xrh2h" Nov 25 10:39:45 crc kubenswrapper[4821]: I1125 10:39:45.355136 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b7cb462a-d71d-4f64-8c00-178d0fff072d-trusted-ca\") pod \"image-registry-66df7c8f76-xrh2h\" (UID: \"b7cb462a-d71d-4f64-8c00-178d0fff072d\") " pod="openshift-image-registry/image-registry-66df7c8f76-xrh2h" Nov 25 10:39:45 crc kubenswrapper[4821]: I1125 10:39:45.355969 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/b7cb462a-d71d-4f64-8c00-178d0fff072d-registry-certificates\") pod \"image-registry-66df7c8f76-xrh2h\" (UID: \"b7cb462a-d71d-4f64-8c00-178d0fff072d\") " pod="openshift-image-registry/image-registry-66df7c8f76-xrh2h" Nov 25 10:39:45 crc kubenswrapper[4821]: I1125 10:39:45.356086 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b7cb462a-d71d-4f64-8c00-178d0fff072d-trusted-ca\") pod \"image-registry-66df7c8f76-xrh2h\" (UID: \"b7cb462a-d71d-4f64-8c00-178d0fff072d\") " pod="openshift-image-registry/image-registry-66df7c8f76-xrh2h" Nov 25 10:39:45 crc kubenswrapper[4821]: I1125 10:39:45.355155 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b7cb462a-d71d-4f64-8c00-178d0fff072d-bound-sa-token\") pod \"image-registry-66df7c8f76-xrh2h\" (UID: \"b7cb462a-d71d-4f64-8c00-178d0fff072d\") " pod="openshift-image-registry/image-registry-66df7c8f76-xrh2h" Nov 25 10:39:45 crc kubenswrapper[4821]: I1125 10:39:45.364072 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/b7cb462a-d71d-4f64-8c00-178d0fff072d-registry-tls\") pod \"image-registry-66df7c8f76-xrh2h\" (UID: \"b7cb462a-d71d-4f64-8c00-178d0fff072d\") " pod="openshift-image-registry/image-registry-66df7c8f76-xrh2h" Nov 25 10:39:45 crc kubenswrapper[4821]: I1125 10:39:45.365153 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/b7cb462a-d71d-4f64-8c00-178d0fff072d-installation-pull-secrets\") pod \"image-registry-66df7c8f76-xrh2h\" (UID: \"b7cb462a-d71d-4f64-8c00-178d0fff072d\") " pod="openshift-image-registry/image-registry-66df7c8f76-xrh2h" Nov 25 10:39:45 crc kubenswrapper[4821]: I1125 10:39:45.370895 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b7cb462a-d71d-4f64-8c00-178d0fff072d-bound-sa-token\") pod \"image-registry-66df7c8f76-xrh2h\" (UID: \"b7cb462a-d71d-4f64-8c00-178d0fff072d\") " pod="openshift-image-registry/image-registry-66df7c8f76-xrh2h" Nov 25 10:39:45 crc kubenswrapper[4821]: I1125 10:39:45.371065 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9nhfz\" (UniqueName: \"kubernetes.io/projected/b7cb462a-d71d-4f64-8c00-178d0fff072d-kube-api-access-9nhfz\") pod \"image-registry-66df7c8f76-xrh2h\" (UID: \"b7cb462a-d71d-4f64-8c00-178d0fff072d\") " pod="openshift-image-registry/image-registry-66df7c8f76-xrh2h" Nov 25 10:39:45 crc kubenswrapper[4821]: I1125 10:39:45.502520 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-xrh2h" Nov 25 10:39:45 crc kubenswrapper[4821]: I1125 10:39:45.684660 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-xrh2h"] Nov 25 10:39:46 crc kubenswrapper[4821]: I1125 10:39:46.460424 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-xrh2h" event={"ID":"b7cb462a-d71d-4f64-8c00-178d0fff072d","Type":"ContainerStarted","Data":"e73f35644bc44ced086a2b000e25d7ca53d45b170582bb1df3d896d32041767a"} Nov 25 10:39:46 crc kubenswrapper[4821]: I1125 10:39:46.460485 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-xrh2h" event={"ID":"b7cb462a-d71d-4f64-8c00-178d0fff072d","Type":"ContainerStarted","Data":"fc69619394fc48fe0e03873e47466ebcbd556bdba10810d60cc1f660d7d9e1e2"} Nov 25 10:39:46 crc kubenswrapper[4821]: I1125 10:39:46.460625 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-xrh2h" Nov 25 10:39:46 crc kubenswrapper[4821]: I1125 10:39:46.480319 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-xrh2h" podStartSLOduration=1.4803016119999999 podStartE2EDuration="1.480301612s" podCreationTimestamp="2025-11-25 10:39:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:39:46.47705488 +0000 UTC m=+457.013374757" watchObservedRunningTime="2025-11-25 10:39:46.480301612 +0000 UTC m=+457.016621449" Nov 25 10:40:05 crc kubenswrapper[4821]: I1125 10:40:05.510688 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-xrh2h" Nov 25 10:40:05 crc kubenswrapper[4821]: I1125 10:40:05.566430 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-xnlvh"] Nov 25 10:40:30 crc kubenswrapper[4821]: I1125 10:40:30.607692 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" podUID="3fb2355d-5ec2-44f1-8665-a3d73c24f8a6" containerName="registry" containerID="cri-o://dfeb5a5d613f3284ed07af1e11e0bc565beb1e5303175ab7e151caf082203aed" gracePeriod=30 Nov 25 10:40:30 crc kubenswrapper[4821]: I1125 10:40:30.915453 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:40:31 crc kubenswrapper[4821]: I1125 10:40:31.073464 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/3fb2355d-5ec2-44f1-8665-a3d73c24f8a6-ca-trust-extracted\") pod \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " Nov 25 10:40:31 crc kubenswrapper[4821]: I1125 10:40:31.073551 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3fb2355d-5ec2-44f1-8665-a3d73c24f8a6-trusted-ca\") pod \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " Nov 25 10:40:31 crc kubenswrapper[4821]: I1125 10:40:31.073610 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/3fb2355d-5ec2-44f1-8665-a3d73c24f8a6-registry-tls\") pod \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " Nov 25 10:40:31 crc kubenswrapper[4821]: I1125 10:40:31.073655 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3fb2355d-5ec2-44f1-8665-a3d73c24f8a6-bound-sa-token\") pod \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " Nov 25 10:40:31 crc kubenswrapper[4821]: I1125 10:40:31.073734 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/3fb2355d-5ec2-44f1-8665-a3d73c24f8a6-installation-pull-secrets\") pod \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " Nov 25 10:40:31 crc kubenswrapper[4821]: I1125 10:40:31.073826 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lgwn7\" (UniqueName: \"kubernetes.io/projected/3fb2355d-5ec2-44f1-8665-a3d73c24f8a6-kube-api-access-lgwn7\") pod \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " Nov 25 10:40:31 crc kubenswrapper[4821]: I1125 10:40:31.073878 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/3fb2355d-5ec2-44f1-8665-a3d73c24f8a6-registry-certificates\") pod \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " Nov 25 10:40:31 crc kubenswrapper[4821]: I1125 10:40:31.074083 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\" (UID: \"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6\") " Nov 25 10:40:31 crc kubenswrapper[4821]: I1125 10:40:31.075712 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3fb2355d-5ec2-44f1-8665-a3d73c24f8a6-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:40:31 crc kubenswrapper[4821]: I1125 10:40:31.075754 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3fb2355d-5ec2-44f1-8665-a3d73c24f8a6-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:40:31 crc kubenswrapper[4821]: I1125 10:40:31.080881 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3fb2355d-5ec2-44f1-8665-a3d73c24f8a6-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:40:31 crc kubenswrapper[4821]: I1125 10:40:31.081794 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3fb2355d-5ec2-44f1-8665-a3d73c24f8a6-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:40:31 crc kubenswrapper[4821]: I1125 10:40:31.082584 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3fb2355d-5ec2-44f1-8665-a3d73c24f8a6-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:40:31 crc kubenswrapper[4821]: I1125 10:40:31.083351 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3fb2355d-5ec2-44f1-8665-a3d73c24f8a6-kube-api-access-lgwn7" (OuterVolumeSpecName: "kube-api-access-lgwn7") pod "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6"). InnerVolumeSpecName "kube-api-access-lgwn7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:40:31 crc kubenswrapper[4821]: I1125 10:40:31.087432 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 25 10:40:31 crc kubenswrapper[4821]: I1125 10:40:31.093224 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3fb2355d-5ec2-44f1-8665-a3d73c24f8a6-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6" (UID: "3fb2355d-5ec2-44f1-8665-a3d73c24f8a6"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:40:31 crc kubenswrapper[4821]: I1125 10:40:31.175373 4821 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/3fb2355d-5ec2-44f1-8665-a3d73c24f8a6-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 25 10:40:31 crc kubenswrapper[4821]: I1125 10:40:31.175745 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lgwn7\" (UniqueName: \"kubernetes.io/projected/3fb2355d-5ec2-44f1-8665-a3d73c24f8a6-kube-api-access-lgwn7\") on node \"crc\" DevicePath \"\"" Nov 25 10:40:31 crc kubenswrapper[4821]: I1125 10:40:31.175829 4821 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/3fb2355d-5ec2-44f1-8665-a3d73c24f8a6-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 25 10:40:31 crc kubenswrapper[4821]: I1125 10:40:31.175898 4821 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/3fb2355d-5ec2-44f1-8665-a3d73c24f8a6-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 25 10:40:31 crc kubenswrapper[4821]: I1125 10:40:31.175964 4821 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3fb2355d-5ec2-44f1-8665-a3d73c24f8a6-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 10:40:31 crc kubenswrapper[4821]: I1125 10:40:31.176026 4821 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3fb2355d-5ec2-44f1-8665-a3d73c24f8a6-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 25 10:40:31 crc kubenswrapper[4821]: I1125 10:40:31.176082 4821 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/3fb2355d-5ec2-44f1-8665-a3d73c24f8a6-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 25 10:40:31 crc kubenswrapper[4821]: I1125 10:40:31.735557 4821 generic.go:334] "Generic (PLEG): container finished" podID="3fb2355d-5ec2-44f1-8665-a3d73c24f8a6" containerID="dfeb5a5d613f3284ed07af1e11e0bc565beb1e5303175ab7e151caf082203aed" exitCode=0 Nov 25 10:40:31 crc kubenswrapper[4821]: I1125 10:40:31.735605 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" event={"ID":"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6","Type":"ContainerDied","Data":"dfeb5a5d613f3284ed07af1e11e0bc565beb1e5303175ab7e151caf082203aed"} Nov 25 10:40:31 crc kubenswrapper[4821]: I1125 10:40:31.735611 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" Nov 25 10:40:31 crc kubenswrapper[4821]: I1125 10:40:31.735633 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-xnlvh" event={"ID":"3fb2355d-5ec2-44f1-8665-a3d73c24f8a6","Type":"ContainerDied","Data":"02e2ec598c0e3b1f1aaf65b4f15ea64649f966c51706c45004ed2b0fed21fef6"} Nov 25 10:40:31 crc kubenswrapper[4821]: I1125 10:40:31.735653 4821 scope.go:117] "RemoveContainer" containerID="dfeb5a5d613f3284ed07af1e11e0bc565beb1e5303175ab7e151caf082203aed" Nov 25 10:40:31 crc kubenswrapper[4821]: I1125 10:40:31.756852 4821 scope.go:117] "RemoveContainer" containerID="dfeb5a5d613f3284ed07af1e11e0bc565beb1e5303175ab7e151caf082203aed" Nov 25 10:40:31 crc kubenswrapper[4821]: E1125 10:40:31.757389 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dfeb5a5d613f3284ed07af1e11e0bc565beb1e5303175ab7e151caf082203aed\": container with ID starting with dfeb5a5d613f3284ed07af1e11e0bc565beb1e5303175ab7e151caf082203aed not found: ID does not exist" containerID="dfeb5a5d613f3284ed07af1e11e0bc565beb1e5303175ab7e151caf082203aed" Nov 25 10:40:31 crc kubenswrapper[4821]: I1125 10:40:31.757439 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dfeb5a5d613f3284ed07af1e11e0bc565beb1e5303175ab7e151caf082203aed"} err="failed to get container status \"dfeb5a5d613f3284ed07af1e11e0bc565beb1e5303175ab7e151caf082203aed\": rpc error: code = NotFound desc = could not find container \"dfeb5a5d613f3284ed07af1e11e0bc565beb1e5303175ab7e151caf082203aed\": container with ID starting with dfeb5a5d613f3284ed07af1e11e0bc565beb1e5303175ab7e151caf082203aed not found: ID does not exist" Nov 25 10:40:31 crc kubenswrapper[4821]: I1125 10:40:31.772397 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-xnlvh"] Nov 25 10:40:31 crc kubenswrapper[4821]: I1125 10:40:31.774975 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-xnlvh"] Nov 25 10:40:32 crc kubenswrapper[4821]: I1125 10:40:32.123815 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3fb2355d-5ec2-44f1-8665-a3d73c24f8a6" path="/var/lib/kubelet/pods/3fb2355d-5ec2-44f1-8665-a3d73c24f8a6/volumes" Nov 25 10:41:34 crc kubenswrapper[4821]: I1125 10:41:34.714736 4821 patch_prober.go:28] interesting pod/machine-config-daemon-2krbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:41:34 crc kubenswrapper[4821]: I1125 10:41:34.715258 4821 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:41:51 crc kubenswrapper[4821]: I1125 10:41:51.490220 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-g7gr9"] Nov 25 10:41:51 crc kubenswrapper[4821]: E1125 10:41:51.492029 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fb2355d-5ec2-44f1-8665-a3d73c24f8a6" containerName="registry" Nov 25 10:41:51 crc kubenswrapper[4821]: I1125 10:41:51.492133 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fb2355d-5ec2-44f1-8665-a3d73c24f8a6" containerName="registry" Nov 25 10:41:51 crc kubenswrapper[4821]: I1125 10:41:51.492384 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="3fb2355d-5ec2-44f1-8665-a3d73c24f8a6" containerName="registry" Nov 25 10:41:51 crc kubenswrapper[4821]: I1125 10:41:51.492923 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-g7gr9" Nov 25 10:41:51 crc kubenswrapper[4821]: I1125 10:41:51.494761 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Nov 25 10:41:51 crc kubenswrapper[4821]: I1125 10:41:51.495082 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Nov 25 10:41:51 crc kubenswrapper[4821]: I1125 10:41:51.495499 4821 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-8n882" Nov 25 10:41:51 crc kubenswrapper[4821]: I1125 10:41:51.503399 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-g7gr9"] Nov 25 10:41:51 crc kubenswrapper[4821]: I1125 10:41:51.516630 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-4qh5c"] Nov 25 10:41:51 crc kubenswrapper[4821]: I1125 10:41:51.517277 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-4qh5c" Nov 25 10:41:51 crc kubenswrapper[4821]: I1125 10:41:51.519852 4821 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-jvb4x" Nov 25 10:41:51 crc kubenswrapper[4821]: I1125 10:41:51.522192 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-crm7d"] Nov 25 10:41:51 crc kubenswrapper[4821]: I1125 10:41:51.523024 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-crm7d" Nov 25 10:41:51 crc kubenswrapper[4821]: I1125 10:41:51.526647 4821 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-twjvx" Nov 25 10:41:51 crc kubenswrapper[4821]: I1125 10:41:51.541873 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-crm7d"] Nov 25 10:41:51 crc kubenswrapper[4821]: I1125 10:41:51.549580 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-4qh5c"] Nov 25 10:41:51 crc kubenswrapper[4821]: I1125 10:41:51.648329 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phpcs\" (UniqueName: \"kubernetes.io/projected/4096527e-0fc1-4991-ad0c-f47a84cd6e84-kube-api-access-phpcs\") pod \"cert-manager-webhook-5655c58dd6-4qh5c\" (UID: \"4096527e-0fc1-4991-ad0c-f47a84cd6e84\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-4qh5c" Nov 25 10:41:51 crc kubenswrapper[4821]: I1125 10:41:51.648372 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvjhp\" (UniqueName: \"kubernetes.io/projected/c7f95e74-b964-455f-bacd-60e1fceb87fb-kube-api-access-vvjhp\") pod \"cert-manager-5b446d88c5-crm7d\" (UID: \"c7f95e74-b964-455f-bacd-60e1fceb87fb\") " pod="cert-manager/cert-manager-5b446d88c5-crm7d" Nov 25 10:41:51 crc kubenswrapper[4821]: I1125 10:41:51.648433 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5ks9\" (UniqueName: \"kubernetes.io/projected/baa81eb8-c73e-41d1-bd01-9958477aa031-kube-api-access-n5ks9\") pod \"cert-manager-cainjector-7f985d654d-g7gr9\" (UID: \"baa81eb8-c73e-41d1-bd01-9958477aa031\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-g7gr9" Nov 25 10:41:51 crc kubenswrapper[4821]: I1125 10:41:51.749656 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-phpcs\" (UniqueName: \"kubernetes.io/projected/4096527e-0fc1-4991-ad0c-f47a84cd6e84-kube-api-access-phpcs\") pod \"cert-manager-webhook-5655c58dd6-4qh5c\" (UID: \"4096527e-0fc1-4991-ad0c-f47a84cd6e84\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-4qh5c" Nov 25 10:41:51 crc kubenswrapper[4821]: I1125 10:41:51.750328 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvjhp\" (UniqueName: \"kubernetes.io/projected/c7f95e74-b964-455f-bacd-60e1fceb87fb-kube-api-access-vvjhp\") pod \"cert-manager-5b446d88c5-crm7d\" (UID: \"c7f95e74-b964-455f-bacd-60e1fceb87fb\") " pod="cert-manager/cert-manager-5b446d88c5-crm7d" Nov 25 10:41:51 crc kubenswrapper[4821]: I1125 10:41:51.750518 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5ks9\" (UniqueName: \"kubernetes.io/projected/baa81eb8-c73e-41d1-bd01-9958477aa031-kube-api-access-n5ks9\") pod \"cert-manager-cainjector-7f985d654d-g7gr9\" (UID: \"baa81eb8-c73e-41d1-bd01-9958477aa031\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-g7gr9" Nov 25 10:41:51 crc kubenswrapper[4821]: I1125 10:41:51.767721 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvjhp\" (UniqueName: \"kubernetes.io/projected/c7f95e74-b964-455f-bacd-60e1fceb87fb-kube-api-access-vvjhp\") pod \"cert-manager-5b446d88c5-crm7d\" (UID: \"c7f95e74-b964-455f-bacd-60e1fceb87fb\") " pod="cert-manager/cert-manager-5b446d88c5-crm7d" Nov 25 10:41:51 crc kubenswrapper[4821]: I1125 10:41:51.767793 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5ks9\" (UniqueName: \"kubernetes.io/projected/baa81eb8-c73e-41d1-bd01-9958477aa031-kube-api-access-n5ks9\") pod \"cert-manager-cainjector-7f985d654d-g7gr9\" (UID: \"baa81eb8-c73e-41d1-bd01-9958477aa031\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-g7gr9" Nov 25 10:41:51 crc kubenswrapper[4821]: I1125 10:41:51.767740 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-phpcs\" (UniqueName: \"kubernetes.io/projected/4096527e-0fc1-4991-ad0c-f47a84cd6e84-kube-api-access-phpcs\") pod \"cert-manager-webhook-5655c58dd6-4qh5c\" (UID: \"4096527e-0fc1-4991-ad0c-f47a84cd6e84\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-4qh5c" Nov 25 10:41:51 crc kubenswrapper[4821]: I1125 10:41:51.811073 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-g7gr9" Nov 25 10:41:51 crc kubenswrapper[4821]: I1125 10:41:51.834021 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-4qh5c" Nov 25 10:41:51 crc kubenswrapper[4821]: I1125 10:41:51.844332 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-crm7d" Nov 25 10:41:52 crc kubenswrapper[4821]: I1125 10:41:52.046901 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-g7gr9"] Nov 25 10:41:52 crc kubenswrapper[4821]: I1125 10:41:52.054107 4821 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 10:41:52 crc kubenswrapper[4821]: I1125 10:41:52.127517 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-g7gr9" event={"ID":"baa81eb8-c73e-41d1-bd01-9958477aa031","Type":"ContainerStarted","Data":"cd8ada7f628cb1cc2d120848c4407e19147179d5cec5d817958a9b2fd499f4e9"} Nov 25 10:41:52 crc kubenswrapper[4821]: I1125 10:41:52.306009 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-crm7d"] Nov 25 10:41:52 crc kubenswrapper[4821]: I1125 10:41:52.309177 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-4qh5c"] Nov 25 10:41:52 crc kubenswrapper[4821]: W1125 10:41:52.317924 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc7f95e74_b964_455f_bacd_60e1fceb87fb.slice/crio-9c6fa6fd1fdfc5b1c949c91f62f6656e1748e197be34b769be452da06196ee46 WatchSource:0}: Error finding container 9c6fa6fd1fdfc5b1c949c91f62f6656e1748e197be34b769be452da06196ee46: Status 404 returned error can't find the container with id 9c6fa6fd1fdfc5b1c949c91f62f6656e1748e197be34b769be452da06196ee46 Nov 25 10:41:52 crc kubenswrapper[4821]: W1125 10:41:52.318191 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4096527e_0fc1_4991_ad0c_f47a84cd6e84.slice/crio-e12e367f12bc33bacafd908c95436f0ff854133687f3357a2bc6847721062f69 WatchSource:0}: Error finding container e12e367f12bc33bacafd908c95436f0ff854133687f3357a2bc6847721062f69: Status 404 returned error can't find the container with id e12e367f12bc33bacafd908c95436f0ff854133687f3357a2bc6847721062f69 Nov 25 10:41:53 crc kubenswrapper[4821]: I1125 10:41:53.148883 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-crm7d" event={"ID":"c7f95e74-b964-455f-bacd-60e1fceb87fb","Type":"ContainerStarted","Data":"9c6fa6fd1fdfc5b1c949c91f62f6656e1748e197be34b769be452da06196ee46"} Nov 25 10:41:53 crc kubenswrapper[4821]: I1125 10:41:53.149972 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-4qh5c" event={"ID":"4096527e-0fc1-4991-ad0c-f47a84cd6e84","Type":"ContainerStarted","Data":"e12e367f12bc33bacafd908c95436f0ff854133687f3357a2bc6847721062f69"} Nov 25 10:41:56 crc kubenswrapper[4821]: I1125 10:41:56.168782 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-g7gr9" event={"ID":"baa81eb8-c73e-41d1-bd01-9958477aa031","Type":"ContainerStarted","Data":"f5cbfa385af5a064d7d7ff9b3dbe26bea8af0d8915620f01763c582d661ca58f"} Nov 25 10:41:56 crc kubenswrapper[4821]: I1125 10:41:56.171689 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-4qh5c" event={"ID":"4096527e-0fc1-4991-ad0c-f47a84cd6e84","Type":"ContainerStarted","Data":"21fd198ba620cd65862ba2869ee194e0d9cdf3b967acf443f5ddd8f84a280f74"} Nov 25 10:41:56 crc kubenswrapper[4821]: I1125 10:41:56.171790 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-4qh5c" Nov 25 10:41:56 crc kubenswrapper[4821]: I1125 10:41:56.172914 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-crm7d" event={"ID":"c7f95e74-b964-455f-bacd-60e1fceb87fb","Type":"ContainerStarted","Data":"6de583956f79af10ca8010076d3571793ddd475854e82275d6f9e70aba8b96ec"} Nov 25 10:41:56 crc kubenswrapper[4821]: I1125 10:41:56.182907 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-g7gr9" podStartSLOduration=2.121641239 podStartE2EDuration="5.182890655s" podCreationTimestamp="2025-11-25 10:41:51 +0000 UTC" firstStartedPulling="2025-11-25 10:41:52.053784237 +0000 UTC m=+582.590104084" lastFinishedPulling="2025-11-25 10:41:55.115033653 +0000 UTC m=+585.651353500" observedRunningTime="2025-11-25 10:41:56.182691481 +0000 UTC m=+586.719011378" watchObservedRunningTime="2025-11-25 10:41:56.182890655 +0000 UTC m=+586.719210502" Nov 25 10:41:56 crc kubenswrapper[4821]: I1125 10:41:56.201998 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-4qh5c" podStartSLOduration=2.410485531 podStartE2EDuration="5.20197306s" podCreationTimestamp="2025-11-25 10:41:51 +0000 UTC" firstStartedPulling="2025-11-25 10:41:52.320478006 +0000 UTC m=+582.856797853" lastFinishedPulling="2025-11-25 10:41:55.111965525 +0000 UTC m=+585.648285382" observedRunningTime="2025-11-25 10:41:56.197084056 +0000 UTC m=+586.733403903" watchObservedRunningTime="2025-11-25 10:41:56.20197306 +0000 UTC m=+586.738292927" Nov 25 10:41:56 crc kubenswrapper[4821]: I1125 10:41:56.215312 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-crm7d" podStartSLOduration=2.353169065 podStartE2EDuration="5.215285254s" podCreationTimestamp="2025-11-25 10:41:51 +0000 UTC" firstStartedPulling="2025-11-25 10:41:52.32066352 +0000 UTC m=+582.856983387" lastFinishedPulling="2025-11-25 10:41:55.182779729 +0000 UTC m=+585.719099576" observedRunningTime="2025-11-25 10:41:56.209903041 +0000 UTC m=+586.746222928" watchObservedRunningTime="2025-11-25 10:41:56.215285254 +0000 UTC m=+586.751605121" Nov 25 10:42:01 crc kubenswrapper[4821]: I1125 10:42:01.837052 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-4qh5c" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.094656 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-mzhnw"] Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.095459 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" podUID="954085fa-d03a-4e88-9dfd-566257e3558d" containerName="ovn-controller" containerID="cri-o://ed2e7cad7bb24b5b56a1295098a6a7c230c8f9b9c29a88ef653ae8b6ac31667d" gracePeriod=30 Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.095520 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" podUID="954085fa-d03a-4e88-9dfd-566257e3558d" containerName="nbdb" containerID="cri-o://e9d7254196b3702e48083ae98e829b73c539adf69ed5b2bbdc62189d09ae1762" gracePeriod=30 Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.095528 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" podUID="954085fa-d03a-4e88-9dfd-566257e3558d" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://1a14a078ba7d2aed95992b36502008f315f0d3a634b64ed49da70398b2562d44" gracePeriod=30 Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.095591 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" podUID="954085fa-d03a-4e88-9dfd-566257e3558d" containerName="kube-rbac-proxy-node" containerID="cri-o://2f00746930c4588e548f0e2b60d52bf448aff7401b676a7855cb0f4cce00d329" gracePeriod=30 Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.095602 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" podUID="954085fa-d03a-4e88-9dfd-566257e3558d" containerName="northd" containerID="cri-o://d727430108977574365f0e9c04b44e03307bd7ff89cdbb1e48a6e916f504d6b0" gracePeriod=30 Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.095661 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" podUID="954085fa-d03a-4e88-9dfd-566257e3558d" containerName="sbdb" containerID="cri-o://7dbf499f4ae6161752abaa7a466a027c6f874016be90a534e69f07e3dd3f3d9f" gracePeriod=30 Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.095631 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" podUID="954085fa-d03a-4e88-9dfd-566257e3558d" containerName="ovn-acl-logging" containerID="cri-o://94c7949f077b1ca2b310bedb090aa72fc4bd7e8f2c8bf438096edfca8568f36e" gracePeriod=30 Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.127823 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" podUID="954085fa-d03a-4e88-9dfd-566257e3558d" containerName="ovnkube-controller" containerID="cri-o://32de39e5841057630116280f4a9e5e5728b786db001c2c0db0d57e7475a3a16d" gracePeriod=30 Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.203666 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-ldzbz_b84e32ae-4897-4644-a656-6bd82012cc81/kube-multus/2.log" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.204202 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-ldzbz_b84e32ae-4897-4644-a656-6bd82012cc81/kube-multus/1.log" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.204243 4821 generic.go:334] "Generic (PLEG): container finished" podID="b84e32ae-4897-4644-a656-6bd82012cc81" containerID="86b9194a6cea3434f70eb5f6d91aa470cb4321b386c78735cf4ca3da718f1026" exitCode=2 Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.204274 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-ldzbz" event={"ID":"b84e32ae-4897-4644-a656-6bd82012cc81","Type":"ContainerDied","Data":"86b9194a6cea3434f70eb5f6d91aa470cb4321b386c78735cf4ca3da718f1026"} Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.204310 4821 scope.go:117] "RemoveContainer" containerID="418dc8577fdc98324677dbccfcf9317b67cab4408432b30df2818b22f88d426e" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.204845 4821 scope.go:117] "RemoveContainer" containerID="86b9194a6cea3434f70eb5f6d91aa470cb4321b386c78735cf4ca3da718f1026" Nov 25 10:42:02 crc kubenswrapper[4821]: E1125 10:42:02.205029 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-ldzbz_openshift-multus(b84e32ae-4897-4644-a656-6bd82012cc81)\"" pod="openshift-multus/multus-ldzbz" podUID="b84e32ae-4897-4644-a656-6bd82012cc81" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.431786 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mzhnw_954085fa-d03a-4e88-9dfd-566257e3558d/ovnkube-controller/3.log" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.434738 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mzhnw_954085fa-d03a-4e88-9dfd-566257e3558d/ovn-acl-logging/0.log" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.435656 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mzhnw_954085fa-d03a-4e88-9dfd-566257e3558d/ovn-controller/0.log" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.436102 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.486048 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-fvjc2"] Nov 25 10:42:02 crc kubenswrapper[4821]: E1125 10:42:02.486324 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="954085fa-d03a-4e88-9dfd-566257e3558d" containerName="northd" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.486342 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="954085fa-d03a-4e88-9dfd-566257e3558d" containerName="northd" Nov 25 10:42:02 crc kubenswrapper[4821]: E1125 10:42:02.486355 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="954085fa-d03a-4e88-9dfd-566257e3558d" containerName="kubecfg-setup" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.486364 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="954085fa-d03a-4e88-9dfd-566257e3558d" containerName="kubecfg-setup" Nov 25 10:42:02 crc kubenswrapper[4821]: E1125 10:42:02.486374 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="954085fa-d03a-4e88-9dfd-566257e3558d" containerName="ovnkube-controller" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.486383 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="954085fa-d03a-4e88-9dfd-566257e3558d" containerName="ovnkube-controller" Nov 25 10:42:02 crc kubenswrapper[4821]: E1125 10:42:02.486391 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="954085fa-d03a-4e88-9dfd-566257e3558d" containerName="ovnkube-controller" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.486398 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="954085fa-d03a-4e88-9dfd-566257e3558d" containerName="ovnkube-controller" Nov 25 10:42:02 crc kubenswrapper[4821]: E1125 10:42:02.486406 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="954085fa-d03a-4e88-9dfd-566257e3558d" containerName="kube-rbac-proxy-node" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.486414 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="954085fa-d03a-4e88-9dfd-566257e3558d" containerName="kube-rbac-proxy-node" Nov 25 10:42:02 crc kubenswrapper[4821]: E1125 10:42:02.486433 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="954085fa-d03a-4e88-9dfd-566257e3558d" containerName="ovn-acl-logging" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.486441 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="954085fa-d03a-4e88-9dfd-566257e3558d" containerName="ovn-acl-logging" Nov 25 10:42:02 crc kubenswrapper[4821]: E1125 10:42:02.486449 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="954085fa-d03a-4e88-9dfd-566257e3558d" containerName="ovnkube-controller" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.486455 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="954085fa-d03a-4e88-9dfd-566257e3558d" containerName="ovnkube-controller" Nov 25 10:42:02 crc kubenswrapper[4821]: E1125 10:42:02.486464 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="954085fa-d03a-4e88-9dfd-566257e3558d" containerName="kube-rbac-proxy-ovn-metrics" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.486471 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="954085fa-d03a-4e88-9dfd-566257e3558d" containerName="kube-rbac-proxy-ovn-metrics" Nov 25 10:42:02 crc kubenswrapper[4821]: E1125 10:42:02.486478 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="954085fa-d03a-4e88-9dfd-566257e3558d" containerName="nbdb" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.486486 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="954085fa-d03a-4e88-9dfd-566257e3558d" containerName="nbdb" Nov 25 10:42:02 crc kubenswrapper[4821]: E1125 10:42:02.486495 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="954085fa-d03a-4e88-9dfd-566257e3558d" containerName="ovn-controller" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.486502 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="954085fa-d03a-4e88-9dfd-566257e3558d" containerName="ovn-controller" Nov 25 10:42:02 crc kubenswrapper[4821]: E1125 10:42:02.486516 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="954085fa-d03a-4e88-9dfd-566257e3558d" containerName="sbdb" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.486524 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="954085fa-d03a-4e88-9dfd-566257e3558d" containerName="sbdb" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.487061 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="954085fa-d03a-4e88-9dfd-566257e3558d" containerName="ovn-controller" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.487073 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="954085fa-d03a-4e88-9dfd-566257e3558d" containerName="sbdb" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.487082 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="954085fa-d03a-4e88-9dfd-566257e3558d" containerName="kube-rbac-proxy-ovn-metrics" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.487091 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="954085fa-d03a-4e88-9dfd-566257e3558d" containerName="ovnkube-controller" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.487098 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="954085fa-d03a-4e88-9dfd-566257e3558d" containerName="nbdb" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.487111 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="954085fa-d03a-4e88-9dfd-566257e3558d" containerName="kube-rbac-proxy-node" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.487149 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="954085fa-d03a-4e88-9dfd-566257e3558d" containerName="ovnkube-controller" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.487176 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="954085fa-d03a-4e88-9dfd-566257e3558d" containerName="ovnkube-controller" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.487186 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="954085fa-d03a-4e88-9dfd-566257e3558d" containerName="ovn-acl-logging" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.487195 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="954085fa-d03a-4e88-9dfd-566257e3558d" containerName="northd" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.487205 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="954085fa-d03a-4e88-9dfd-566257e3558d" containerName="ovnkube-controller" Nov 25 10:42:02 crc kubenswrapper[4821]: E1125 10:42:02.487312 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="954085fa-d03a-4e88-9dfd-566257e3558d" containerName="ovnkube-controller" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.487323 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="954085fa-d03a-4e88-9dfd-566257e3558d" containerName="ovnkube-controller" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.487423 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="954085fa-d03a-4e88-9dfd-566257e3558d" containerName="ovnkube-controller" Nov 25 10:42:02 crc kubenswrapper[4821]: E1125 10:42:02.487554 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="954085fa-d03a-4e88-9dfd-566257e3558d" containerName="ovnkube-controller" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.487565 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="954085fa-d03a-4e88-9dfd-566257e3558d" containerName="ovnkube-controller" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.489326 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.575688 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/954085fa-d03a-4e88-9dfd-566257e3558d-env-overrides\") pod \"954085fa-d03a-4e88-9dfd-566257e3558d\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.575961 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/954085fa-d03a-4e88-9dfd-566257e3558d-ovnkube-config\") pod \"954085fa-d03a-4e88-9dfd-566257e3558d\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.576069 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-host-kubelet\") pod \"954085fa-d03a-4e88-9dfd-566257e3558d\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.576185 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-run-ovn\") pod \"954085fa-d03a-4e88-9dfd-566257e3558d\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.576286 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-host-cni-bin\") pod \"954085fa-d03a-4e88-9dfd-566257e3558d\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.576492 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"954085fa-d03a-4e88-9dfd-566257e3558d\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.576395 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "954085fa-d03a-4e88-9dfd-566257e3558d" (UID: "954085fa-d03a-4e88-9dfd-566257e3558d"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.576423 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "954085fa-d03a-4e88-9dfd-566257e3558d" (UID: "954085fa-d03a-4e88-9dfd-566257e3558d"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.576465 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "954085fa-d03a-4e88-9dfd-566257e3558d" (UID: "954085fa-d03a-4e88-9dfd-566257e3558d"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.576534 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "954085fa-d03a-4e88-9dfd-566257e3558d" (UID: "954085fa-d03a-4e88-9dfd-566257e3558d"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.576607 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-node-log\") pod \"954085fa-d03a-4e88-9dfd-566257e3558d\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.576701 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-systemd-units\") pod \"954085fa-d03a-4e88-9dfd-566257e3558d\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.576733 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-host-run-ovn-kubernetes\") pod \"954085fa-d03a-4e88-9dfd-566257e3558d\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.576755 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-log-socket\") pod \"954085fa-d03a-4e88-9dfd-566257e3558d\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.576769 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/954085fa-d03a-4e88-9dfd-566257e3558d-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "954085fa-d03a-4e88-9dfd-566257e3558d" (UID: "954085fa-d03a-4e88-9dfd-566257e3558d"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.576788 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-run-systemd\") pod \"954085fa-d03a-4e88-9dfd-566257e3558d\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.576798 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "954085fa-d03a-4e88-9dfd-566257e3558d" (UID: "954085fa-d03a-4e88-9dfd-566257e3558d"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.576797 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/954085fa-d03a-4e88-9dfd-566257e3558d-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "954085fa-d03a-4e88-9dfd-566257e3558d" (UID: "954085fa-d03a-4e88-9dfd-566257e3558d"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.576818 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-log-socket" (OuterVolumeSpecName: "log-socket") pod "954085fa-d03a-4e88-9dfd-566257e3558d" (UID: "954085fa-d03a-4e88-9dfd-566257e3558d"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.576824 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "954085fa-d03a-4e88-9dfd-566257e3558d" (UID: "954085fa-d03a-4e88-9dfd-566257e3558d"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.576881 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "954085fa-d03a-4e88-9dfd-566257e3558d" (UID: "954085fa-d03a-4e88-9dfd-566257e3558d"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.576807 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-etc-openvswitch\") pod \"954085fa-d03a-4e88-9dfd-566257e3558d\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.576992 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-var-lib-openvswitch\") pod \"954085fa-d03a-4e88-9dfd-566257e3558d\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.577019 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-host-run-netns\") pod \"954085fa-d03a-4e88-9dfd-566257e3558d\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.577048 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/954085fa-d03a-4e88-9dfd-566257e3558d-ovn-node-metrics-cert\") pod \"954085fa-d03a-4e88-9dfd-566257e3558d\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.577076 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-run-openvswitch\") pod \"954085fa-d03a-4e88-9dfd-566257e3558d\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.577098 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "954085fa-d03a-4e88-9dfd-566257e3558d" (UID: "954085fa-d03a-4e88-9dfd-566257e3558d"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.577105 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/954085fa-d03a-4e88-9dfd-566257e3558d-ovnkube-script-lib\") pod \"954085fa-d03a-4e88-9dfd-566257e3558d\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.577191 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-host-cni-netd\") pod \"954085fa-d03a-4e88-9dfd-566257e3558d\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.577217 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hrb87\" (UniqueName: \"kubernetes.io/projected/954085fa-d03a-4e88-9dfd-566257e3558d-kube-api-access-hrb87\") pod \"954085fa-d03a-4e88-9dfd-566257e3558d\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.577255 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-host-slash\") pod \"954085fa-d03a-4e88-9dfd-566257e3558d\" (UID: \"954085fa-d03a-4e88-9dfd-566257e3558d\") " Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.577552 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/954085fa-d03a-4e88-9dfd-566257e3558d-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "954085fa-d03a-4e88-9dfd-566257e3558d" (UID: "954085fa-d03a-4e88-9dfd-566257e3558d"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.577595 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-host-slash" (OuterVolumeSpecName: "host-slash") pod "954085fa-d03a-4e88-9dfd-566257e3558d" (UID: "954085fa-d03a-4e88-9dfd-566257e3558d"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.577621 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "954085fa-d03a-4e88-9dfd-566257e3558d" (UID: "954085fa-d03a-4e88-9dfd-566257e3558d"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.577558 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-node-log" (OuterVolumeSpecName: "node-log") pod "954085fa-d03a-4e88-9dfd-566257e3558d" (UID: "954085fa-d03a-4e88-9dfd-566257e3558d"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.577767 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "954085fa-d03a-4e88-9dfd-566257e3558d" (UID: "954085fa-d03a-4e88-9dfd-566257e3558d"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.577801 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "954085fa-d03a-4e88-9dfd-566257e3558d" (UID: "954085fa-d03a-4e88-9dfd-566257e3558d"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.577778 4821 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.577956 4821 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.578025 4821 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/954085fa-d03a-4e88-9dfd-566257e3558d-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.578087 4821 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-host-cni-netd\") on node \"crc\" DevicePath \"\"" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.578152 4821 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-host-slash\") on node \"crc\" DevicePath \"\"" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.578248 4821 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/954085fa-d03a-4e88-9dfd-566257e3558d-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.578308 4821 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/954085fa-d03a-4e88-9dfd-566257e3558d-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.578377 4821 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-host-kubelet\") on node \"crc\" DevicePath \"\"" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.578434 4821 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.578493 4821 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-host-cni-bin\") on node \"crc\" DevicePath \"\"" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.578554 4821 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.578617 4821 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-systemd-units\") on node \"crc\" DevicePath \"\"" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.578678 4821 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.578739 4821 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-log-socket\") on node \"crc\" DevicePath \"\"" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.582454 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/954085fa-d03a-4e88-9dfd-566257e3558d-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "954085fa-d03a-4e88-9dfd-566257e3558d" (UID: "954085fa-d03a-4e88-9dfd-566257e3558d"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.583249 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/954085fa-d03a-4e88-9dfd-566257e3558d-kube-api-access-hrb87" (OuterVolumeSpecName: "kube-api-access-hrb87") pod "954085fa-d03a-4e88-9dfd-566257e3558d" (UID: "954085fa-d03a-4e88-9dfd-566257e3558d"). InnerVolumeSpecName "kube-api-access-hrb87". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.589600 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "954085fa-d03a-4e88-9dfd-566257e3558d" (UID: "954085fa-d03a-4e88-9dfd-566257e3558d"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.680251 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7-host-run-ovn-kubernetes\") pod \"ovnkube-node-fvjc2\" (UID: \"d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7\") " pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.680307 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7-ovnkube-config\") pod \"ovnkube-node-fvjc2\" (UID: \"d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7\") " pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.680336 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7-ovnkube-script-lib\") pod \"ovnkube-node-fvjc2\" (UID: \"d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7\") " pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.680354 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7-systemd-units\") pod \"ovnkube-node-fvjc2\" (UID: \"d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7\") " pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.680370 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7-ovn-node-metrics-cert\") pod \"ovnkube-node-fvjc2\" (UID: \"d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7\") " pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.680385 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7-run-openvswitch\") pod \"ovnkube-node-fvjc2\" (UID: \"d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7\") " pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.680409 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7-var-lib-openvswitch\") pod \"ovnkube-node-fvjc2\" (UID: \"d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7\") " pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.680487 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-fvjc2\" (UID: \"d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7\") " pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.680554 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7-env-overrides\") pod \"ovnkube-node-fvjc2\" (UID: \"d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7\") " pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.680593 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7-host-cni-bin\") pod \"ovnkube-node-fvjc2\" (UID: \"d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7\") " pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.680684 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7-run-systemd\") pod \"ovnkube-node-fvjc2\" (UID: \"d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7\") " pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.680714 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7-host-kubelet\") pod \"ovnkube-node-fvjc2\" (UID: \"d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7\") " pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.680745 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7-run-ovn\") pod \"ovnkube-node-fvjc2\" (UID: \"d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7\") " pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.680769 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7-host-slash\") pod \"ovnkube-node-fvjc2\" (UID: \"d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7\") " pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.680808 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7-host-cni-netd\") pod \"ovnkube-node-fvjc2\" (UID: \"d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7\") " pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.680835 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7-log-socket\") pod \"ovnkube-node-fvjc2\" (UID: \"d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7\") " pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.680857 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7-etc-openvswitch\") pod \"ovnkube-node-fvjc2\" (UID: \"d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7\") " pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.680887 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7-host-run-netns\") pod \"ovnkube-node-fvjc2\" (UID: \"d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7\") " pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.680914 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7-node-log\") pod \"ovnkube-node-fvjc2\" (UID: \"d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7\") " pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.680946 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hg8df\" (UniqueName: \"kubernetes.io/projected/d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7-kube-api-access-hg8df\") pod \"ovnkube-node-fvjc2\" (UID: \"d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7\") " pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.681023 4821 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-node-log\") on node \"crc\" DevicePath \"\"" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.681039 4821 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-run-systemd\") on node \"crc\" DevicePath \"\"" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.681051 4821 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-host-run-netns\") on node \"crc\" DevicePath \"\"" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.681064 4821 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/954085fa-d03a-4e88-9dfd-566257e3558d-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.681076 4821 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/954085fa-d03a-4e88-9dfd-566257e3558d-run-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.681088 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hrb87\" (UniqueName: \"kubernetes.io/projected/954085fa-d03a-4e88-9dfd-566257e3558d-kube-api-access-hrb87\") on node \"crc\" DevicePath \"\"" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.782036 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7-run-systemd\") pod \"ovnkube-node-fvjc2\" (UID: \"d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7\") " pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.782109 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7-host-kubelet\") pod \"ovnkube-node-fvjc2\" (UID: \"d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7\") " pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.782132 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7-run-ovn\") pod \"ovnkube-node-fvjc2\" (UID: \"d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7\") " pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.782156 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7-host-slash\") pod \"ovnkube-node-fvjc2\" (UID: \"d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7\") " pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.782222 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7-host-cni-netd\") pod \"ovnkube-node-fvjc2\" (UID: \"d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7\") " pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.782240 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7-log-socket\") pod \"ovnkube-node-fvjc2\" (UID: \"d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7\") " pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.782255 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7-etc-openvswitch\") pod \"ovnkube-node-fvjc2\" (UID: \"d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7\") " pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.782260 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7-run-systemd\") pod \"ovnkube-node-fvjc2\" (UID: \"d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7\") " pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.782279 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7-host-kubelet\") pod \"ovnkube-node-fvjc2\" (UID: \"d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7\") " pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.782329 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7-host-slash\") pod \"ovnkube-node-fvjc2\" (UID: \"d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7\") " pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.782341 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7-log-socket\") pod \"ovnkube-node-fvjc2\" (UID: \"d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7\") " pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.782376 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7-host-cni-netd\") pod \"ovnkube-node-fvjc2\" (UID: \"d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7\") " pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.782378 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7-etc-openvswitch\") pod \"ovnkube-node-fvjc2\" (UID: \"d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7\") " pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.782273 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7-host-run-netns\") pod \"ovnkube-node-fvjc2\" (UID: \"d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7\") " pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.782396 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7-run-ovn\") pod \"ovnkube-node-fvjc2\" (UID: \"d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7\") " pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.782311 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7-host-run-netns\") pod \"ovnkube-node-fvjc2\" (UID: \"d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7\") " pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.782434 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7-node-log\") pod \"ovnkube-node-fvjc2\" (UID: \"d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7\") " pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.782611 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hg8df\" (UniqueName: \"kubernetes.io/projected/d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7-kube-api-access-hg8df\") pod \"ovnkube-node-fvjc2\" (UID: \"d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7\") " pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.782457 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7-node-log\") pod \"ovnkube-node-fvjc2\" (UID: \"d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7\") " pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.782683 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7-host-run-ovn-kubernetes\") pod \"ovnkube-node-fvjc2\" (UID: \"d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7\") " pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.782723 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7-ovnkube-config\") pod \"ovnkube-node-fvjc2\" (UID: \"d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7\") " pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.782788 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7-ovnkube-script-lib\") pod \"ovnkube-node-fvjc2\" (UID: \"d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7\") " pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.782815 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7-systemd-units\") pod \"ovnkube-node-fvjc2\" (UID: \"d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7\") " pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.782848 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7-ovn-node-metrics-cert\") pod \"ovnkube-node-fvjc2\" (UID: \"d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7\") " pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.782879 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7-run-openvswitch\") pod \"ovnkube-node-fvjc2\" (UID: \"d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7\") " pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.782950 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7-var-lib-openvswitch\") pod \"ovnkube-node-fvjc2\" (UID: \"d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7\") " pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.782977 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-fvjc2\" (UID: \"d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7\") " pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.783012 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7-env-overrides\") pod \"ovnkube-node-fvjc2\" (UID: \"d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7\") " pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.783057 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7-host-cni-bin\") pod \"ovnkube-node-fvjc2\" (UID: \"d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7\") " pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.783224 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7-host-cni-bin\") pod \"ovnkube-node-fvjc2\" (UID: \"d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7\") " pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.783269 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7-run-openvswitch\") pod \"ovnkube-node-fvjc2\" (UID: \"d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7\") " pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.783310 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-fvjc2\" (UID: \"d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7\") " pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.783369 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7-host-run-ovn-kubernetes\") pod \"ovnkube-node-fvjc2\" (UID: \"d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7\") " pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.783369 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7-var-lib-openvswitch\") pod \"ovnkube-node-fvjc2\" (UID: \"d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7\") " pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.783439 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7-systemd-units\") pod \"ovnkube-node-fvjc2\" (UID: \"d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7\") " pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.784420 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7-ovnkube-script-lib\") pod \"ovnkube-node-fvjc2\" (UID: \"d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7\") " pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.785089 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7-env-overrides\") pod \"ovnkube-node-fvjc2\" (UID: \"d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7\") " pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.785462 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7-ovnkube-config\") pod \"ovnkube-node-fvjc2\" (UID: \"d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7\") " pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.789080 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7-ovn-node-metrics-cert\") pod \"ovnkube-node-fvjc2\" (UID: \"d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7\") " pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.799234 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hg8df\" (UniqueName: \"kubernetes.io/projected/d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7-kube-api-access-hg8df\") pod \"ovnkube-node-fvjc2\" (UID: \"d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7\") " pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" Nov 25 10:42:02 crc kubenswrapper[4821]: I1125 10:42:02.807119 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.214022 4821 generic.go:334] "Generic (PLEG): container finished" podID="d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7" containerID="bb9131c18c30706eb17a899587d8a22a55e76fb6891f3f15a954cf3ab45df88f" exitCode=0 Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.214088 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" event={"ID":"d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7","Type":"ContainerDied","Data":"bb9131c18c30706eb17a899587d8a22a55e76fb6891f3f15a954cf3ab45df88f"} Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.214113 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" event={"ID":"d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7","Type":"ContainerStarted","Data":"f89a69baec9e4c479cf6886f36b806a568674b4ba9557fc7322cdfd9f0683ff4"} Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.217638 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-ldzbz_b84e32ae-4897-4644-a656-6bd82012cc81/kube-multus/2.log" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.223855 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mzhnw_954085fa-d03a-4e88-9dfd-566257e3558d/ovnkube-controller/3.log" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.226590 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mzhnw_954085fa-d03a-4e88-9dfd-566257e3558d/ovn-acl-logging/0.log" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.227106 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-mzhnw_954085fa-d03a-4e88-9dfd-566257e3558d/ovn-controller/0.log" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.227471 4821 generic.go:334] "Generic (PLEG): container finished" podID="954085fa-d03a-4e88-9dfd-566257e3558d" containerID="32de39e5841057630116280f4a9e5e5728b786db001c2c0db0d57e7475a3a16d" exitCode=0 Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.227491 4821 generic.go:334] "Generic (PLEG): container finished" podID="954085fa-d03a-4e88-9dfd-566257e3558d" containerID="7dbf499f4ae6161752abaa7a466a027c6f874016be90a534e69f07e3dd3f3d9f" exitCode=0 Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.227499 4821 generic.go:334] "Generic (PLEG): container finished" podID="954085fa-d03a-4e88-9dfd-566257e3558d" containerID="e9d7254196b3702e48083ae98e829b73c539adf69ed5b2bbdc62189d09ae1762" exitCode=0 Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.227549 4821 generic.go:334] "Generic (PLEG): container finished" podID="954085fa-d03a-4e88-9dfd-566257e3558d" containerID="d727430108977574365f0e9c04b44e03307bd7ff89cdbb1e48a6e916f504d6b0" exitCode=0 Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.227557 4821 generic.go:334] "Generic (PLEG): container finished" podID="954085fa-d03a-4e88-9dfd-566257e3558d" containerID="1a14a078ba7d2aed95992b36502008f315f0d3a634b64ed49da70398b2562d44" exitCode=0 Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.227563 4821 generic.go:334] "Generic (PLEG): container finished" podID="954085fa-d03a-4e88-9dfd-566257e3558d" containerID="2f00746930c4588e548f0e2b60d52bf448aff7401b676a7855cb0f4cce00d329" exitCode=0 Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.227569 4821 generic.go:334] "Generic (PLEG): container finished" podID="954085fa-d03a-4e88-9dfd-566257e3558d" containerID="94c7949f077b1ca2b310bedb090aa72fc4bd7e8f2c8bf438096edfca8568f36e" exitCode=143 Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.227575 4821 generic.go:334] "Generic (PLEG): container finished" podID="954085fa-d03a-4e88-9dfd-566257e3558d" containerID="ed2e7cad7bb24b5b56a1295098a6a7c230c8f9b9c29a88ef653ae8b6ac31667d" exitCode=143 Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.227577 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.227528 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" event={"ID":"954085fa-d03a-4e88-9dfd-566257e3558d","Type":"ContainerDied","Data":"32de39e5841057630116280f4a9e5e5728b786db001c2c0db0d57e7475a3a16d"} Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.227684 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" event={"ID":"954085fa-d03a-4e88-9dfd-566257e3558d","Type":"ContainerDied","Data":"7dbf499f4ae6161752abaa7a466a027c6f874016be90a534e69f07e3dd3f3d9f"} Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.227695 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" event={"ID":"954085fa-d03a-4e88-9dfd-566257e3558d","Type":"ContainerDied","Data":"e9d7254196b3702e48083ae98e829b73c539adf69ed5b2bbdc62189d09ae1762"} Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.227703 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" event={"ID":"954085fa-d03a-4e88-9dfd-566257e3558d","Type":"ContainerDied","Data":"d727430108977574365f0e9c04b44e03307bd7ff89cdbb1e48a6e916f504d6b0"} Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.227711 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" event={"ID":"954085fa-d03a-4e88-9dfd-566257e3558d","Type":"ContainerDied","Data":"1a14a078ba7d2aed95992b36502008f315f0d3a634b64ed49da70398b2562d44"} Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.227720 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" event={"ID":"954085fa-d03a-4e88-9dfd-566257e3558d","Type":"ContainerDied","Data":"2f00746930c4588e548f0e2b60d52bf448aff7401b676a7855cb0f4cce00d329"} Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.227729 4821 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"90d8792a79993c04a0eb774498395d955deff5c42c693892426685d157cf9cff"} Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.227740 4821 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7dbf499f4ae6161752abaa7a466a027c6f874016be90a534e69f07e3dd3f3d9f"} Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.227745 4821 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e9d7254196b3702e48083ae98e829b73c539adf69ed5b2bbdc62189d09ae1762"} Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.227751 4821 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d727430108977574365f0e9c04b44e03307bd7ff89cdbb1e48a6e916f504d6b0"} Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.227755 4821 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1a14a078ba7d2aed95992b36502008f315f0d3a634b64ed49da70398b2562d44"} Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.227760 4821 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2f00746930c4588e548f0e2b60d52bf448aff7401b676a7855cb0f4cce00d329"} Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.227765 4821 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"94c7949f077b1ca2b310bedb090aa72fc4bd7e8f2c8bf438096edfca8568f36e"} Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.227769 4821 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ed2e7cad7bb24b5b56a1295098a6a7c230c8f9b9c29a88ef653ae8b6ac31667d"} Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.227774 4821 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3"} Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.227781 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" event={"ID":"954085fa-d03a-4e88-9dfd-566257e3558d","Type":"ContainerDied","Data":"94c7949f077b1ca2b310bedb090aa72fc4bd7e8f2c8bf438096edfca8568f36e"} Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.227788 4821 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"32de39e5841057630116280f4a9e5e5728b786db001c2c0db0d57e7475a3a16d"} Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.227794 4821 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"90d8792a79993c04a0eb774498395d955deff5c42c693892426685d157cf9cff"} Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.227799 4821 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7dbf499f4ae6161752abaa7a466a027c6f874016be90a534e69f07e3dd3f3d9f"} Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.227804 4821 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e9d7254196b3702e48083ae98e829b73c539adf69ed5b2bbdc62189d09ae1762"} Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.227808 4821 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d727430108977574365f0e9c04b44e03307bd7ff89cdbb1e48a6e916f504d6b0"} Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.227813 4821 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1a14a078ba7d2aed95992b36502008f315f0d3a634b64ed49da70398b2562d44"} Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.227818 4821 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2f00746930c4588e548f0e2b60d52bf448aff7401b676a7855cb0f4cce00d329"} Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.227823 4821 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"94c7949f077b1ca2b310bedb090aa72fc4bd7e8f2c8bf438096edfca8568f36e"} Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.227829 4821 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ed2e7cad7bb24b5b56a1295098a6a7c230c8f9b9c29a88ef653ae8b6ac31667d"} Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.227834 4821 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3"} Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.227840 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" event={"ID":"954085fa-d03a-4e88-9dfd-566257e3558d","Type":"ContainerDied","Data":"ed2e7cad7bb24b5b56a1295098a6a7c230c8f9b9c29a88ef653ae8b6ac31667d"} Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.227848 4821 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"32de39e5841057630116280f4a9e5e5728b786db001c2c0db0d57e7475a3a16d"} Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.227854 4821 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"90d8792a79993c04a0eb774498395d955deff5c42c693892426685d157cf9cff"} Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.227859 4821 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7dbf499f4ae6161752abaa7a466a027c6f874016be90a534e69f07e3dd3f3d9f"} Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.227864 4821 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e9d7254196b3702e48083ae98e829b73c539adf69ed5b2bbdc62189d09ae1762"} Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.227869 4821 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d727430108977574365f0e9c04b44e03307bd7ff89cdbb1e48a6e916f504d6b0"} Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.227874 4821 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1a14a078ba7d2aed95992b36502008f315f0d3a634b64ed49da70398b2562d44"} Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.227879 4821 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2f00746930c4588e548f0e2b60d52bf448aff7401b676a7855cb0f4cce00d329"} Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.227885 4821 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"94c7949f077b1ca2b310bedb090aa72fc4bd7e8f2c8bf438096edfca8568f36e"} Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.227889 4821 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ed2e7cad7bb24b5b56a1295098a6a7c230c8f9b9c29a88ef653ae8b6ac31667d"} Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.227936 4821 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3"} Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.227944 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mzhnw" event={"ID":"954085fa-d03a-4e88-9dfd-566257e3558d","Type":"ContainerDied","Data":"0ebf341b9565ca43636f838f682a2a1805f094bdccd2c8362fd1dd4236b179ad"} Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.227951 4821 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"32de39e5841057630116280f4a9e5e5728b786db001c2c0db0d57e7475a3a16d"} Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.227957 4821 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"90d8792a79993c04a0eb774498395d955deff5c42c693892426685d157cf9cff"} Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.227962 4821 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7dbf499f4ae6161752abaa7a466a027c6f874016be90a534e69f07e3dd3f3d9f"} Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.227967 4821 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e9d7254196b3702e48083ae98e829b73c539adf69ed5b2bbdc62189d09ae1762"} Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.227972 4821 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d727430108977574365f0e9c04b44e03307bd7ff89cdbb1e48a6e916f504d6b0"} Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.227977 4821 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1a14a078ba7d2aed95992b36502008f315f0d3a634b64ed49da70398b2562d44"} Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.227982 4821 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2f00746930c4588e548f0e2b60d52bf448aff7401b676a7855cb0f4cce00d329"} Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.227987 4821 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"94c7949f077b1ca2b310bedb090aa72fc4bd7e8f2c8bf438096edfca8568f36e"} Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.227992 4821 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ed2e7cad7bb24b5b56a1295098a6a7c230c8f9b9c29a88ef653ae8b6ac31667d"} Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.227996 4821 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3"} Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.228008 4821 scope.go:117] "RemoveContainer" containerID="32de39e5841057630116280f4a9e5e5728b786db001c2c0db0d57e7475a3a16d" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.263396 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-mzhnw"] Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.263507 4821 scope.go:117] "RemoveContainer" containerID="90d8792a79993c04a0eb774498395d955deff5c42c693892426685d157cf9cff" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.268714 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-mzhnw"] Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.305811 4821 scope.go:117] "RemoveContainer" containerID="7dbf499f4ae6161752abaa7a466a027c6f874016be90a534e69f07e3dd3f3d9f" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.323765 4821 scope.go:117] "RemoveContainer" containerID="e9d7254196b3702e48083ae98e829b73c539adf69ed5b2bbdc62189d09ae1762" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.350219 4821 scope.go:117] "RemoveContainer" containerID="d727430108977574365f0e9c04b44e03307bd7ff89cdbb1e48a6e916f504d6b0" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.365915 4821 scope.go:117] "RemoveContainer" containerID="1a14a078ba7d2aed95992b36502008f315f0d3a634b64ed49da70398b2562d44" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.376944 4821 scope.go:117] "RemoveContainer" containerID="2f00746930c4588e548f0e2b60d52bf448aff7401b676a7855cb0f4cce00d329" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.388498 4821 scope.go:117] "RemoveContainer" containerID="94c7949f077b1ca2b310bedb090aa72fc4bd7e8f2c8bf438096edfca8568f36e" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.401073 4821 scope.go:117] "RemoveContainer" containerID="ed2e7cad7bb24b5b56a1295098a6a7c230c8f9b9c29a88ef653ae8b6ac31667d" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.418845 4821 scope.go:117] "RemoveContainer" containerID="932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.442776 4821 scope.go:117] "RemoveContainer" containerID="32de39e5841057630116280f4a9e5e5728b786db001c2c0db0d57e7475a3a16d" Nov 25 10:42:03 crc kubenswrapper[4821]: E1125 10:42:03.443145 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"32de39e5841057630116280f4a9e5e5728b786db001c2c0db0d57e7475a3a16d\": container with ID starting with 32de39e5841057630116280f4a9e5e5728b786db001c2c0db0d57e7475a3a16d not found: ID does not exist" containerID="32de39e5841057630116280f4a9e5e5728b786db001c2c0db0d57e7475a3a16d" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.443191 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32de39e5841057630116280f4a9e5e5728b786db001c2c0db0d57e7475a3a16d"} err="failed to get container status \"32de39e5841057630116280f4a9e5e5728b786db001c2c0db0d57e7475a3a16d\": rpc error: code = NotFound desc = could not find container \"32de39e5841057630116280f4a9e5e5728b786db001c2c0db0d57e7475a3a16d\": container with ID starting with 32de39e5841057630116280f4a9e5e5728b786db001c2c0db0d57e7475a3a16d not found: ID does not exist" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.443220 4821 scope.go:117] "RemoveContainer" containerID="90d8792a79993c04a0eb774498395d955deff5c42c693892426685d157cf9cff" Nov 25 10:42:03 crc kubenswrapper[4821]: E1125 10:42:03.443666 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90d8792a79993c04a0eb774498395d955deff5c42c693892426685d157cf9cff\": container with ID starting with 90d8792a79993c04a0eb774498395d955deff5c42c693892426685d157cf9cff not found: ID does not exist" containerID="90d8792a79993c04a0eb774498395d955deff5c42c693892426685d157cf9cff" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.443686 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90d8792a79993c04a0eb774498395d955deff5c42c693892426685d157cf9cff"} err="failed to get container status \"90d8792a79993c04a0eb774498395d955deff5c42c693892426685d157cf9cff\": rpc error: code = NotFound desc = could not find container \"90d8792a79993c04a0eb774498395d955deff5c42c693892426685d157cf9cff\": container with ID starting with 90d8792a79993c04a0eb774498395d955deff5c42c693892426685d157cf9cff not found: ID does not exist" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.443717 4821 scope.go:117] "RemoveContainer" containerID="7dbf499f4ae6161752abaa7a466a027c6f874016be90a534e69f07e3dd3f3d9f" Nov 25 10:42:03 crc kubenswrapper[4821]: E1125 10:42:03.443977 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7dbf499f4ae6161752abaa7a466a027c6f874016be90a534e69f07e3dd3f3d9f\": container with ID starting with 7dbf499f4ae6161752abaa7a466a027c6f874016be90a534e69f07e3dd3f3d9f not found: ID does not exist" containerID="7dbf499f4ae6161752abaa7a466a027c6f874016be90a534e69f07e3dd3f3d9f" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.444040 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7dbf499f4ae6161752abaa7a466a027c6f874016be90a534e69f07e3dd3f3d9f"} err="failed to get container status \"7dbf499f4ae6161752abaa7a466a027c6f874016be90a534e69f07e3dd3f3d9f\": rpc error: code = NotFound desc = could not find container \"7dbf499f4ae6161752abaa7a466a027c6f874016be90a534e69f07e3dd3f3d9f\": container with ID starting with 7dbf499f4ae6161752abaa7a466a027c6f874016be90a534e69f07e3dd3f3d9f not found: ID does not exist" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.444066 4821 scope.go:117] "RemoveContainer" containerID="e9d7254196b3702e48083ae98e829b73c539adf69ed5b2bbdc62189d09ae1762" Nov 25 10:42:03 crc kubenswrapper[4821]: E1125 10:42:03.444377 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e9d7254196b3702e48083ae98e829b73c539adf69ed5b2bbdc62189d09ae1762\": container with ID starting with e9d7254196b3702e48083ae98e829b73c539adf69ed5b2bbdc62189d09ae1762 not found: ID does not exist" containerID="e9d7254196b3702e48083ae98e829b73c539adf69ed5b2bbdc62189d09ae1762" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.444401 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9d7254196b3702e48083ae98e829b73c539adf69ed5b2bbdc62189d09ae1762"} err="failed to get container status \"e9d7254196b3702e48083ae98e829b73c539adf69ed5b2bbdc62189d09ae1762\": rpc error: code = NotFound desc = could not find container \"e9d7254196b3702e48083ae98e829b73c539adf69ed5b2bbdc62189d09ae1762\": container with ID starting with e9d7254196b3702e48083ae98e829b73c539adf69ed5b2bbdc62189d09ae1762 not found: ID does not exist" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.444420 4821 scope.go:117] "RemoveContainer" containerID="d727430108977574365f0e9c04b44e03307bd7ff89cdbb1e48a6e916f504d6b0" Nov 25 10:42:03 crc kubenswrapper[4821]: E1125 10:42:03.444671 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d727430108977574365f0e9c04b44e03307bd7ff89cdbb1e48a6e916f504d6b0\": container with ID starting with d727430108977574365f0e9c04b44e03307bd7ff89cdbb1e48a6e916f504d6b0 not found: ID does not exist" containerID="d727430108977574365f0e9c04b44e03307bd7ff89cdbb1e48a6e916f504d6b0" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.444690 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d727430108977574365f0e9c04b44e03307bd7ff89cdbb1e48a6e916f504d6b0"} err="failed to get container status \"d727430108977574365f0e9c04b44e03307bd7ff89cdbb1e48a6e916f504d6b0\": rpc error: code = NotFound desc = could not find container \"d727430108977574365f0e9c04b44e03307bd7ff89cdbb1e48a6e916f504d6b0\": container with ID starting with d727430108977574365f0e9c04b44e03307bd7ff89cdbb1e48a6e916f504d6b0 not found: ID does not exist" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.444702 4821 scope.go:117] "RemoveContainer" containerID="1a14a078ba7d2aed95992b36502008f315f0d3a634b64ed49da70398b2562d44" Nov 25 10:42:03 crc kubenswrapper[4821]: E1125 10:42:03.444908 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a14a078ba7d2aed95992b36502008f315f0d3a634b64ed49da70398b2562d44\": container with ID starting with 1a14a078ba7d2aed95992b36502008f315f0d3a634b64ed49da70398b2562d44 not found: ID does not exist" containerID="1a14a078ba7d2aed95992b36502008f315f0d3a634b64ed49da70398b2562d44" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.444931 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a14a078ba7d2aed95992b36502008f315f0d3a634b64ed49da70398b2562d44"} err="failed to get container status \"1a14a078ba7d2aed95992b36502008f315f0d3a634b64ed49da70398b2562d44\": rpc error: code = NotFound desc = could not find container \"1a14a078ba7d2aed95992b36502008f315f0d3a634b64ed49da70398b2562d44\": container with ID starting with 1a14a078ba7d2aed95992b36502008f315f0d3a634b64ed49da70398b2562d44 not found: ID does not exist" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.444950 4821 scope.go:117] "RemoveContainer" containerID="2f00746930c4588e548f0e2b60d52bf448aff7401b676a7855cb0f4cce00d329" Nov 25 10:42:03 crc kubenswrapper[4821]: E1125 10:42:03.445201 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f00746930c4588e548f0e2b60d52bf448aff7401b676a7855cb0f4cce00d329\": container with ID starting with 2f00746930c4588e548f0e2b60d52bf448aff7401b676a7855cb0f4cce00d329 not found: ID does not exist" containerID="2f00746930c4588e548f0e2b60d52bf448aff7401b676a7855cb0f4cce00d329" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.445228 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f00746930c4588e548f0e2b60d52bf448aff7401b676a7855cb0f4cce00d329"} err="failed to get container status \"2f00746930c4588e548f0e2b60d52bf448aff7401b676a7855cb0f4cce00d329\": rpc error: code = NotFound desc = could not find container \"2f00746930c4588e548f0e2b60d52bf448aff7401b676a7855cb0f4cce00d329\": container with ID starting with 2f00746930c4588e548f0e2b60d52bf448aff7401b676a7855cb0f4cce00d329 not found: ID does not exist" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.445245 4821 scope.go:117] "RemoveContainer" containerID="94c7949f077b1ca2b310bedb090aa72fc4bd7e8f2c8bf438096edfca8568f36e" Nov 25 10:42:03 crc kubenswrapper[4821]: E1125 10:42:03.445535 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"94c7949f077b1ca2b310bedb090aa72fc4bd7e8f2c8bf438096edfca8568f36e\": container with ID starting with 94c7949f077b1ca2b310bedb090aa72fc4bd7e8f2c8bf438096edfca8568f36e not found: ID does not exist" containerID="94c7949f077b1ca2b310bedb090aa72fc4bd7e8f2c8bf438096edfca8568f36e" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.445560 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"94c7949f077b1ca2b310bedb090aa72fc4bd7e8f2c8bf438096edfca8568f36e"} err="failed to get container status \"94c7949f077b1ca2b310bedb090aa72fc4bd7e8f2c8bf438096edfca8568f36e\": rpc error: code = NotFound desc = could not find container \"94c7949f077b1ca2b310bedb090aa72fc4bd7e8f2c8bf438096edfca8568f36e\": container with ID starting with 94c7949f077b1ca2b310bedb090aa72fc4bd7e8f2c8bf438096edfca8568f36e not found: ID does not exist" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.445578 4821 scope.go:117] "RemoveContainer" containerID="ed2e7cad7bb24b5b56a1295098a6a7c230c8f9b9c29a88ef653ae8b6ac31667d" Nov 25 10:42:03 crc kubenswrapper[4821]: E1125 10:42:03.445833 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed2e7cad7bb24b5b56a1295098a6a7c230c8f9b9c29a88ef653ae8b6ac31667d\": container with ID starting with ed2e7cad7bb24b5b56a1295098a6a7c230c8f9b9c29a88ef653ae8b6ac31667d not found: ID does not exist" containerID="ed2e7cad7bb24b5b56a1295098a6a7c230c8f9b9c29a88ef653ae8b6ac31667d" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.445856 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed2e7cad7bb24b5b56a1295098a6a7c230c8f9b9c29a88ef653ae8b6ac31667d"} err="failed to get container status \"ed2e7cad7bb24b5b56a1295098a6a7c230c8f9b9c29a88ef653ae8b6ac31667d\": rpc error: code = NotFound desc = could not find container \"ed2e7cad7bb24b5b56a1295098a6a7c230c8f9b9c29a88ef653ae8b6ac31667d\": container with ID starting with ed2e7cad7bb24b5b56a1295098a6a7c230c8f9b9c29a88ef653ae8b6ac31667d not found: ID does not exist" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.445870 4821 scope.go:117] "RemoveContainer" containerID="932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3" Nov 25 10:42:03 crc kubenswrapper[4821]: E1125 10:42:03.446207 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\": container with ID starting with 932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3 not found: ID does not exist" containerID="932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.446226 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3"} err="failed to get container status \"932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\": rpc error: code = NotFound desc = could not find container \"932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\": container with ID starting with 932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3 not found: ID does not exist" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.446238 4821 scope.go:117] "RemoveContainer" containerID="32de39e5841057630116280f4a9e5e5728b786db001c2c0db0d57e7475a3a16d" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.446506 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32de39e5841057630116280f4a9e5e5728b786db001c2c0db0d57e7475a3a16d"} err="failed to get container status \"32de39e5841057630116280f4a9e5e5728b786db001c2c0db0d57e7475a3a16d\": rpc error: code = NotFound desc = could not find container \"32de39e5841057630116280f4a9e5e5728b786db001c2c0db0d57e7475a3a16d\": container with ID starting with 32de39e5841057630116280f4a9e5e5728b786db001c2c0db0d57e7475a3a16d not found: ID does not exist" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.446522 4821 scope.go:117] "RemoveContainer" containerID="90d8792a79993c04a0eb774498395d955deff5c42c693892426685d157cf9cff" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.446748 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90d8792a79993c04a0eb774498395d955deff5c42c693892426685d157cf9cff"} err="failed to get container status \"90d8792a79993c04a0eb774498395d955deff5c42c693892426685d157cf9cff\": rpc error: code = NotFound desc = could not find container \"90d8792a79993c04a0eb774498395d955deff5c42c693892426685d157cf9cff\": container with ID starting with 90d8792a79993c04a0eb774498395d955deff5c42c693892426685d157cf9cff not found: ID does not exist" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.446769 4821 scope.go:117] "RemoveContainer" containerID="7dbf499f4ae6161752abaa7a466a027c6f874016be90a534e69f07e3dd3f3d9f" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.447042 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7dbf499f4ae6161752abaa7a466a027c6f874016be90a534e69f07e3dd3f3d9f"} err="failed to get container status \"7dbf499f4ae6161752abaa7a466a027c6f874016be90a534e69f07e3dd3f3d9f\": rpc error: code = NotFound desc = could not find container \"7dbf499f4ae6161752abaa7a466a027c6f874016be90a534e69f07e3dd3f3d9f\": container with ID starting with 7dbf499f4ae6161752abaa7a466a027c6f874016be90a534e69f07e3dd3f3d9f not found: ID does not exist" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.447064 4821 scope.go:117] "RemoveContainer" containerID="e9d7254196b3702e48083ae98e829b73c539adf69ed5b2bbdc62189d09ae1762" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.447298 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9d7254196b3702e48083ae98e829b73c539adf69ed5b2bbdc62189d09ae1762"} err="failed to get container status \"e9d7254196b3702e48083ae98e829b73c539adf69ed5b2bbdc62189d09ae1762\": rpc error: code = NotFound desc = could not find container \"e9d7254196b3702e48083ae98e829b73c539adf69ed5b2bbdc62189d09ae1762\": container with ID starting with e9d7254196b3702e48083ae98e829b73c539adf69ed5b2bbdc62189d09ae1762 not found: ID does not exist" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.447319 4821 scope.go:117] "RemoveContainer" containerID="d727430108977574365f0e9c04b44e03307bd7ff89cdbb1e48a6e916f504d6b0" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.447511 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d727430108977574365f0e9c04b44e03307bd7ff89cdbb1e48a6e916f504d6b0"} err="failed to get container status \"d727430108977574365f0e9c04b44e03307bd7ff89cdbb1e48a6e916f504d6b0\": rpc error: code = NotFound desc = could not find container \"d727430108977574365f0e9c04b44e03307bd7ff89cdbb1e48a6e916f504d6b0\": container with ID starting with d727430108977574365f0e9c04b44e03307bd7ff89cdbb1e48a6e916f504d6b0 not found: ID does not exist" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.447531 4821 scope.go:117] "RemoveContainer" containerID="1a14a078ba7d2aed95992b36502008f315f0d3a634b64ed49da70398b2562d44" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.447722 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a14a078ba7d2aed95992b36502008f315f0d3a634b64ed49da70398b2562d44"} err="failed to get container status \"1a14a078ba7d2aed95992b36502008f315f0d3a634b64ed49da70398b2562d44\": rpc error: code = NotFound desc = could not find container \"1a14a078ba7d2aed95992b36502008f315f0d3a634b64ed49da70398b2562d44\": container with ID starting with 1a14a078ba7d2aed95992b36502008f315f0d3a634b64ed49da70398b2562d44 not found: ID does not exist" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.447742 4821 scope.go:117] "RemoveContainer" containerID="2f00746930c4588e548f0e2b60d52bf448aff7401b676a7855cb0f4cce00d329" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.447947 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f00746930c4588e548f0e2b60d52bf448aff7401b676a7855cb0f4cce00d329"} err="failed to get container status \"2f00746930c4588e548f0e2b60d52bf448aff7401b676a7855cb0f4cce00d329\": rpc error: code = NotFound desc = could not find container \"2f00746930c4588e548f0e2b60d52bf448aff7401b676a7855cb0f4cce00d329\": container with ID starting with 2f00746930c4588e548f0e2b60d52bf448aff7401b676a7855cb0f4cce00d329 not found: ID does not exist" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.447969 4821 scope.go:117] "RemoveContainer" containerID="94c7949f077b1ca2b310bedb090aa72fc4bd7e8f2c8bf438096edfca8568f36e" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.448283 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"94c7949f077b1ca2b310bedb090aa72fc4bd7e8f2c8bf438096edfca8568f36e"} err="failed to get container status \"94c7949f077b1ca2b310bedb090aa72fc4bd7e8f2c8bf438096edfca8568f36e\": rpc error: code = NotFound desc = could not find container \"94c7949f077b1ca2b310bedb090aa72fc4bd7e8f2c8bf438096edfca8568f36e\": container with ID starting with 94c7949f077b1ca2b310bedb090aa72fc4bd7e8f2c8bf438096edfca8568f36e not found: ID does not exist" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.448306 4821 scope.go:117] "RemoveContainer" containerID="ed2e7cad7bb24b5b56a1295098a6a7c230c8f9b9c29a88ef653ae8b6ac31667d" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.448591 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed2e7cad7bb24b5b56a1295098a6a7c230c8f9b9c29a88ef653ae8b6ac31667d"} err="failed to get container status \"ed2e7cad7bb24b5b56a1295098a6a7c230c8f9b9c29a88ef653ae8b6ac31667d\": rpc error: code = NotFound desc = could not find container \"ed2e7cad7bb24b5b56a1295098a6a7c230c8f9b9c29a88ef653ae8b6ac31667d\": container with ID starting with ed2e7cad7bb24b5b56a1295098a6a7c230c8f9b9c29a88ef653ae8b6ac31667d not found: ID does not exist" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.448616 4821 scope.go:117] "RemoveContainer" containerID="932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.448894 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3"} err="failed to get container status \"932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\": rpc error: code = NotFound desc = could not find container \"932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\": container with ID starting with 932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3 not found: ID does not exist" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.448914 4821 scope.go:117] "RemoveContainer" containerID="32de39e5841057630116280f4a9e5e5728b786db001c2c0db0d57e7475a3a16d" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.449105 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32de39e5841057630116280f4a9e5e5728b786db001c2c0db0d57e7475a3a16d"} err="failed to get container status \"32de39e5841057630116280f4a9e5e5728b786db001c2c0db0d57e7475a3a16d\": rpc error: code = NotFound desc = could not find container \"32de39e5841057630116280f4a9e5e5728b786db001c2c0db0d57e7475a3a16d\": container with ID starting with 32de39e5841057630116280f4a9e5e5728b786db001c2c0db0d57e7475a3a16d not found: ID does not exist" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.449125 4821 scope.go:117] "RemoveContainer" containerID="90d8792a79993c04a0eb774498395d955deff5c42c693892426685d157cf9cff" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.449422 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90d8792a79993c04a0eb774498395d955deff5c42c693892426685d157cf9cff"} err="failed to get container status \"90d8792a79993c04a0eb774498395d955deff5c42c693892426685d157cf9cff\": rpc error: code = NotFound desc = could not find container \"90d8792a79993c04a0eb774498395d955deff5c42c693892426685d157cf9cff\": container with ID starting with 90d8792a79993c04a0eb774498395d955deff5c42c693892426685d157cf9cff not found: ID does not exist" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.449445 4821 scope.go:117] "RemoveContainer" containerID="7dbf499f4ae6161752abaa7a466a027c6f874016be90a534e69f07e3dd3f3d9f" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.449663 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7dbf499f4ae6161752abaa7a466a027c6f874016be90a534e69f07e3dd3f3d9f"} err="failed to get container status \"7dbf499f4ae6161752abaa7a466a027c6f874016be90a534e69f07e3dd3f3d9f\": rpc error: code = NotFound desc = could not find container \"7dbf499f4ae6161752abaa7a466a027c6f874016be90a534e69f07e3dd3f3d9f\": container with ID starting with 7dbf499f4ae6161752abaa7a466a027c6f874016be90a534e69f07e3dd3f3d9f not found: ID does not exist" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.449678 4821 scope.go:117] "RemoveContainer" containerID="e9d7254196b3702e48083ae98e829b73c539adf69ed5b2bbdc62189d09ae1762" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.449986 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9d7254196b3702e48083ae98e829b73c539adf69ed5b2bbdc62189d09ae1762"} err="failed to get container status \"e9d7254196b3702e48083ae98e829b73c539adf69ed5b2bbdc62189d09ae1762\": rpc error: code = NotFound desc = could not find container \"e9d7254196b3702e48083ae98e829b73c539adf69ed5b2bbdc62189d09ae1762\": container with ID starting with e9d7254196b3702e48083ae98e829b73c539adf69ed5b2bbdc62189d09ae1762 not found: ID does not exist" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.450016 4821 scope.go:117] "RemoveContainer" containerID="d727430108977574365f0e9c04b44e03307bd7ff89cdbb1e48a6e916f504d6b0" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.450301 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d727430108977574365f0e9c04b44e03307bd7ff89cdbb1e48a6e916f504d6b0"} err="failed to get container status \"d727430108977574365f0e9c04b44e03307bd7ff89cdbb1e48a6e916f504d6b0\": rpc error: code = NotFound desc = could not find container \"d727430108977574365f0e9c04b44e03307bd7ff89cdbb1e48a6e916f504d6b0\": container with ID starting with d727430108977574365f0e9c04b44e03307bd7ff89cdbb1e48a6e916f504d6b0 not found: ID does not exist" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.450341 4821 scope.go:117] "RemoveContainer" containerID="1a14a078ba7d2aed95992b36502008f315f0d3a634b64ed49da70398b2562d44" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.450582 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a14a078ba7d2aed95992b36502008f315f0d3a634b64ed49da70398b2562d44"} err="failed to get container status \"1a14a078ba7d2aed95992b36502008f315f0d3a634b64ed49da70398b2562d44\": rpc error: code = NotFound desc = could not find container \"1a14a078ba7d2aed95992b36502008f315f0d3a634b64ed49da70398b2562d44\": container with ID starting with 1a14a078ba7d2aed95992b36502008f315f0d3a634b64ed49da70398b2562d44 not found: ID does not exist" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.450605 4821 scope.go:117] "RemoveContainer" containerID="2f00746930c4588e548f0e2b60d52bf448aff7401b676a7855cb0f4cce00d329" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.450829 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f00746930c4588e548f0e2b60d52bf448aff7401b676a7855cb0f4cce00d329"} err="failed to get container status \"2f00746930c4588e548f0e2b60d52bf448aff7401b676a7855cb0f4cce00d329\": rpc error: code = NotFound desc = could not find container \"2f00746930c4588e548f0e2b60d52bf448aff7401b676a7855cb0f4cce00d329\": container with ID starting with 2f00746930c4588e548f0e2b60d52bf448aff7401b676a7855cb0f4cce00d329 not found: ID does not exist" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.450848 4821 scope.go:117] "RemoveContainer" containerID="94c7949f077b1ca2b310bedb090aa72fc4bd7e8f2c8bf438096edfca8568f36e" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.451146 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"94c7949f077b1ca2b310bedb090aa72fc4bd7e8f2c8bf438096edfca8568f36e"} err="failed to get container status \"94c7949f077b1ca2b310bedb090aa72fc4bd7e8f2c8bf438096edfca8568f36e\": rpc error: code = NotFound desc = could not find container \"94c7949f077b1ca2b310bedb090aa72fc4bd7e8f2c8bf438096edfca8568f36e\": container with ID starting with 94c7949f077b1ca2b310bedb090aa72fc4bd7e8f2c8bf438096edfca8568f36e not found: ID does not exist" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.451181 4821 scope.go:117] "RemoveContainer" containerID="ed2e7cad7bb24b5b56a1295098a6a7c230c8f9b9c29a88ef653ae8b6ac31667d" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.451400 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed2e7cad7bb24b5b56a1295098a6a7c230c8f9b9c29a88ef653ae8b6ac31667d"} err="failed to get container status \"ed2e7cad7bb24b5b56a1295098a6a7c230c8f9b9c29a88ef653ae8b6ac31667d\": rpc error: code = NotFound desc = could not find container \"ed2e7cad7bb24b5b56a1295098a6a7c230c8f9b9c29a88ef653ae8b6ac31667d\": container with ID starting with ed2e7cad7bb24b5b56a1295098a6a7c230c8f9b9c29a88ef653ae8b6ac31667d not found: ID does not exist" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.451423 4821 scope.go:117] "RemoveContainer" containerID="932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.451727 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3"} err="failed to get container status \"932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\": rpc error: code = NotFound desc = could not find container \"932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\": container with ID starting with 932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3 not found: ID does not exist" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.451745 4821 scope.go:117] "RemoveContainer" containerID="32de39e5841057630116280f4a9e5e5728b786db001c2c0db0d57e7475a3a16d" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.452062 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32de39e5841057630116280f4a9e5e5728b786db001c2c0db0d57e7475a3a16d"} err="failed to get container status \"32de39e5841057630116280f4a9e5e5728b786db001c2c0db0d57e7475a3a16d\": rpc error: code = NotFound desc = could not find container \"32de39e5841057630116280f4a9e5e5728b786db001c2c0db0d57e7475a3a16d\": container with ID starting with 32de39e5841057630116280f4a9e5e5728b786db001c2c0db0d57e7475a3a16d not found: ID does not exist" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.452086 4821 scope.go:117] "RemoveContainer" containerID="90d8792a79993c04a0eb774498395d955deff5c42c693892426685d157cf9cff" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.452372 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90d8792a79993c04a0eb774498395d955deff5c42c693892426685d157cf9cff"} err="failed to get container status \"90d8792a79993c04a0eb774498395d955deff5c42c693892426685d157cf9cff\": rpc error: code = NotFound desc = could not find container \"90d8792a79993c04a0eb774498395d955deff5c42c693892426685d157cf9cff\": container with ID starting with 90d8792a79993c04a0eb774498395d955deff5c42c693892426685d157cf9cff not found: ID does not exist" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.452390 4821 scope.go:117] "RemoveContainer" containerID="7dbf499f4ae6161752abaa7a466a027c6f874016be90a534e69f07e3dd3f3d9f" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.452730 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7dbf499f4ae6161752abaa7a466a027c6f874016be90a534e69f07e3dd3f3d9f"} err="failed to get container status \"7dbf499f4ae6161752abaa7a466a027c6f874016be90a534e69f07e3dd3f3d9f\": rpc error: code = NotFound desc = could not find container \"7dbf499f4ae6161752abaa7a466a027c6f874016be90a534e69f07e3dd3f3d9f\": container with ID starting with 7dbf499f4ae6161752abaa7a466a027c6f874016be90a534e69f07e3dd3f3d9f not found: ID does not exist" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.452751 4821 scope.go:117] "RemoveContainer" containerID="e9d7254196b3702e48083ae98e829b73c539adf69ed5b2bbdc62189d09ae1762" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.452984 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9d7254196b3702e48083ae98e829b73c539adf69ed5b2bbdc62189d09ae1762"} err="failed to get container status \"e9d7254196b3702e48083ae98e829b73c539adf69ed5b2bbdc62189d09ae1762\": rpc error: code = NotFound desc = could not find container \"e9d7254196b3702e48083ae98e829b73c539adf69ed5b2bbdc62189d09ae1762\": container with ID starting with e9d7254196b3702e48083ae98e829b73c539adf69ed5b2bbdc62189d09ae1762 not found: ID does not exist" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.453008 4821 scope.go:117] "RemoveContainer" containerID="d727430108977574365f0e9c04b44e03307bd7ff89cdbb1e48a6e916f504d6b0" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.453369 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d727430108977574365f0e9c04b44e03307bd7ff89cdbb1e48a6e916f504d6b0"} err="failed to get container status \"d727430108977574365f0e9c04b44e03307bd7ff89cdbb1e48a6e916f504d6b0\": rpc error: code = NotFound desc = could not find container \"d727430108977574365f0e9c04b44e03307bd7ff89cdbb1e48a6e916f504d6b0\": container with ID starting with d727430108977574365f0e9c04b44e03307bd7ff89cdbb1e48a6e916f504d6b0 not found: ID does not exist" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.453392 4821 scope.go:117] "RemoveContainer" containerID="1a14a078ba7d2aed95992b36502008f315f0d3a634b64ed49da70398b2562d44" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.453673 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a14a078ba7d2aed95992b36502008f315f0d3a634b64ed49da70398b2562d44"} err="failed to get container status \"1a14a078ba7d2aed95992b36502008f315f0d3a634b64ed49da70398b2562d44\": rpc error: code = NotFound desc = could not find container \"1a14a078ba7d2aed95992b36502008f315f0d3a634b64ed49da70398b2562d44\": container with ID starting with 1a14a078ba7d2aed95992b36502008f315f0d3a634b64ed49da70398b2562d44 not found: ID does not exist" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.453698 4821 scope.go:117] "RemoveContainer" containerID="2f00746930c4588e548f0e2b60d52bf448aff7401b676a7855cb0f4cce00d329" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.453971 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f00746930c4588e548f0e2b60d52bf448aff7401b676a7855cb0f4cce00d329"} err="failed to get container status \"2f00746930c4588e548f0e2b60d52bf448aff7401b676a7855cb0f4cce00d329\": rpc error: code = NotFound desc = could not find container \"2f00746930c4588e548f0e2b60d52bf448aff7401b676a7855cb0f4cce00d329\": container with ID starting with 2f00746930c4588e548f0e2b60d52bf448aff7401b676a7855cb0f4cce00d329 not found: ID does not exist" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.453994 4821 scope.go:117] "RemoveContainer" containerID="94c7949f077b1ca2b310bedb090aa72fc4bd7e8f2c8bf438096edfca8568f36e" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.454230 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"94c7949f077b1ca2b310bedb090aa72fc4bd7e8f2c8bf438096edfca8568f36e"} err="failed to get container status \"94c7949f077b1ca2b310bedb090aa72fc4bd7e8f2c8bf438096edfca8568f36e\": rpc error: code = NotFound desc = could not find container \"94c7949f077b1ca2b310bedb090aa72fc4bd7e8f2c8bf438096edfca8568f36e\": container with ID starting with 94c7949f077b1ca2b310bedb090aa72fc4bd7e8f2c8bf438096edfca8568f36e not found: ID does not exist" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.454271 4821 scope.go:117] "RemoveContainer" containerID="ed2e7cad7bb24b5b56a1295098a6a7c230c8f9b9c29a88ef653ae8b6ac31667d" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.454584 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed2e7cad7bb24b5b56a1295098a6a7c230c8f9b9c29a88ef653ae8b6ac31667d"} err="failed to get container status \"ed2e7cad7bb24b5b56a1295098a6a7c230c8f9b9c29a88ef653ae8b6ac31667d\": rpc error: code = NotFound desc = could not find container \"ed2e7cad7bb24b5b56a1295098a6a7c230c8f9b9c29a88ef653ae8b6ac31667d\": container with ID starting with ed2e7cad7bb24b5b56a1295098a6a7c230c8f9b9c29a88ef653ae8b6ac31667d not found: ID does not exist" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.454601 4821 scope.go:117] "RemoveContainer" containerID="932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.454830 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3"} err="failed to get container status \"932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\": rpc error: code = NotFound desc = could not find container \"932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3\": container with ID starting with 932a6494e13e8344afb1262c70880bd7a89e7df316bc9ff2d1512b8bcfef73d3 not found: ID does not exist" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.454851 4821 scope.go:117] "RemoveContainer" containerID="32de39e5841057630116280f4a9e5e5728b786db001c2c0db0d57e7475a3a16d" Nov 25 10:42:03 crc kubenswrapper[4821]: I1125 10:42:03.455085 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32de39e5841057630116280f4a9e5e5728b786db001c2c0db0d57e7475a3a16d"} err="failed to get container status \"32de39e5841057630116280f4a9e5e5728b786db001c2c0db0d57e7475a3a16d\": rpc error: code = NotFound desc = could not find container \"32de39e5841057630116280f4a9e5e5728b786db001c2c0db0d57e7475a3a16d\": container with ID starting with 32de39e5841057630116280f4a9e5e5728b786db001c2c0db0d57e7475a3a16d not found: ID does not exist" Nov 25 10:42:04 crc kubenswrapper[4821]: I1125 10:42:04.122268 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="954085fa-d03a-4e88-9dfd-566257e3558d" path="/var/lib/kubelet/pods/954085fa-d03a-4e88-9dfd-566257e3558d/volumes" Nov 25 10:42:04 crc kubenswrapper[4821]: I1125 10:42:04.238131 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" event={"ID":"d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7","Type":"ContainerStarted","Data":"b1f2371091ce8977464c89a56ffe7b583f25506f01aa75e7d8737fe08a0c7cd5"} Nov 25 10:42:04 crc kubenswrapper[4821]: I1125 10:42:04.238501 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" event={"ID":"d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7","Type":"ContainerStarted","Data":"7c3cd10a400dc22cf433839ee2629c1b865e77a74737ee15894a653880833703"} Nov 25 10:42:04 crc kubenswrapper[4821]: I1125 10:42:04.238527 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" event={"ID":"d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7","Type":"ContainerStarted","Data":"cedf21cb2a1f6241a33399c24249726e728b3d688fdd415e9d4529cd8452747c"} Nov 25 10:42:04 crc kubenswrapper[4821]: I1125 10:42:04.238554 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" event={"ID":"d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7","Type":"ContainerStarted","Data":"19b40f5b86a7c345f370012f59780d5d5692880662735fd4c12a4aa36f5d2c64"} Nov 25 10:42:04 crc kubenswrapper[4821]: I1125 10:42:04.238576 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" event={"ID":"d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7","Type":"ContainerStarted","Data":"bd665a06e50ae1ae80b29fc5208f8adbddf4337cc5f8751bd5abff3e0266de07"} Nov 25 10:42:04 crc kubenswrapper[4821]: I1125 10:42:04.238597 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" event={"ID":"d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7","Type":"ContainerStarted","Data":"5610b669216dc0369761d2384d5c32f24e5081dca4a8216e03d7fd4ebc5da07f"} Nov 25 10:42:04 crc kubenswrapper[4821]: I1125 10:42:04.715179 4821 patch_prober.go:28] interesting pod/machine-config-daemon-2krbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:42:04 crc kubenswrapper[4821]: I1125 10:42:04.715265 4821 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:42:06 crc kubenswrapper[4821]: I1125 10:42:06.260126 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" event={"ID":"d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7","Type":"ContainerStarted","Data":"de3d6644037aa82893e09e5ad54458bea7b9dab26d75adbcf074506670503ee9"} Nov 25 10:42:09 crc kubenswrapper[4821]: I1125 10:42:09.279284 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" event={"ID":"d0aa1fce-c7d0-477f-b7d3-76b2b7a1d7a7","Type":"ContainerStarted","Data":"8459475eb3af21953a0f44ead697a717131fa018d04b795b651351012b4083fb"} Nov 25 10:42:09 crc kubenswrapper[4821]: I1125 10:42:09.279798 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" Nov 25 10:42:09 crc kubenswrapper[4821]: I1125 10:42:09.311669 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" podStartSLOduration=7.311647883 podStartE2EDuration="7.311647883s" podCreationTimestamp="2025-11-25 10:42:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:42:09.310516452 +0000 UTC m=+599.846836299" watchObservedRunningTime="2025-11-25 10:42:09.311647883 +0000 UTC m=+599.847967740" Nov 25 10:42:09 crc kubenswrapper[4821]: I1125 10:42:09.315422 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" Nov 25 10:42:10 crc kubenswrapper[4821]: I1125 10:42:10.284598 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" Nov 25 10:42:10 crc kubenswrapper[4821]: I1125 10:42:10.284895 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" Nov 25 10:42:10 crc kubenswrapper[4821]: I1125 10:42:10.315554 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" Nov 25 10:42:14 crc kubenswrapper[4821]: I1125 10:42:14.113895 4821 scope.go:117] "RemoveContainer" containerID="86b9194a6cea3434f70eb5f6d91aa470cb4321b386c78735cf4ca3da718f1026" Nov 25 10:42:14 crc kubenswrapper[4821]: E1125 10:42:14.114696 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-ldzbz_openshift-multus(b84e32ae-4897-4644-a656-6bd82012cc81)\"" pod="openshift-multus/multus-ldzbz" podUID="b84e32ae-4897-4644-a656-6bd82012cc81" Nov 25 10:42:27 crc kubenswrapper[4821]: I1125 10:42:27.115152 4821 scope.go:117] "RemoveContainer" containerID="86b9194a6cea3434f70eb5f6d91aa470cb4321b386c78735cf4ca3da718f1026" Nov 25 10:42:27 crc kubenswrapper[4821]: I1125 10:42:27.369594 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-ldzbz_b84e32ae-4897-4644-a656-6bd82012cc81/kube-multus/2.log" Nov 25 10:42:27 crc kubenswrapper[4821]: I1125 10:42:27.370231 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-ldzbz" event={"ID":"b84e32ae-4897-4644-a656-6bd82012cc81","Type":"ContainerStarted","Data":"f79146aa52c1727f4c592c0ad36c4d04c0b9a88d3488dcdd6dc3aee00e2497b1"} Nov 25 10:42:32 crc kubenswrapper[4821]: I1125 10:42:32.829630 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-fvjc2" Nov 25 10:42:34 crc kubenswrapper[4821]: I1125 10:42:34.715240 4821 patch_prober.go:28] interesting pod/machine-config-daemon-2krbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:42:34 crc kubenswrapper[4821]: I1125 10:42:34.715726 4821 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:42:34 crc kubenswrapper[4821]: I1125 10:42:34.715777 4821 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" Nov 25 10:42:34 crc kubenswrapper[4821]: I1125 10:42:34.716493 4821 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"808518ddff79187bc281fef8f1f64e070e5abe86921d58acacca289efda8d1d4"} pod="openshift-machine-config-operator/machine-config-daemon-2krbf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 10:42:34 crc kubenswrapper[4821]: I1125 10:42:34.716554 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerName="machine-config-daemon" containerID="cri-o://808518ddff79187bc281fef8f1f64e070e5abe86921d58acacca289efda8d1d4" gracePeriod=600 Nov 25 10:42:35 crc kubenswrapper[4821]: I1125 10:42:35.412856 4821 generic.go:334] "Generic (PLEG): container finished" podID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerID="808518ddff79187bc281fef8f1f64e070e5abe86921d58acacca289efda8d1d4" exitCode=0 Nov 25 10:42:35 crc kubenswrapper[4821]: I1125 10:42:35.412947 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" event={"ID":"5f948b87-ac86-4de6-ad64-c2ef947f84d4","Type":"ContainerDied","Data":"808518ddff79187bc281fef8f1f64e070e5abe86921d58acacca289efda8d1d4"} Nov 25 10:42:35 crc kubenswrapper[4821]: I1125 10:42:35.413178 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" event={"ID":"5f948b87-ac86-4de6-ad64-c2ef947f84d4","Type":"ContainerStarted","Data":"d231cd7c0910d85e280090ca76750286ce837c9e610ff6ebf83f2f5b9e078c96"} Nov 25 10:42:35 crc kubenswrapper[4821]: I1125 10:42:35.413214 4821 scope.go:117] "RemoveContainer" containerID="131eaad8c97dd44aa2f047258a3256b958757d6189e9d69d307bd69c27cbcd08" Nov 25 10:42:40 crc kubenswrapper[4821]: I1125 10:42:40.067458 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6jtmt"] Nov 25 10:42:40 crc kubenswrapper[4821]: I1125 10:42:40.068933 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6jtmt" Nov 25 10:42:40 crc kubenswrapper[4821]: I1125 10:42:40.070550 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 25 10:42:40 crc kubenswrapper[4821]: I1125 10:42:40.074692 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6jtmt"] Nov 25 10:42:40 crc kubenswrapper[4821]: I1125 10:42:40.245601 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t228v\" (UniqueName: \"kubernetes.io/projected/faf2c05e-02b4-4b92-b614-d3de8e443677-kube-api-access-t228v\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6jtmt\" (UID: \"faf2c05e-02b4-4b92-b614-d3de8e443677\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6jtmt" Nov 25 10:42:40 crc kubenswrapper[4821]: I1125 10:42:40.245778 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/faf2c05e-02b4-4b92-b614-d3de8e443677-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6jtmt\" (UID: \"faf2c05e-02b4-4b92-b614-d3de8e443677\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6jtmt" Nov 25 10:42:40 crc kubenswrapper[4821]: I1125 10:42:40.246907 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/faf2c05e-02b4-4b92-b614-d3de8e443677-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6jtmt\" (UID: \"faf2c05e-02b4-4b92-b614-d3de8e443677\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6jtmt" Nov 25 10:42:40 crc kubenswrapper[4821]: I1125 10:42:40.348416 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t228v\" (UniqueName: \"kubernetes.io/projected/faf2c05e-02b4-4b92-b614-d3de8e443677-kube-api-access-t228v\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6jtmt\" (UID: \"faf2c05e-02b4-4b92-b614-d3de8e443677\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6jtmt" Nov 25 10:42:40 crc kubenswrapper[4821]: I1125 10:42:40.349108 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/faf2c05e-02b4-4b92-b614-d3de8e443677-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6jtmt\" (UID: \"faf2c05e-02b4-4b92-b614-d3de8e443677\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6jtmt" Nov 25 10:42:40 crc kubenswrapper[4821]: I1125 10:42:40.350086 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/faf2c05e-02b4-4b92-b614-d3de8e443677-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6jtmt\" (UID: \"faf2c05e-02b4-4b92-b614-d3de8e443677\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6jtmt" Nov 25 10:42:40 crc kubenswrapper[4821]: I1125 10:42:40.350366 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/faf2c05e-02b4-4b92-b614-d3de8e443677-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6jtmt\" (UID: \"faf2c05e-02b4-4b92-b614-d3de8e443677\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6jtmt" Nov 25 10:42:40 crc kubenswrapper[4821]: I1125 10:42:40.350809 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/faf2c05e-02b4-4b92-b614-d3de8e443677-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6jtmt\" (UID: \"faf2c05e-02b4-4b92-b614-d3de8e443677\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6jtmt" Nov 25 10:42:40 crc kubenswrapper[4821]: I1125 10:42:40.371928 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t228v\" (UniqueName: \"kubernetes.io/projected/faf2c05e-02b4-4b92-b614-d3de8e443677-kube-api-access-t228v\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6jtmt\" (UID: \"faf2c05e-02b4-4b92-b614-d3de8e443677\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6jtmt" Nov 25 10:42:40 crc kubenswrapper[4821]: I1125 10:42:40.411665 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6jtmt" Nov 25 10:42:40 crc kubenswrapper[4821]: I1125 10:42:40.600353 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6jtmt"] Nov 25 10:42:41 crc kubenswrapper[4821]: I1125 10:42:41.444595 4821 generic.go:334] "Generic (PLEG): container finished" podID="faf2c05e-02b4-4b92-b614-d3de8e443677" containerID="6b3857c384a56fa4bb4c0519ff15954ad461d5e82beed09626ccb66830f13383" exitCode=0 Nov 25 10:42:41 crc kubenswrapper[4821]: I1125 10:42:41.444642 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6jtmt" event={"ID":"faf2c05e-02b4-4b92-b614-d3de8e443677","Type":"ContainerDied","Data":"6b3857c384a56fa4bb4c0519ff15954ad461d5e82beed09626ccb66830f13383"} Nov 25 10:42:41 crc kubenswrapper[4821]: I1125 10:42:41.444881 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6jtmt" event={"ID":"faf2c05e-02b4-4b92-b614-d3de8e443677","Type":"ContainerStarted","Data":"6308187d731899a7b50cf417f682f4c2b1169b1ce7c5e792dbbcf47697748b74"} Nov 25 10:42:43 crc kubenswrapper[4821]: I1125 10:42:43.455559 4821 generic.go:334] "Generic (PLEG): container finished" podID="faf2c05e-02b4-4b92-b614-d3de8e443677" containerID="6b7ecdcdb8aa513dd029c6afb3a3d7667a41a1f41bb4ada87bc2f0e3aaf8cf6c" exitCode=0 Nov 25 10:42:43 crc kubenswrapper[4821]: I1125 10:42:43.455874 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6jtmt" event={"ID":"faf2c05e-02b4-4b92-b614-d3de8e443677","Type":"ContainerDied","Data":"6b7ecdcdb8aa513dd029c6afb3a3d7667a41a1f41bb4ada87bc2f0e3aaf8cf6c"} Nov 25 10:42:44 crc kubenswrapper[4821]: I1125 10:42:44.462063 4821 generic.go:334] "Generic (PLEG): container finished" podID="faf2c05e-02b4-4b92-b614-d3de8e443677" containerID="6cba6ce802371cc1367aff9d4881a444a7f6fbf7b308ebc56f79f18802bdfb94" exitCode=0 Nov 25 10:42:44 crc kubenswrapper[4821]: I1125 10:42:44.462183 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6jtmt" event={"ID":"faf2c05e-02b4-4b92-b614-d3de8e443677","Type":"ContainerDied","Data":"6cba6ce802371cc1367aff9d4881a444a7f6fbf7b308ebc56f79f18802bdfb94"} Nov 25 10:42:45 crc kubenswrapper[4821]: I1125 10:42:45.673591 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6jtmt" Nov 25 10:42:45 crc kubenswrapper[4821]: I1125 10:42:45.814319 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/faf2c05e-02b4-4b92-b614-d3de8e443677-bundle\") pod \"faf2c05e-02b4-4b92-b614-d3de8e443677\" (UID: \"faf2c05e-02b4-4b92-b614-d3de8e443677\") " Nov 25 10:42:45 crc kubenswrapper[4821]: I1125 10:42:45.814568 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/faf2c05e-02b4-4b92-b614-d3de8e443677-util\") pod \"faf2c05e-02b4-4b92-b614-d3de8e443677\" (UID: \"faf2c05e-02b4-4b92-b614-d3de8e443677\") " Nov 25 10:42:45 crc kubenswrapper[4821]: I1125 10:42:45.814652 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t228v\" (UniqueName: \"kubernetes.io/projected/faf2c05e-02b4-4b92-b614-d3de8e443677-kube-api-access-t228v\") pod \"faf2c05e-02b4-4b92-b614-d3de8e443677\" (UID: \"faf2c05e-02b4-4b92-b614-d3de8e443677\") " Nov 25 10:42:45 crc kubenswrapper[4821]: I1125 10:42:45.815906 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/faf2c05e-02b4-4b92-b614-d3de8e443677-bundle" (OuterVolumeSpecName: "bundle") pod "faf2c05e-02b4-4b92-b614-d3de8e443677" (UID: "faf2c05e-02b4-4b92-b614-d3de8e443677"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:42:45 crc kubenswrapper[4821]: I1125 10:42:45.819620 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/faf2c05e-02b4-4b92-b614-d3de8e443677-kube-api-access-t228v" (OuterVolumeSpecName: "kube-api-access-t228v") pod "faf2c05e-02b4-4b92-b614-d3de8e443677" (UID: "faf2c05e-02b4-4b92-b614-d3de8e443677"). InnerVolumeSpecName "kube-api-access-t228v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:42:45 crc kubenswrapper[4821]: I1125 10:42:45.829272 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/faf2c05e-02b4-4b92-b614-d3de8e443677-util" (OuterVolumeSpecName: "util") pod "faf2c05e-02b4-4b92-b614-d3de8e443677" (UID: "faf2c05e-02b4-4b92-b614-d3de8e443677"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:42:45 crc kubenswrapper[4821]: I1125 10:42:45.917441 4821 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/faf2c05e-02b4-4b92-b614-d3de8e443677-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:42:45 crc kubenswrapper[4821]: I1125 10:42:45.917517 4821 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/faf2c05e-02b4-4b92-b614-d3de8e443677-util\") on node \"crc\" DevicePath \"\"" Nov 25 10:42:45 crc kubenswrapper[4821]: I1125 10:42:45.917531 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t228v\" (UniqueName: \"kubernetes.io/projected/faf2c05e-02b4-4b92-b614-d3de8e443677-kube-api-access-t228v\") on node \"crc\" DevicePath \"\"" Nov 25 10:42:46 crc kubenswrapper[4821]: I1125 10:42:46.474493 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6jtmt" event={"ID":"faf2c05e-02b4-4b92-b614-d3de8e443677","Type":"ContainerDied","Data":"6308187d731899a7b50cf417f682f4c2b1169b1ce7c5e792dbbcf47697748b74"} Nov 25 10:42:46 crc kubenswrapper[4821]: I1125 10:42:46.474534 4821 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6308187d731899a7b50cf417f682f4c2b1169b1ce7c5e792dbbcf47697748b74" Nov 25 10:42:46 crc kubenswrapper[4821]: I1125 10:42:46.474586 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6jtmt" Nov 25 10:42:48 crc kubenswrapper[4821]: I1125 10:42:48.763445 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-d8wjl"] Nov 25 10:42:48 crc kubenswrapper[4821]: E1125 10:42:48.763947 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="faf2c05e-02b4-4b92-b614-d3de8e443677" containerName="util" Nov 25 10:42:48 crc kubenswrapper[4821]: I1125 10:42:48.763961 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="faf2c05e-02b4-4b92-b614-d3de8e443677" containerName="util" Nov 25 10:42:48 crc kubenswrapper[4821]: E1125 10:42:48.763972 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="faf2c05e-02b4-4b92-b614-d3de8e443677" containerName="extract" Nov 25 10:42:48 crc kubenswrapper[4821]: I1125 10:42:48.763979 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="faf2c05e-02b4-4b92-b614-d3de8e443677" containerName="extract" Nov 25 10:42:48 crc kubenswrapper[4821]: E1125 10:42:48.763989 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="faf2c05e-02b4-4b92-b614-d3de8e443677" containerName="pull" Nov 25 10:42:48 crc kubenswrapper[4821]: I1125 10:42:48.763995 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="faf2c05e-02b4-4b92-b614-d3de8e443677" containerName="pull" Nov 25 10:42:48 crc kubenswrapper[4821]: I1125 10:42:48.764077 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="faf2c05e-02b4-4b92-b614-d3de8e443677" containerName="extract" Nov 25 10:42:48 crc kubenswrapper[4821]: I1125 10:42:48.764470 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-557fdffb88-d8wjl" Nov 25 10:42:48 crc kubenswrapper[4821]: I1125 10:42:48.766205 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Nov 25 10:42:48 crc kubenswrapper[4821]: I1125 10:42:48.766301 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Nov 25 10:42:48 crc kubenswrapper[4821]: I1125 10:42:48.767825 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-cdnwf" Nov 25 10:42:48 crc kubenswrapper[4821]: I1125 10:42:48.775848 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-d8wjl"] Nov 25 10:42:48 crc kubenswrapper[4821]: I1125 10:42:48.856379 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8dsmt\" (UniqueName: \"kubernetes.io/projected/7218bb0e-8630-47d7-a259-3d4f79f1a470-kube-api-access-8dsmt\") pod \"nmstate-operator-557fdffb88-d8wjl\" (UID: \"7218bb0e-8630-47d7-a259-3d4f79f1a470\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-d8wjl" Nov 25 10:42:48 crc kubenswrapper[4821]: I1125 10:42:48.957187 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8dsmt\" (UniqueName: \"kubernetes.io/projected/7218bb0e-8630-47d7-a259-3d4f79f1a470-kube-api-access-8dsmt\") pod \"nmstate-operator-557fdffb88-d8wjl\" (UID: \"7218bb0e-8630-47d7-a259-3d4f79f1a470\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-d8wjl" Nov 25 10:42:48 crc kubenswrapper[4821]: I1125 10:42:48.972553 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8dsmt\" (UniqueName: \"kubernetes.io/projected/7218bb0e-8630-47d7-a259-3d4f79f1a470-kube-api-access-8dsmt\") pod \"nmstate-operator-557fdffb88-d8wjl\" (UID: \"7218bb0e-8630-47d7-a259-3d4f79f1a470\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-d8wjl" Nov 25 10:42:49 crc kubenswrapper[4821]: I1125 10:42:49.085282 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-557fdffb88-d8wjl" Nov 25 10:42:49 crc kubenswrapper[4821]: I1125 10:42:49.259861 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-d8wjl"] Nov 25 10:42:49 crc kubenswrapper[4821]: I1125 10:42:49.489403 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-557fdffb88-d8wjl" event={"ID":"7218bb0e-8630-47d7-a259-3d4f79f1a470","Type":"ContainerStarted","Data":"62d11caf2657b8294289208f5251facfdd0cb826b705cf1dea57dad2c5a32e1f"} Nov 25 10:42:52 crc kubenswrapper[4821]: I1125 10:42:52.507820 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-557fdffb88-d8wjl" event={"ID":"7218bb0e-8630-47d7-a259-3d4f79f1a470","Type":"ContainerStarted","Data":"303dc8b950a5d7fee18487ce2c93218f6519d2d9d8463d0b59108e754f5e692a"} Nov 25 10:42:52 crc kubenswrapper[4821]: I1125 10:42:52.525149 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-557fdffb88-d8wjl" podStartSLOduration=2.425721261 podStartE2EDuration="4.525128639s" podCreationTimestamp="2025-11-25 10:42:48 +0000 UTC" firstStartedPulling="2025-11-25 10:42:49.272797403 +0000 UTC m=+639.809117240" lastFinishedPulling="2025-11-25 10:42:51.372204771 +0000 UTC m=+641.908524618" observedRunningTime="2025-11-25 10:42:52.521717803 +0000 UTC m=+643.058037650" watchObservedRunningTime="2025-11-25 10:42:52.525128639 +0000 UTC m=+643.061448486" Nov 25 10:42:53 crc kubenswrapper[4821]: I1125 10:42:53.574320 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-79cx4"] Nov 25 10:42:53 crc kubenswrapper[4821]: I1125 10:42:53.575326 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-79cx4" Nov 25 10:42:53 crc kubenswrapper[4821]: I1125 10:42:53.579177 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-98g7n" Nov 25 10:42:53 crc kubenswrapper[4821]: I1125 10:42:53.587579 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-c9bsx"] Nov 25 10:42:53 crc kubenswrapper[4821]: I1125 10:42:53.588294 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-c9bsx" Nov 25 10:42:53 crc kubenswrapper[4821]: I1125 10:42:53.590084 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Nov 25 10:42:53 crc kubenswrapper[4821]: I1125 10:42:53.595640 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-79cx4"] Nov 25 10:42:53 crc kubenswrapper[4821]: I1125 10:42:53.610127 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nlz4b\" (UniqueName: \"kubernetes.io/projected/221a1f74-bcbb-4788-83d1-4c5c92d3fedd-kube-api-access-nlz4b\") pod \"nmstate-metrics-5dcf9c57c5-79cx4\" (UID: \"221a1f74-bcbb-4788-83d1-4c5c92d3fedd\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-79cx4" Nov 25 10:42:53 crc kubenswrapper[4821]: I1125 10:42:53.610254 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/de41e739-6f86-44cc-80a2-99bb874d259f-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-c9bsx\" (UID: \"de41e739-6f86-44cc-80a2-99bb874d259f\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-c9bsx" Nov 25 10:42:53 crc kubenswrapper[4821]: I1125 10:42:53.610306 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhrq5\" (UniqueName: \"kubernetes.io/projected/de41e739-6f86-44cc-80a2-99bb874d259f-kube-api-access-zhrq5\") pod \"nmstate-webhook-6b89b748d8-c9bsx\" (UID: \"de41e739-6f86-44cc-80a2-99bb874d259f\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-c9bsx" Nov 25 10:42:53 crc kubenswrapper[4821]: I1125 10:42:53.613061 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-89r7d"] Nov 25 10:42:53 crc kubenswrapper[4821]: I1125 10:42:53.625054 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-c9bsx"] Nov 25 10:42:53 crc kubenswrapper[4821]: I1125 10:42:53.625192 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-89r7d" Nov 25 10:42:53 crc kubenswrapper[4821]: I1125 10:42:53.711271 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/de41e739-6f86-44cc-80a2-99bb874d259f-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-c9bsx\" (UID: \"de41e739-6f86-44cc-80a2-99bb874d259f\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-c9bsx" Nov 25 10:42:53 crc kubenswrapper[4821]: I1125 10:42:53.711329 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhrq5\" (UniqueName: \"kubernetes.io/projected/de41e739-6f86-44cc-80a2-99bb874d259f-kube-api-access-zhrq5\") pod \"nmstate-webhook-6b89b748d8-c9bsx\" (UID: \"de41e739-6f86-44cc-80a2-99bb874d259f\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-c9bsx" Nov 25 10:42:53 crc kubenswrapper[4821]: I1125 10:42:53.711380 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nlz4b\" (UniqueName: \"kubernetes.io/projected/221a1f74-bcbb-4788-83d1-4c5c92d3fedd-kube-api-access-nlz4b\") pod \"nmstate-metrics-5dcf9c57c5-79cx4\" (UID: \"221a1f74-bcbb-4788-83d1-4c5c92d3fedd\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-79cx4" Nov 25 10:42:53 crc kubenswrapper[4821]: I1125 10:42:53.718915 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/de41e739-6f86-44cc-80a2-99bb874d259f-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-c9bsx\" (UID: \"de41e739-6f86-44cc-80a2-99bb874d259f\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-c9bsx" Nov 25 10:42:53 crc kubenswrapper[4821]: I1125 10:42:53.724048 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-7mnfd"] Nov 25 10:42:53 crc kubenswrapper[4821]: I1125 10:42:53.724649 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-7mnfd" Nov 25 10:42:53 crc kubenswrapper[4821]: I1125 10:42:53.726622 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Nov 25 10:42:53 crc kubenswrapper[4821]: I1125 10:42:53.726831 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-m9q7f" Nov 25 10:42:53 crc kubenswrapper[4821]: I1125 10:42:53.727269 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Nov 25 10:42:53 crc kubenswrapper[4821]: I1125 10:42:53.731064 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nlz4b\" (UniqueName: \"kubernetes.io/projected/221a1f74-bcbb-4788-83d1-4c5c92d3fedd-kube-api-access-nlz4b\") pod \"nmstate-metrics-5dcf9c57c5-79cx4\" (UID: \"221a1f74-bcbb-4788-83d1-4c5c92d3fedd\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-79cx4" Nov 25 10:42:53 crc kubenswrapper[4821]: I1125 10:42:53.734935 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhrq5\" (UniqueName: \"kubernetes.io/projected/de41e739-6f86-44cc-80a2-99bb874d259f-kube-api-access-zhrq5\") pod \"nmstate-webhook-6b89b748d8-c9bsx\" (UID: \"de41e739-6f86-44cc-80a2-99bb874d259f\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-c9bsx" Nov 25 10:42:53 crc kubenswrapper[4821]: I1125 10:42:53.741816 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-7mnfd"] Nov 25 10:42:53 crc kubenswrapper[4821]: I1125 10:42:53.813027 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/e29f6151-8a55-4b3e-8a15-8f8ec4b5ac66-nmstate-lock\") pod \"nmstate-handler-89r7d\" (UID: \"e29f6151-8a55-4b3e-8a15-8f8ec4b5ac66\") " pod="openshift-nmstate/nmstate-handler-89r7d" Nov 25 10:42:53 crc kubenswrapper[4821]: I1125 10:42:53.813087 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lz4pl\" (UniqueName: \"kubernetes.io/projected/e29f6151-8a55-4b3e-8a15-8f8ec4b5ac66-kube-api-access-lz4pl\") pod \"nmstate-handler-89r7d\" (UID: \"e29f6151-8a55-4b3e-8a15-8f8ec4b5ac66\") " pod="openshift-nmstate/nmstate-handler-89r7d" Nov 25 10:42:53 crc kubenswrapper[4821]: I1125 10:42:53.813740 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/e29f6151-8a55-4b3e-8a15-8f8ec4b5ac66-ovs-socket\") pod \"nmstate-handler-89r7d\" (UID: \"e29f6151-8a55-4b3e-8a15-8f8ec4b5ac66\") " pod="openshift-nmstate/nmstate-handler-89r7d" Nov 25 10:42:53 crc kubenswrapper[4821]: I1125 10:42:53.813784 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/e29f6151-8a55-4b3e-8a15-8f8ec4b5ac66-dbus-socket\") pod \"nmstate-handler-89r7d\" (UID: \"e29f6151-8a55-4b3e-8a15-8f8ec4b5ac66\") " pod="openshift-nmstate/nmstate-handler-89r7d" Nov 25 10:42:53 crc kubenswrapper[4821]: I1125 10:42:53.893315 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-66744675f6-5krzk"] Nov 25 10:42:53 crc kubenswrapper[4821]: I1125 10:42:53.894031 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-66744675f6-5krzk" Nov 25 10:42:53 crc kubenswrapper[4821]: I1125 10:42:53.894445 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-79cx4" Nov 25 10:42:53 crc kubenswrapper[4821]: I1125 10:42:53.906400 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-c9bsx" Nov 25 10:42:53 crc kubenswrapper[4821]: I1125 10:42:53.907470 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-66744675f6-5krzk"] Nov 25 10:42:53 crc kubenswrapper[4821]: I1125 10:42:53.918446 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9q7pg\" (UniqueName: \"kubernetes.io/projected/dff35e42-91ef-41f1-81b9-c12191388f31-kube-api-access-9q7pg\") pod \"nmstate-console-plugin-5874bd7bc5-7mnfd\" (UID: \"dff35e42-91ef-41f1-81b9-c12191388f31\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-7mnfd" Nov 25 10:42:53 crc kubenswrapper[4821]: I1125 10:42:53.918509 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/dff35e42-91ef-41f1-81b9-c12191388f31-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-7mnfd\" (UID: \"dff35e42-91ef-41f1-81b9-c12191388f31\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-7mnfd" Nov 25 10:42:53 crc kubenswrapper[4821]: I1125 10:42:53.918542 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/dff35e42-91ef-41f1-81b9-c12191388f31-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-7mnfd\" (UID: \"dff35e42-91ef-41f1-81b9-c12191388f31\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-7mnfd" Nov 25 10:42:53 crc kubenswrapper[4821]: I1125 10:42:53.918579 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/e29f6151-8a55-4b3e-8a15-8f8ec4b5ac66-ovs-socket\") pod \"nmstate-handler-89r7d\" (UID: \"e29f6151-8a55-4b3e-8a15-8f8ec4b5ac66\") " pod="openshift-nmstate/nmstate-handler-89r7d" Nov 25 10:42:53 crc kubenswrapper[4821]: I1125 10:42:53.918608 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/e29f6151-8a55-4b3e-8a15-8f8ec4b5ac66-dbus-socket\") pod \"nmstate-handler-89r7d\" (UID: \"e29f6151-8a55-4b3e-8a15-8f8ec4b5ac66\") " pod="openshift-nmstate/nmstate-handler-89r7d" Nov 25 10:42:53 crc kubenswrapper[4821]: I1125 10:42:53.918642 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/e29f6151-8a55-4b3e-8a15-8f8ec4b5ac66-nmstate-lock\") pod \"nmstate-handler-89r7d\" (UID: \"e29f6151-8a55-4b3e-8a15-8f8ec4b5ac66\") " pod="openshift-nmstate/nmstate-handler-89r7d" Nov 25 10:42:53 crc kubenswrapper[4821]: I1125 10:42:53.918667 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lz4pl\" (UniqueName: \"kubernetes.io/projected/e29f6151-8a55-4b3e-8a15-8f8ec4b5ac66-kube-api-access-lz4pl\") pod \"nmstate-handler-89r7d\" (UID: \"e29f6151-8a55-4b3e-8a15-8f8ec4b5ac66\") " pod="openshift-nmstate/nmstate-handler-89r7d" Nov 25 10:42:53 crc kubenswrapper[4821]: I1125 10:42:53.919822 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/e29f6151-8a55-4b3e-8a15-8f8ec4b5ac66-ovs-socket\") pod \"nmstate-handler-89r7d\" (UID: \"e29f6151-8a55-4b3e-8a15-8f8ec4b5ac66\") " pod="openshift-nmstate/nmstate-handler-89r7d" Nov 25 10:42:53 crc kubenswrapper[4821]: I1125 10:42:53.920112 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/e29f6151-8a55-4b3e-8a15-8f8ec4b5ac66-dbus-socket\") pod \"nmstate-handler-89r7d\" (UID: \"e29f6151-8a55-4b3e-8a15-8f8ec4b5ac66\") " pod="openshift-nmstate/nmstate-handler-89r7d" Nov 25 10:42:53 crc kubenswrapper[4821]: I1125 10:42:53.920143 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/e29f6151-8a55-4b3e-8a15-8f8ec4b5ac66-nmstate-lock\") pod \"nmstate-handler-89r7d\" (UID: \"e29f6151-8a55-4b3e-8a15-8f8ec4b5ac66\") " pod="openshift-nmstate/nmstate-handler-89r7d" Nov 25 10:42:53 crc kubenswrapper[4821]: I1125 10:42:53.951924 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lz4pl\" (UniqueName: \"kubernetes.io/projected/e29f6151-8a55-4b3e-8a15-8f8ec4b5ac66-kube-api-access-lz4pl\") pod \"nmstate-handler-89r7d\" (UID: \"e29f6151-8a55-4b3e-8a15-8f8ec4b5ac66\") " pod="openshift-nmstate/nmstate-handler-89r7d" Nov 25 10:42:54 crc kubenswrapper[4821]: I1125 10:42:54.019605 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/99be33ca-2610-4006-a231-663dee9a75c4-console-oauth-config\") pod \"console-66744675f6-5krzk\" (UID: \"99be33ca-2610-4006-a231-663dee9a75c4\") " pod="openshift-console/console-66744675f6-5krzk" Nov 25 10:42:54 crc kubenswrapper[4821]: I1125 10:42:54.019655 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/99be33ca-2610-4006-a231-663dee9a75c4-trusted-ca-bundle\") pod \"console-66744675f6-5krzk\" (UID: \"99be33ca-2610-4006-a231-663dee9a75c4\") " pod="openshift-console/console-66744675f6-5krzk" Nov 25 10:42:54 crc kubenswrapper[4821]: I1125 10:42:54.019687 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/99be33ca-2610-4006-a231-663dee9a75c4-console-serving-cert\") pod \"console-66744675f6-5krzk\" (UID: \"99be33ca-2610-4006-a231-663dee9a75c4\") " pod="openshift-console/console-66744675f6-5krzk" Nov 25 10:42:54 crc kubenswrapper[4821]: I1125 10:42:54.019713 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/99be33ca-2610-4006-a231-663dee9a75c4-service-ca\") pod \"console-66744675f6-5krzk\" (UID: \"99be33ca-2610-4006-a231-663dee9a75c4\") " pod="openshift-console/console-66744675f6-5krzk" Nov 25 10:42:54 crc kubenswrapper[4821]: I1125 10:42:54.019746 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/99be33ca-2610-4006-a231-663dee9a75c4-oauth-serving-cert\") pod \"console-66744675f6-5krzk\" (UID: \"99be33ca-2610-4006-a231-663dee9a75c4\") " pod="openshift-console/console-66744675f6-5krzk" Nov 25 10:42:54 crc kubenswrapper[4821]: I1125 10:42:54.019776 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9q7pg\" (UniqueName: \"kubernetes.io/projected/dff35e42-91ef-41f1-81b9-c12191388f31-kube-api-access-9q7pg\") pod \"nmstate-console-plugin-5874bd7bc5-7mnfd\" (UID: \"dff35e42-91ef-41f1-81b9-c12191388f31\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-7mnfd" Nov 25 10:42:54 crc kubenswrapper[4821]: I1125 10:42:54.019803 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/99be33ca-2610-4006-a231-663dee9a75c4-console-config\") pod \"console-66744675f6-5krzk\" (UID: \"99be33ca-2610-4006-a231-663dee9a75c4\") " pod="openshift-console/console-66744675f6-5krzk" Nov 25 10:42:54 crc kubenswrapper[4821]: I1125 10:42:54.019840 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/dff35e42-91ef-41f1-81b9-c12191388f31-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-7mnfd\" (UID: \"dff35e42-91ef-41f1-81b9-c12191388f31\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-7mnfd" Nov 25 10:42:54 crc kubenswrapper[4821]: I1125 10:42:54.019879 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/dff35e42-91ef-41f1-81b9-c12191388f31-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-7mnfd\" (UID: \"dff35e42-91ef-41f1-81b9-c12191388f31\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-7mnfd" Nov 25 10:42:54 crc kubenswrapper[4821]: I1125 10:42:54.019912 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p698w\" (UniqueName: \"kubernetes.io/projected/99be33ca-2610-4006-a231-663dee9a75c4-kube-api-access-p698w\") pod \"console-66744675f6-5krzk\" (UID: \"99be33ca-2610-4006-a231-663dee9a75c4\") " pod="openshift-console/console-66744675f6-5krzk" Nov 25 10:42:54 crc kubenswrapper[4821]: I1125 10:42:54.023689 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/dff35e42-91ef-41f1-81b9-c12191388f31-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-7mnfd\" (UID: \"dff35e42-91ef-41f1-81b9-c12191388f31\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-7mnfd" Nov 25 10:42:54 crc kubenswrapper[4821]: I1125 10:42:54.030681 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/dff35e42-91ef-41f1-81b9-c12191388f31-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-7mnfd\" (UID: \"dff35e42-91ef-41f1-81b9-c12191388f31\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-7mnfd" Nov 25 10:42:54 crc kubenswrapper[4821]: I1125 10:42:54.042922 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9q7pg\" (UniqueName: \"kubernetes.io/projected/dff35e42-91ef-41f1-81b9-c12191388f31-kube-api-access-9q7pg\") pod \"nmstate-console-plugin-5874bd7bc5-7mnfd\" (UID: \"dff35e42-91ef-41f1-81b9-c12191388f31\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-7mnfd" Nov 25 10:42:54 crc kubenswrapper[4821]: I1125 10:42:54.079452 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-7mnfd" Nov 25 10:42:54 crc kubenswrapper[4821]: I1125 10:42:54.120788 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p698w\" (UniqueName: \"kubernetes.io/projected/99be33ca-2610-4006-a231-663dee9a75c4-kube-api-access-p698w\") pod \"console-66744675f6-5krzk\" (UID: \"99be33ca-2610-4006-a231-663dee9a75c4\") " pod="openshift-console/console-66744675f6-5krzk" Nov 25 10:42:54 crc kubenswrapper[4821]: I1125 10:42:54.120871 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/99be33ca-2610-4006-a231-663dee9a75c4-console-oauth-config\") pod \"console-66744675f6-5krzk\" (UID: \"99be33ca-2610-4006-a231-663dee9a75c4\") " pod="openshift-console/console-66744675f6-5krzk" Nov 25 10:42:54 crc kubenswrapper[4821]: I1125 10:42:54.120906 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/99be33ca-2610-4006-a231-663dee9a75c4-trusted-ca-bundle\") pod \"console-66744675f6-5krzk\" (UID: \"99be33ca-2610-4006-a231-663dee9a75c4\") " pod="openshift-console/console-66744675f6-5krzk" Nov 25 10:42:54 crc kubenswrapper[4821]: I1125 10:42:54.120935 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/99be33ca-2610-4006-a231-663dee9a75c4-console-serving-cert\") pod \"console-66744675f6-5krzk\" (UID: \"99be33ca-2610-4006-a231-663dee9a75c4\") " pod="openshift-console/console-66744675f6-5krzk" Nov 25 10:42:54 crc kubenswrapper[4821]: I1125 10:42:54.120965 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/99be33ca-2610-4006-a231-663dee9a75c4-service-ca\") pod \"console-66744675f6-5krzk\" (UID: \"99be33ca-2610-4006-a231-663dee9a75c4\") " pod="openshift-console/console-66744675f6-5krzk" Nov 25 10:42:54 crc kubenswrapper[4821]: I1125 10:42:54.121060 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/99be33ca-2610-4006-a231-663dee9a75c4-oauth-serving-cert\") pod \"console-66744675f6-5krzk\" (UID: \"99be33ca-2610-4006-a231-663dee9a75c4\") " pod="openshift-console/console-66744675f6-5krzk" Nov 25 10:42:54 crc kubenswrapper[4821]: I1125 10:42:54.121360 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/99be33ca-2610-4006-a231-663dee9a75c4-console-config\") pod \"console-66744675f6-5krzk\" (UID: \"99be33ca-2610-4006-a231-663dee9a75c4\") " pod="openshift-console/console-66744675f6-5krzk" Nov 25 10:42:54 crc kubenswrapper[4821]: I1125 10:42:54.122238 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/99be33ca-2610-4006-a231-663dee9a75c4-oauth-serving-cert\") pod \"console-66744675f6-5krzk\" (UID: \"99be33ca-2610-4006-a231-663dee9a75c4\") " pod="openshift-console/console-66744675f6-5krzk" Nov 25 10:42:54 crc kubenswrapper[4821]: I1125 10:42:54.122348 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/99be33ca-2610-4006-a231-663dee9a75c4-service-ca\") pod \"console-66744675f6-5krzk\" (UID: \"99be33ca-2610-4006-a231-663dee9a75c4\") " pod="openshift-console/console-66744675f6-5krzk" Nov 25 10:42:54 crc kubenswrapper[4821]: I1125 10:42:54.123043 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/99be33ca-2610-4006-a231-663dee9a75c4-trusted-ca-bundle\") pod \"console-66744675f6-5krzk\" (UID: \"99be33ca-2610-4006-a231-663dee9a75c4\") " pod="openshift-console/console-66744675f6-5krzk" Nov 25 10:42:54 crc kubenswrapper[4821]: I1125 10:42:54.124084 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/99be33ca-2610-4006-a231-663dee9a75c4-console-config\") pod \"console-66744675f6-5krzk\" (UID: \"99be33ca-2610-4006-a231-663dee9a75c4\") " pod="openshift-console/console-66744675f6-5krzk" Nov 25 10:42:54 crc kubenswrapper[4821]: I1125 10:42:54.126489 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/99be33ca-2610-4006-a231-663dee9a75c4-console-serving-cert\") pod \"console-66744675f6-5krzk\" (UID: \"99be33ca-2610-4006-a231-663dee9a75c4\") " pod="openshift-console/console-66744675f6-5krzk" Nov 25 10:42:54 crc kubenswrapper[4821]: I1125 10:42:54.126680 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/99be33ca-2610-4006-a231-663dee9a75c4-console-oauth-config\") pod \"console-66744675f6-5krzk\" (UID: \"99be33ca-2610-4006-a231-663dee9a75c4\") " pod="openshift-console/console-66744675f6-5krzk" Nov 25 10:42:54 crc kubenswrapper[4821]: I1125 10:42:54.138073 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p698w\" (UniqueName: \"kubernetes.io/projected/99be33ca-2610-4006-a231-663dee9a75c4-kube-api-access-p698w\") pod \"console-66744675f6-5krzk\" (UID: \"99be33ca-2610-4006-a231-663dee9a75c4\") " pod="openshift-console/console-66744675f6-5krzk" Nov 25 10:42:54 crc kubenswrapper[4821]: I1125 10:42:54.149972 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-79cx4"] Nov 25 10:42:54 crc kubenswrapper[4821]: W1125 10:42:54.156681 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod221a1f74_bcbb_4788_83d1_4c5c92d3fedd.slice/crio-04562de51029c9e5c8a6426e9c331b5f022b3aa97bcc6b5ab7b12443ecda2751 WatchSource:0}: Error finding container 04562de51029c9e5c8a6426e9c331b5f022b3aa97bcc6b5ab7b12443ecda2751: Status 404 returned error can't find the container with id 04562de51029c9e5c8a6426e9c331b5f022b3aa97bcc6b5ab7b12443ecda2751 Nov 25 10:42:54 crc kubenswrapper[4821]: I1125 10:42:54.242675 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-89r7d" Nov 25 10:42:54 crc kubenswrapper[4821]: I1125 10:42:54.270404 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-7mnfd"] Nov 25 10:42:54 crc kubenswrapper[4821]: I1125 10:42:54.272091 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-66744675f6-5krzk" Nov 25 10:42:54 crc kubenswrapper[4821]: W1125 10:42:54.273366 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddff35e42_91ef_41f1_81b9_c12191388f31.slice/crio-4bfab23faeaec5a021a4caf1a567f2b4cbff97e2b9ec9cdf269462217f8884a9 WatchSource:0}: Error finding container 4bfab23faeaec5a021a4caf1a567f2b4cbff97e2b9ec9cdf269462217f8884a9: Status 404 returned error can't find the container with id 4bfab23faeaec5a021a4caf1a567f2b4cbff97e2b9ec9cdf269462217f8884a9 Nov 25 10:42:54 crc kubenswrapper[4821]: W1125 10:42:54.274746 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode29f6151_8a55_4b3e_8a15_8f8ec4b5ac66.slice/crio-81f4961bf44f03724ba6e74242513a1d8c4e41db1a3df55c36b897f5510ccdaf WatchSource:0}: Error finding container 81f4961bf44f03724ba6e74242513a1d8c4e41db1a3df55c36b897f5510ccdaf: Status 404 returned error can't find the container with id 81f4961bf44f03724ba6e74242513a1d8c4e41db1a3df55c36b897f5510ccdaf Nov 25 10:42:54 crc kubenswrapper[4821]: I1125 10:42:54.386699 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-c9bsx"] Nov 25 10:42:54 crc kubenswrapper[4821]: W1125 10:42:54.395719 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podde41e739_6f86_44cc_80a2_99bb874d259f.slice/crio-8782d819a98b380de5537c174aa8bcffcc588ab114335f28aa53402f0c802b91 WatchSource:0}: Error finding container 8782d819a98b380de5537c174aa8bcffcc588ab114335f28aa53402f0c802b91: Status 404 returned error can't find the container with id 8782d819a98b380de5537c174aa8bcffcc588ab114335f28aa53402f0c802b91 Nov 25 10:42:54 crc kubenswrapper[4821]: I1125 10:42:54.438012 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-66744675f6-5krzk"] Nov 25 10:42:54 crc kubenswrapper[4821]: W1125 10:42:54.443039 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod99be33ca_2610_4006_a231_663dee9a75c4.slice/crio-cc802d504cfbf5f1c43f398b594797bbc2ddf05893fae603dec9091609b23758 WatchSource:0}: Error finding container cc802d504cfbf5f1c43f398b594797bbc2ddf05893fae603dec9091609b23758: Status 404 returned error can't find the container with id cc802d504cfbf5f1c43f398b594797bbc2ddf05893fae603dec9091609b23758 Nov 25 10:42:54 crc kubenswrapper[4821]: I1125 10:42:54.519545 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-c9bsx" event={"ID":"de41e739-6f86-44cc-80a2-99bb874d259f","Type":"ContainerStarted","Data":"8782d819a98b380de5537c174aa8bcffcc588ab114335f28aa53402f0c802b91"} Nov 25 10:42:54 crc kubenswrapper[4821]: I1125 10:42:54.520474 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-79cx4" event={"ID":"221a1f74-bcbb-4788-83d1-4c5c92d3fedd","Type":"ContainerStarted","Data":"04562de51029c9e5c8a6426e9c331b5f022b3aa97bcc6b5ab7b12443ecda2751"} Nov 25 10:42:54 crc kubenswrapper[4821]: I1125 10:42:54.523322 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-89r7d" event={"ID":"e29f6151-8a55-4b3e-8a15-8f8ec4b5ac66","Type":"ContainerStarted","Data":"81f4961bf44f03724ba6e74242513a1d8c4e41db1a3df55c36b897f5510ccdaf"} Nov 25 10:42:54 crc kubenswrapper[4821]: I1125 10:42:54.524020 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-7mnfd" event={"ID":"dff35e42-91ef-41f1-81b9-c12191388f31","Type":"ContainerStarted","Data":"4bfab23faeaec5a021a4caf1a567f2b4cbff97e2b9ec9cdf269462217f8884a9"} Nov 25 10:42:54 crc kubenswrapper[4821]: I1125 10:42:54.525915 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-66744675f6-5krzk" event={"ID":"99be33ca-2610-4006-a231-663dee9a75c4","Type":"ContainerStarted","Data":"cc802d504cfbf5f1c43f398b594797bbc2ddf05893fae603dec9091609b23758"} Nov 25 10:42:55 crc kubenswrapper[4821]: I1125 10:42:55.533776 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-66744675f6-5krzk" event={"ID":"99be33ca-2610-4006-a231-663dee9a75c4","Type":"ContainerStarted","Data":"90821341b7c893340d7c4930ac2aedeb0c0f9952a503ad47d47ae7575266c4c9"} Nov 25 10:42:55 crc kubenswrapper[4821]: I1125 10:42:55.553102 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-66744675f6-5krzk" podStartSLOduration=2.553084885 podStartE2EDuration="2.553084885s" podCreationTimestamp="2025-11-25 10:42:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:42:55.550302053 +0000 UTC m=+646.086621910" watchObservedRunningTime="2025-11-25 10:42:55.553084885 +0000 UTC m=+646.089404732" Nov 25 10:42:58 crc kubenswrapper[4821]: I1125 10:42:58.551279 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-89r7d" event={"ID":"e29f6151-8a55-4b3e-8a15-8f8ec4b5ac66","Type":"ContainerStarted","Data":"3335edafe1401ef8b9fd5c73f3cd4aa55133dbfc2819aeb5a205ce38f2748d1c"} Nov 25 10:42:58 crc kubenswrapper[4821]: I1125 10:42:58.551866 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-89r7d" Nov 25 10:42:58 crc kubenswrapper[4821]: I1125 10:42:58.552708 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-7mnfd" event={"ID":"dff35e42-91ef-41f1-81b9-c12191388f31","Type":"ContainerStarted","Data":"b66bee01f8d746374e339a930ff0708776adbf8ed83f5d7c092f94f498d54515"} Nov 25 10:42:58 crc kubenswrapper[4821]: I1125 10:42:58.554719 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-c9bsx" event={"ID":"de41e739-6f86-44cc-80a2-99bb874d259f","Type":"ContainerStarted","Data":"0532db17a36f599899fc66a9c418a5a5a18c0b363a15e0edee0da4e3eebd54c6"} Nov 25 10:42:58 crc kubenswrapper[4821]: I1125 10:42:58.554775 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-c9bsx" Nov 25 10:42:58 crc kubenswrapper[4821]: I1125 10:42:58.556910 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-79cx4" event={"ID":"221a1f74-bcbb-4788-83d1-4c5c92d3fedd","Type":"ContainerStarted","Data":"0d5731e29ddfec1c584ca683056d94220a7f1899f0bd57e9cad7d4a555685333"} Nov 25 10:42:58 crc kubenswrapper[4821]: I1125 10:42:58.568667 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-89r7d" podStartSLOduration=2.209661716 podStartE2EDuration="5.568648662s" podCreationTimestamp="2025-11-25 10:42:53 +0000 UTC" firstStartedPulling="2025-11-25 10:42:54.279641874 +0000 UTC m=+644.815961741" lastFinishedPulling="2025-11-25 10:42:57.63862884 +0000 UTC m=+648.174948687" observedRunningTime="2025-11-25 10:42:58.568324964 +0000 UTC m=+649.104644811" watchObservedRunningTime="2025-11-25 10:42:58.568648662 +0000 UTC m=+649.104968509" Nov 25 10:42:58 crc kubenswrapper[4821]: I1125 10:42:58.583385 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-c9bsx" podStartSLOduration=2.350968303 podStartE2EDuration="5.58336946s" podCreationTimestamp="2025-11-25 10:42:53 +0000 UTC" firstStartedPulling="2025-11-25 10:42:54.401214655 +0000 UTC m=+644.937534502" lastFinishedPulling="2025-11-25 10:42:57.633615812 +0000 UTC m=+648.169935659" observedRunningTime="2025-11-25 10:42:58.581287796 +0000 UTC m=+649.117607653" watchObservedRunningTime="2025-11-25 10:42:58.58336946 +0000 UTC m=+649.119689307" Nov 25 10:42:58 crc kubenswrapper[4821]: I1125 10:42:58.640052 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-7mnfd" podStartSLOduration=2.300535961 podStartE2EDuration="5.640032049s" podCreationTimestamp="2025-11-25 10:42:53 +0000 UTC" firstStartedPulling="2025-11-25 10:42:54.279461609 +0000 UTC m=+644.815781456" lastFinishedPulling="2025-11-25 10:42:57.618957697 +0000 UTC m=+648.155277544" observedRunningTime="2025-11-25 10:42:58.638586622 +0000 UTC m=+649.174906469" watchObservedRunningTime="2025-11-25 10:42:58.640032049 +0000 UTC m=+649.176351896" Nov 25 10:43:00 crc kubenswrapper[4821]: I1125 10:43:00.573211 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-79cx4" event={"ID":"221a1f74-bcbb-4788-83d1-4c5c92d3fedd","Type":"ContainerStarted","Data":"3f82fc794f04756690cc7ffe04423c29baf880619a2985a594409caf40cdbcf7"} Nov 25 10:43:00 crc kubenswrapper[4821]: I1125 10:43:00.598343 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-79cx4" podStartSLOduration=1.915283367 podStartE2EDuration="7.598321822s" podCreationTimestamp="2025-11-25 10:42:53 +0000 UTC" firstStartedPulling="2025-11-25 10:42:54.159190551 +0000 UTC m=+644.695510398" lastFinishedPulling="2025-11-25 10:42:59.842229006 +0000 UTC m=+650.378548853" observedRunningTime="2025-11-25 10:43:00.597378127 +0000 UTC m=+651.133697984" watchObservedRunningTime="2025-11-25 10:43:00.598321822 +0000 UTC m=+651.134641669" Nov 25 10:43:04 crc kubenswrapper[4821]: I1125 10:43:04.264253 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-89r7d" Nov 25 10:43:04 crc kubenswrapper[4821]: I1125 10:43:04.273640 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-66744675f6-5krzk" Nov 25 10:43:04 crc kubenswrapper[4821]: I1125 10:43:04.273782 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-66744675f6-5krzk" Nov 25 10:43:04 crc kubenswrapper[4821]: I1125 10:43:04.280893 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-66744675f6-5krzk" Nov 25 10:43:04 crc kubenswrapper[4821]: I1125 10:43:04.605113 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-66744675f6-5krzk" Nov 25 10:43:04 crc kubenswrapper[4821]: I1125 10:43:04.650726 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-hf49d"] Nov 25 10:43:13 crc kubenswrapper[4821]: I1125 10:43:13.912364 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-c9bsx" Nov 25 10:43:27 crc kubenswrapper[4821]: I1125 10:43:27.734070 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6hsmqh"] Nov 25 10:43:27 crc kubenswrapper[4821]: I1125 10:43:27.735892 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6hsmqh" Nov 25 10:43:27 crc kubenswrapper[4821]: I1125 10:43:27.737918 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 25 10:43:27 crc kubenswrapper[4821]: I1125 10:43:27.751802 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6hsmqh"] Nov 25 10:43:27 crc kubenswrapper[4821]: I1125 10:43:27.865600 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8c67f775-e13f-4c7d-9d2b-115d492961f1-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6hsmqh\" (UID: \"8c67f775-e13f-4c7d-9d2b-115d492961f1\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6hsmqh" Nov 25 10:43:27 crc kubenswrapper[4821]: I1125 10:43:27.865829 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8c67f775-e13f-4c7d-9d2b-115d492961f1-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6hsmqh\" (UID: \"8c67f775-e13f-4c7d-9d2b-115d492961f1\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6hsmqh" Nov 25 10:43:27 crc kubenswrapper[4821]: I1125 10:43:27.865914 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txlv6\" (UniqueName: \"kubernetes.io/projected/8c67f775-e13f-4c7d-9d2b-115d492961f1-kube-api-access-txlv6\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6hsmqh\" (UID: \"8c67f775-e13f-4c7d-9d2b-115d492961f1\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6hsmqh" Nov 25 10:43:27 crc kubenswrapper[4821]: I1125 10:43:27.967488 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8c67f775-e13f-4c7d-9d2b-115d492961f1-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6hsmqh\" (UID: \"8c67f775-e13f-4c7d-9d2b-115d492961f1\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6hsmqh" Nov 25 10:43:27 crc kubenswrapper[4821]: I1125 10:43:27.967561 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8c67f775-e13f-4c7d-9d2b-115d492961f1-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6hsmqh\" (UID: \"8c67f775-e13f-4c7d-9d2b-115d492961f1\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6hsmqh" Nov 25 10:43:27 crc kubenswrapper[4821]: I1125 10:43:27.967588 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txlv6\" (UniqueName: \"kubernetes.io/projected/8c67f775-e13f-4c7d-9d2b-115d492961f1-kube-api-access-txlv6\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6hsmqh\" (UID: \"8c67f775-e13f-4c7d-9d2b-115d492961f1\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6hsmqh" Nov 25 10:43:27 crc kubenswrapper[4821]: I1125 10:43:27.968154 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8c67f775-e13f-4c7d-9d2b-115d492961f1-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6hsmqh\" (UID: \"8c67f775-e13f-4c7d-9d2b-115d492961f1\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6hsmqh" Nov 25 10:43:27 crc kubenswrapper[4821]: I1125 10:43:27.968352 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8c67f775-e13f-4c7d-9d2b-115d492961f1-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6hsmqh\" (UID: \"8c67f775-e13f-4c7d-9d2b-115d492961f1\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6hsmqh" Nov 25 10:43:27 crc kubenswrapper[4821]: I1125 10:43:27.991230 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txlv6\" (UniqueName: \"kubernetes.io/projected/8c67f775-e13f-4c7d-9d2b-115d492961f1-kube-api-access-txlv6\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6hsmqh\" (UID: \"8c67f775-e13f-4c7d-9d2b-115d492961f1\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6hsmqh" Nov 25 10:43:28 crc kubenswrapper[4821]: I1125 10:43:28.053594 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6hsmqh" Nov 25 10:43:28 crc kubenswrapper[4821]: I1125 10:43:28.508085 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6hsmqh"] Nov 25 10:43:28 crc kubenswrapper[4821]: I1125 10:43:28.736430 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6hsmqh" event={"ID":"8c67f775-e13f-4c7d-9d2b-115d492961f1","Type":"ContainerStarted","Data":"b1d1281f0be96f9806cb89147b85215970a357f6e064a5125f3b7398eecc4de7"} Nov 25 10:43:28 crc kubenswrapper[4821]: I1125 10:43:28.736474 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6hsmqh" event={"ID":"8c67f775-e13f-4c7d-9d2b-115d492961f1","Type":"ContainerStarted","Data":"a3ddf997e3bf3b7dddfd1a50983a8a698ce9273ab4e3c2f4a4e425c4a14564f3"} Nov 25 10:43:29 crc kubenswrapper[4821]: I1125 10:43:29.689907 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-hf49d" podUID="88725123-3da9-4127-9506-80c90c9accd5" containerName="console" containerID="cri-o://3dd116757d3852f02a7dd5f0a100fafa105ad3f81d3d6be8c6c22129dab37315" gracePeriod=15 Nov 25 10:43:29 crc kubenswrapper[4821]: I1125 10:43:29.744923 4821 generic.go:334] "Generic (PLEG): container finished" podID="8c67f775-e13f-4c7d-9d2b-115d492961f1" containerID="b1d1281f0be96f9806cb89147b85215970a357f6e064a5125f3b7398eecc4de7" exitCode=0 Nov 25 10:43:29 crc kubenswrapper[4821]: I1125 10:43:29.745055 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6hsmqh" event={"ID":"8c67f775-e13f-4c7d-9d2b-115d492961f1","Type":"ContainerDied","Data":"b1d1281f0be96f9806cb89147b85215970a357f6e064a5125f3b7398eecc4de7"} Nov 25 10:43:30 crc kubenswrapper[4821]: I1125 10:43:30.028882 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-hf49d_88725123-3da9-4127-9506-80c90c9accd5/console/0.log" Nov 25 10:43:30 crc kubenswrapper[4821]: I1125 10:43:30.028943 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-hf49d" Nov 25 10:43:30 crc kubenswrapper[4821]: I1125 10:43:30.196399 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/88725123-3da9-4127-9506-80c90c9accd5-trusted-ca-bundle\") pod \"88725123-3da9-4127-9506-80c90c9accd5\" (UID: \"88725123-3da9-4127-9506-80c90c9accd5\") " Nov 25 10:43:30 crc kubenswrapper[4821]: I1125 10:43:30.196441 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5x4xr\" (UniqueName: \"kubernetes.io/projected/88725123-3da9-4127-9506-80c90c9accd5-kube-api-access-5x4xr\") pod \"88725123-3da9-4127-9506-80c90c9accd5\" (UID: \"88725123-3da9-4127-9506-80c90c9accd5\") " Nov 25 10:43:30 crc kubenswrapper[4821]: I1125 10:43:30.196473 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/88725123-3da9-4127-9506-80c90c9accd5-console-config\") pod \"88725123-3da9-4127-9506-80c90c9accd5\" (UID: \"88725123-3da9-4127-9506-80c90c9accd5\") " Nov 25 10:43:30 crc kubenswrapper[4821]: I1125 10:43:30.196506 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/88725123-3da9-4127-9506-80c90c9accd5-oauth-serving-cert\") pod \"88725123-3da9-4127-9506-80c90c9accd5\" (UID: \"88725123-3da9-4127-9506-80c90c9accd5\") " Nov 25 10:43:30 crc kubenswrapper[4821]: I1125 10:43:30.196558 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/88725123-3da9-4127-9506-80c90c9accd5-service-ca\") pod \"88725123-3da9-4127-9506-80c90c9accd5\" (UID: \"88725123-3da9-4127-9506-80c90c9accd5\") " Nov 25 10:43:30 crc kubenswrapper[4821]: I1125 10:43:30.196623 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/88725123-3da9-4127-9506-80c90c9accd5-console-oauth-config\") pod \"88725123-3da9-4127-9506-80c90c9accd5\" (UID: \"88725123-3da9-4127-9506-80c90c9accd5\") " Nov 25 10:43:30 crc kubenswrapper[4821]: I1125 10:43:30.196684 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/88725123-3da9-4127-9506-80c90c9accd5-console-serving-cert\") pod \"88725123-3da9-4127-9506-80c90c9accd5\" (UID: \"88725123-3da9-4127-9506-80c90c9accd5\") " Nov 25 10:43:30 crc kubenswrapper[4821]: I1125 10:43:30.197372 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/88725123-3da9-4127-9506-80c90c9accd5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "88725123-3da9-4127-9506-80c90c9accd5" (UID: "88725123-3da9-4127-9506-80c90c9accd5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:43:30 crc kubenswrapper[4821]: I1125 10:43:30.197407 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/88725123-3da9-4127-9506-80c90c9accd5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "88725123-3da9-4127-9506-80c90c9accd5" (UID: "88725123-3da9-4127-9506-80c90c9accd5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:43:30 crc kubenswrapper[4821]: I1125 10:43:30.197671 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/88725123-3da9-4127-9506-80c90c9accd5-service-ca" (OuterVolumeSpecName: "service-ca") pod "88725123-3da9-4127-9506-80c90c9accd5" (UID: "88725123-3da9-4127-9506-80c90c9accd5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:43:30 crc kubenswrapper[4821]: I1125 10:43:30.198001 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/88725123-3da9-4127-9506-80c90c9accd5-console-config" (OuterVolumeSpecName: "console-config") pod "88725123-3da9-4127-9506-80c90c9accd5" (UID: "88725123-3da9-4127-9506-80c90c9accd5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:43:30 crc kubenswrapper[4821]: I1125 10:43:30.202799 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88725123-3da9-4127-9506-80c90c9accd5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "88725123-3da9-4127-9506-80c90c9accd5" (UID: "88725123-3da9-4127-9506-80c90c9accd5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:43:30 crc kubenswrapper[4821]: I1125 10:43:30.203390 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88725123-3da9-4127-9506-80c90c9accd5-kube-api-access-5x4xr" (OuterVolumeSpecName: "kube-api-access-5x4xr") pod "88725123-3da9-4127-9506-80c90c9accd5" (UID: "88725123-3da9-4127-9506-80c90c9accd5"). InnerVolumeSpecName "kube-api-access-5x4xr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:43:30 crc kubenswrapper[4821]: I1125 10:43:30.204302 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88725123-3da9-4127-9506-80c90c9accd5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "88725123-3da9-4127-9506-80c90c9accd5" (UID: "88725123-3da9-4127-9506-80c90c9accd5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:43:30 crc kubenswrapper[4821]: I1125 10:43:30.298577 4821 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/88725123-3da9-4127-9506-80c90c9accd5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:43:30 crc kubenswrapper[4821]: I1125 10:43:30.298621 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5x4xr\" (UniqueName: \"kubernetes.io/projected/88725123-3da9-4127-9506-80c90c9accd5-kube-api-access-5x4xr\") on node \"crc\" DevicePath \"\"" Nov 25 10:43:30 crc kubenswrapper[4821]: I1125 10:43:30.298633 4821 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/88725123-3da9-4127-9506-80c90c9accd5-console-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:43:30 crc kubenswrapper[4821]: I1125 10:43:30.298641 4821 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/88725123-3da9-4127-9506-80c90c9accd5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:43:30 crc kubenswrapper[4821]: I1125 10:43:30.298651 4821 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/88725123-3da9-4127-9506-80c90c9accd5-service-ca\") on node \"crc\" DevicePath \"\"" Nov 25 10:43:30 crc kubenswrapper[4821]: I1125 10:43:30.298661 4821 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/88725123-3da9-4127-9506-80c90c9accd5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:43:30 crc kubenswrapper[4821]: I1125 10:43:30.298669 4821 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/88725123-3da9-4127-9506-80c90c9accd5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:43:30 crc kubenswrapper[4821]: I1125 10:43:30.753659 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-hf49d_88725123-3da9-4127-9506-80c90c9accd5/console/0.log" Nov 25 10:43:30 crc kubenswrapper[4821]: I1125 10:43:30.753733 4821 generic.go:334] "Generic (PLEG): container finished" podID="88725123-3da9-4127-9506-80c90c9accd5" containerID="3dd116757d3852f02a7dd5f0a100fafa105ad3f81d3d6be8c6c22129dab37315" exitCode=2 Nov 25 10:43:30 crc kubenswrapper[4821]: I1125 10:43:30.753778 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-hf49d" event={"ID":"88725123-3da9-4127-9506-80c90c9accd5","Type":"ContainerDied","Data":"3dd116757d3852f02a7dd5f0a100fafa105ad3f81d3d6be8c6c22129dab37315"} Nov 25 10:43:30 crc kubenswrapper[4821]: I1125 10:43:30.753822 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-hf49d" event={"ID":"88725123-3da9-4127-9506-80c90c9accd5","Type":"ContainerDied","Data":"fa6ed8bb408c16d09537a52f993d7362b7b95f6699f1285cbfb2a8959b734528"} Nov 25 10:43:30 crc kubenswrapper[4821]: I1125 10:43:30.753846 4821 scope.go:117] "RemoveContainer" containerID="3dd116757d3852f02a7dd5f0a100fafa105ad3f81d3d6be8c6c22129dab37315" Nov 25 10:43:30 crc kubenswrapper[4821]: I1125 10:43:30.753868 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-hf49d" Nov 25 10:43:30 crc kubenswrapper[4821]: I1125 10:43:30.778005 4821 scope.go:117] "RemoveContainer" containerID="3dd116757d3852f02a7dd5f0a100fafa105ad3f81d3d6be8c6c22129dab37315" Nov 25 10:43:30 crc kubenswrapper[4821]: E1125 10:43:30.778450 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3dd116757d3852f02a7dd5f0a100fafa105ad3f81d3d6be8c6c22129dab37315\": container with ID starting with 3dd116757d3852f02a7dd5f0a100fafa105ad3f81d3d6be8c6c22129dab37315 not found: ID does not exist" containerID="3dd116757d3852f02a7dd5f0a100fafa105ad3f81d3d6be8c6c22129dab37315" Nov 25 10:43:30 crc kubenswrapper[4821]: I1125 10:43:30.778497 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3dd116757d3852f02a7dd5f0a100fafa105ad3f81d3d6be8c6c22129dab37315"} err="failed to get container status \"3dd116757d3852f02a7dd5f0a100fafa105ad3f81d3d6be8c6c22129dab37315\": rpc error: code = NotFound desc = could not find container \"3dd116757d3852f02a7dd5f0a100fafa105ad3f81d3d6be8c6c22129dab37315\": container with ID starting with 3dd116757d3852f02a7dd5f0a100fafa105ad3f81d3d6be8c6c22129dab37315 not found: ID does not exist" Nov 25 10:43:30 crc kubenswrapper[4821]: I1125 10:43:30.787879 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-hf49d"] Nov 25 10:43:30 crc kubenswrapper[4821]: I1125 10:43:30.791415 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-hf49d"] Nov 25 10:43:31 crc kubenswrapper[4821]: I1125 10:43:31.762120 4821 generic.go:334] "Generic (PLEG): container finished" podID="8c67f775-e13f-4c7d-9d2b-115d492961f1" containerID="843611cc091cdb7cb6cd67e3196cb5d12e7faef69d76110b414818aab9a88c37" exitCode=0 Nov 25 10:43:31 crc kubenswrapper[4821]: I1125 10:43:31.762230 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6hsmqh" event={"ID":"8c67f775-e13f-4c7d-9d2b-115d492961f1","Type":"ContainerDied","Data":"843611cc091cdb7cb6cd67e3196cb5d12e7faef69d76110b414818aab9a88c37"} Nov 25 10:43:32 crc kubenswrapper[4821]: I1125 10:43:32.123515 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88725123-3da9-4127-9506-80c90c9accd5" path="/var/lib/kubelet/pods/88725123-3da9-4127-9506-80c90c9accd5/volumes" Nov 25 10:43:32 crc kubenswrapper[4821]: I1125 10:43:32.771374 4821 generic.go:334] "Generic (PLEG): container finished" podID="8c67f775-e13f-4c7d-9d2b-115d492961f1" containerID="9f6520ad45d8656cde300e3c71da6dcc8f762d76dc8d384d4eb0e1ca37b1e9bc" exitCode=0 Nov 25 10:43:32 crc kubenswrapper[4821]: I1125 10:43:32.772180 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6hsmqh" event={"ID":"8c67f775-e13f-4c7d-9d2b-115d492961f1","Type":"ContainerDied","Data":"9f6520ad45d8656cde300e3c71da6dcc8f762d76dc8d384d4eb0e1ca37b1e9bc"} Nov 25 10:43:33 crc kubenswrapper[4821]: I1125 10:43:33.993562 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6hsmqh" Nov 25 10:43:34 crc kubenswrapper[4821]: I1125 10:43:34.142495 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8c67f775-e13f-4c7d-9d2b-115d492961f1-bundle\") pod \"8c67f775-e13f-4c7d-9d2b-115d492961f1\" (UID: \"8c67f775-e13f-4c7d-9d2b-115d492961f1\") " Nov 25 10:43:34 crc kubenswrapper[4821]: I1125 10:43:34.142587 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-txlv6\" (UniqueName: \"kubernetes.io/projected/8c67f775-e13f-4c7d-9d2b-115d492961f1-kube-api-access-txlv6\") pod \"8c67f775-e13f-4c7d-9d2b-115d492961f1\" (UID: \"8c67f775-e13f-4c7d-9d2b-115d492961f1\") " Nov 25 10:43:34 crc kubenswrapper[4821]: I1125 10:43:34.142656 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8c67f775-e13f-4c7d-9d2b-115d492961f1-util\") pod \"8c67f775-e13f-4c7d-9d2b-115d492961f1\" (UID: \"8c67f775-e13f-4c7d-9d2b-115d492961f1\") " Nov 25 10:43:34 crc kubenswrapper[4821]: I1125 10:43:34.143617 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c67f775-e13f-4c7d-9d2b-115d492961f1-bundle" (OuterVolumeSpecName: "bundle") pod "8c67f775-e13f-4c7d-9d2b-115d492961f1" (UID: "8c67f775-e13f-4c7d-9d2b-115d492961f1"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:43:34 crc kubenswrapper[4821]: I1125 10:43:34.149505 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c67f775-e13f-4c7d-9d2b-115d492961f1-kube-api-access-txlv6" (OuterVolumeSpecName: "kube-api-access-txlv6") pod "8c67f775-e13f-4c7d-9d2b-115d492961f1" (UID: "8c67f775-e13f-4c7d-9d2b-115d492961f1"). InnerVolumeSpecName "kube-api-access-txlv6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:43:34 crc kubenswrapper[4821]: I1125 10:43:34.156586 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c67f775-e13f-4c7d-9d2b-115d492961f1-util" (OuterVolumeSpecName: "util") pod "8c67f775-e13f-4c7d-9d2b-115d492961f1" (UID: "8c67f775-e13f-4c7d-9d2b-115d492961f1"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:43:34 crc kubenswrapper[4821]: I1125 10:43:34.243780 4821 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8c67f775-e13f-4c7d-9d2b-115d492961f1-util\") on node \"crc\" DevicePath \"\"" Nov 25 10:43:34 crc kubenswrapper[4821]: I1125 10:43:34.243812 4821 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8c67f775-e13f-4c7d-9d2b-115d492961f1-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:43:34 crc kubenswrapper[4821]: I1125 10:43:34.243824 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-txlv6\" (UniqueName: \"kubernetes.io/projected/8c67f775-e13f-4c7d-9d2b-115d492961f1-kube-api-access-txlv6\") on node \"crc\" DevicePath \"\"" Nov 25 10:43:34 crc kubenswrapper[4821]: I1125 10:43:34.782891 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6hsmqh" event={"ID":"8c67f775-e13f-4c7d-9d2b-115d492961f1","Type":"ContainerDied","Data":"a3ddf997e3bf3b7dddfd1a50983a8a698ce9273ab4e3c2f4a4e425c4a14564f3"} Nov 25 10:43:34 crc kubenswrapper[4821]: I1125 10:43:34.783314 4821 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a3ddf997e3bf3b7dddfd1a50983a8a698ce9273ab4e3c2f4a4e425c4a14564f3" Nov 25 10:43:34 crc kubenswrapper[4821]: I1125 10:43:34.783000 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6hsmqh" Nov 25 10:43:42 crc kubenswrapper[4821]: I1125 10:43:42.572317 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-644d88f8f5-gtq55"] Nov 25 10:43:42 crc kubenswrapper[4821]: E1125 10:43:42.573175 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c67f775-e13f-4c7d-9d2b-115d492961f1" containerName="util" Nov 25 10:43:42 crc kubenswrapper[4821]: I1125 10:43:42.573191 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c67f775-e13f-4c7d-9d2b-115d492961f1" containerName="util" Nov 25 10:43:42 crc kubenswrapper[4821]: E1125 10:43:42.573205 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c67f775-e13f-4c7d-9d2b-115d492961f1" containerName="pull" Nov 25 10:43:42 crc kubenswrapper[4821]: I1125 10:43:42.573213 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c67f775-e13f-4c7d-9d2b-115d492961f1" containerName="pull" Nov 25 10:43:42 crc kubenswrapper[4821]: E1125 10:43:42.573225 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88725123-3da9-4127-9506-80c90c9accd5" containerName="console" Nov 25 10:43:42 crc kubenswrapper[4821]: I1125 10:43:42.573235 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="88725123-3da9-4127-9506-80c90c9accd5" containerName="console" Nov 25 10:43:42 crc kubenswrapper[4821]: E1125 10:43:42.573247 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c67f775-e13f-4c7d-9d2b-115d492961f1" containerName="extract" Nov 25 10:43:42 crc kubenswrapper[4821]: I1125 10:43:42.573256 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c67f775-e13f-4c7d-9d2b-115d492961f1" containerName="extract" Nov 25 10:43:42 crc kubenswrapper[4821]: I1125 10:43:42.573377 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="88725123-3da9-4127-9506-80c90c9accd5" containerName="console" Nov 25 10:43:42 crc kubenswrapper[4821]: I1125 10:43:42.573397 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c67f775-e13f-4c7d-9d2b-115d492961f1" containerName="extract" Nov 25 10:43:42 crc kubenswrapper[4821]: I1125 10:43:42.573858 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-644d88f8f5-gtq55" Nov 25 10:43:42 crc kubenswrapper[4821]: I1125 10:43:42.575632 4821 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Nov 25 10:43:42 crc kubenswrapper[4821]: I1125 10:43:42.575842 4821 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-8qsb6" Nov 25 10:43:42 crc kubenswrapper[4821]: I1125 10:43:42.577536 4821 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Nov 25 10:43:42 crc kubenswrapper[4821]: I1125 10:43:42.577605 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Nov 25 10:43:42 crc kubenswrapper[4821]: I1125 10:43:42.581297 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Nov 25 10:43:42 crc kubenswrapper[4821]: I1125 10:43:42.594719 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-644d88f8f5-gtq55"] Nov 25 10:43:42 crc kubenswrapper[4821]: I1125 10:43:42.750262 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b1969fcd-bedc-42dc-966d-9cd5b8770fd9-apiservice-cert\") pod \"metallb-operator-controller-manager-644d88f8f5-gtq55\" (UID: \"b1969fcd-bedc-42dc-966d-9cd5b8770fd9\") " pod="metallb-system/metallb-operator-controller-manager-644d88f8f5-gtq55" Nov 25 10:43:42 crc kubenswrapper[4821]: I1125 10:43:42.750334 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b1969fcd-bedc-42dc-966d-9cd5b8770fd9-webhook-cert\") pod \"metallb-operator-controller-manager-644d88f8f5-gtq55\" (UID: \"b1969fcd-bedc-42dc-966d-9cd5b8770fd9\") " pod="metallb-system/metallb-operator-controller-manager-644d88f8f5-gtq55" Nov 25 10:43:42 crc kubenswrapper[4821]: I1125 10:43:42.750439 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmkqf\" (UniqueName: \"kubernetes.io/projected/b1969fcd-bedc-42dc-966d-9cd5b8770fd9-kube-api-access-qmkqf\") pod \"metallb-operator-controller-manager-644d88f8f5-gtq55\" (UID: \"b1969fcd-bedc-42dc-966d-9cd5b8770fd9\") " pod="metallb-system/metallb-operator-controller-manager-644d88f8f5-gtq55" Nov 25 10:43:42 crc kubenswrapper[4821]: I1125 10:43:42.851667 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmkqf\" (UniqueName: \"kubernetes.io/projected/b1969fcd-bedc-42dc-966d-9cd5b8770fd9-kube-api-access-qmkqf\") pod \"metallb-operator-controller-manager-644d88f8f5-gtq55\" (UID: \"b1969fcd-bedc-42dc-966d-9cd5b8770fd9\") " pod="metallb-system/metallb-operator-controller-manager-644d88f8f5-gtq55" Nov 25 10:43:42 crc kubenswrapper[4821]: I1125 10:43:42.851986 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b1969fcd-bedc-42dc-966d-9cd5b8770fd9-apiservice-cert\") pod \"metallb-operator-controller-manager-644d88f8f5-gtq55\" (UID: \"b1969fcd-bedc-42dc-966d-9cd5b8770fd9\") " pod="metallb-system/metallb-operator-controller-manager-644d88f8f5-gtq55" Nov 25 10:43:42 crc kubenswrapper[4821]: I1125 10:43:42.852019 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b1969fcd-bedc-42dc-966d-9cd5b8770fd9-webhook-cert\") pod \"metallb-operator-controller-manager-644d88f8f5-gtq55\" (UID: \"b1969fcd-bedc-42dc-966d-9cd5b8770fd9\") " pod="metallb-system/metallb-operator-controller-manager-644d88f8f5-gtq55" Nov 25 10:43:42 crc kubenswrapper[4821]: I1125 10:43:42.857759 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b1969fcd-bedc-42dc-966d-9cd5b8770fd9-webhook-cert\") pod \"metallb-operator-controller-manager-644d88f8f5-gtq55\" (UID: \"b1969fcd-bedc-42dc-966d-9cd5b8770fd9\") " pod="metallb-system/metallb-operator-controller-manager-644d88f8f5-gtq55" Nov 25 10:43:42 crc kubenswrapper[4821]: I1125 10:43:42.860544 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b1969fcd-bedc-42dc-966d-9cd5b8770fd9-apiservice-cert\") pod \"metallb-operator-controller-manager-644d88f8f5-gtq55\" (UID: \"b1969fcd-bedc-42dc-966d-9cd5b8770fd9\") " pod="metallb-system/metallb-operator-controller-manager-644d88f8f5-gtq55" Nov 25 10:43:42 crc kubenswrapper[4821]: I1125 10:43:42.870640 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmkqf\" (UniqueName: \"kubernetes.io/projected/b1969fcd-bedc-42dc-966d-9cd5b8770fd9-kube-api-access-qmkqf\") pod \"metallb-operator-controller-manager-644d88f8f5-gtq55\" (UID: \"b1969fcd-bedc-42dc-966d-9cd5b8770fd9\") " pod="metallb-system/metallb-operator-controller-manager-644d88f8f5-gtq55" Nov 25 10:43:42 crc kubenswrapper[4821]: I1125 10:43:42.891517 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-644d88f8f5-gtq55" Nov 25 10:43:43 crc kubenswrapper[4821]: I1125 10:43:43.108972 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-666cb5dd59-l9nfz"] Nov 25 10:43:43 crc kubenswrapper[4821]: I1125 10:43:43.120934 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-666cb5dd59-l9nfz" Nov 25 10:43:43 crc kubenswrapper[4821]: I1125 10:43:43.131724 4821 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 25 10:43:43 crc kubenswrapper[4821]: I1125 10:43:43.131921 4821 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Nov 25 10:43:43 crc kubenswrapper[4821]: I1125 10:43:43.132038 4821 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-xwcpk" Nov 25 10:43:43 crc kubenswrapper[4821]: I1125 10:43:43.143455 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-666cb5dd59-l9nfz"] Nov 25 10:43:43 crc kubenswrapper[4821]: I1125 10:43:43.258988 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dds85\" (UniqueName: \"kubernetes.io/projected/83c8e5d6-48b9-47ff-8c0a-aad92bbe511a-kube-api-access-dds85\") pod \"metallb-operator-webhook-server-666cb5dd59-l9nfz\" (UID: \"83c8e5d6-48b9-47ff-8c0a-aad92bbe511a\") " pod="metallb-system/metallb-operator-webhook-server-666cb5dd59-l9nfz" Nov 25 10:43:43 crc kubenswrapper[4821]: I1125 10:43:43.259055 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/83c8e5d6-48b9-47ff-8c0a-aad92bbe511a-apiservice-cert\") pod \"metallb-operator-webhook-server-666cb5dd59-l9nfz\" (UID: \"83c8e5d6-48b9-47ff-8c0a-aad92bbe511a\") " pod="metallb-system/metallb-operator-webhook-server-666cb5dd59-l9nfz" Nov 25 10:43:43 crc kubenswrapper[4821]: I1125 10:43:43.259132 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/83c8e5d6-48b9-47ff-8c0a-aad92bbe511a-webhook-cert\") pod \"metallb-operator-webhook-server-666cb5dd59-l9nfz\" (UID: \"83c8e5d6-48b9-47ff-8c0a-aad92bbe511a\") " pod="metallb-system/metallb-operator-webhook-server-666cb5dd59-l9nfz" Nov 25 10:43:43 crc kubenswrapper[4821]: I1125 10:43:43.263597 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-644d88f8f5-gtq55"] Nov 25 10:43:43 crc kubenswrapper[4821]: W1125 10:43:43.272444 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb1969fcd_bedc_42dc_966d_9cd5b8770fd9.slice/crio-7565adef43137cc2c7cc8973d7fc3fb45347a18751149fa157623bbe46515302 WatchSource:0}: Error finding container 7565adef43137cc2c7cc8973d7fc3fb45347a18751149fa157623bbe46515302: Status 404 returned error can't find the container with id 7565adef43137cc2c7cc8973d7fc3fb45347a18751149fa157623bbe46515302 Nov 25 10:43:43 crc kubenswrapper[4821]: I1125 10:43:43.360352 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/83c8e5d6-48b9-47ff-8c0a-aad92bbe511a-apiservice-cert\") pod \"metallb-operator-webhook-server-666cb5dd59-l9nfz\" (UID: \"83c8e5d6-48b9-47ff-8c0a-aad92bbe511a\") " pod="metallb-system/metallb-operator-webhook-server-666cb5dd59-l9nfz" Nov 25 10:43:43 crc kubenswrapper[4821]: I1125 10:43:43.360405 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/83c8e5d6-48b9-47ff-8c0a-aad92bbe511a-webhook-cert\") pod \"metallb-operator-webhook-server-666cb5dd59-l9nfz\" (UID: \"83c8e5d6-48b9-47ff-8c0a-aad92bbe511a\") " pod="metallb-system/metallb-operator-webhook-server-666cb5dd59-l9nfz" Nov 25 10:43:43 crc kubenswrapper[4821]: I1125 10:43:43.360475 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dds85\" (UniqueName: \"kubernetes.io/projected/83c8e5d6-48b9-47ff-8c0a-aad92bbe511a-kube-api-access-dds85\") pod \"metallb-operator-webhook-server-666cb5dd59-l9nfz\" (UID: \"83c8e5d6-48b9-47ff-8c0a-aad92bbe511a\") " pod="metallb-system/metallb-operator-webhook-server-666cb5dd59-l9nfz" Nov 25 10:43:43 crc kubenswrapper[4821]: I1125 10:43:43.366304 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/83c8e5d6-48b9-47ff-8c0a-aad92bbe511a-webhook-cert\") pod \"metallb-operator-webhook-server-666cb5dd59-l9nfz\" (UID: \"83c8e5d6-48b9-47ff-8c0a-aad92bbe511a\") " pod="metallb-system/metallb-operator-webhook-server-666cb5dd59-l9nfz" Nov 25 10:43:43 crc kubenswrapper[4821]: I1125 10:43:43.366703 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/83c8e5d6-48b9-47ff-8c0a-aad92bbe511a-apiservice-cert\") pod \"metallb-operator-webhook-server-666cb5dd59-l9nfz\" (UID: \"83c8e5d6-48b9-47ff-8c0a-aad92bbe511a\") " pod="metallb-system/metallb-operator-webhook-server-666cb5dd59-l9nfz" Nov 25 10:43:43 crc kubenswrapper[4821]: I1125 10:43:43.377127 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dds85\" (UniqueName: \"kubernetes.io/projected/83c8e5d6-48b9-47ff-8c0a-aad92bbe511a-kube-api-access-dds85\") pod \"metallb-operator-webhook-server-666cb5dd59-l9nfz\" (UID: \"83c8e5d6-48b9-47ff-8c0a-aad92bbe511a\") " pod="metallb-system/metallb-operator-webhook-server-666cb5dd59-l9nfz" Nov 25 10:43:43 crc kubenswrapper[4821]: I1125 10:43:43.462068 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-666cb5dd59-l9nfz" Nov 25 10:43:43 crc kubenswrapper[4821]: I1125 10:43:43.673093 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-666cb5dd59-l9nfz"] Nov 25 10:43:43 crc kubenswrapper[4821]: I1125 10:43:43.827658 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-666cb5dd59-l9nfz" event={"ID":"83c8e5d6-48b9-47ff-8c0a-aad92bbe511a","Type":"ContainerStarted","Data":"e3cb547ab1220da5e937495bdf8048490cef6eb3c712c0cf712ad04c9eb8ad3f"} Nov 25 10:43:43 crc kubenswrapper[4821]: I1125 10:43:43.829904 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-644d88f8f5-gtq55" event={"ID":"b1969fcd-bedc-42dc-966d-9cd5b8770fd9","Type":"ContainerStarted","Data":"7565adef43137cc2c7cc8973d7fc3fb45347a18751149fa157623bbe46515302"} Nov 25 10:43:46 crc kubenswrapper[4821]: I1125 10:43:46.846302 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-644d88f8f5-gtq55" event={"ID":"b1969fcd-bedc-42dc-966d-9cd5b8770fd9","Type":"ContainerStarted","Data":"1406b8cf86bd1ec02c599fc6edb379834a924b2f17ddd94bc860eea242eb5c4c"} Nov 25 10:43:46 crc kubenswrapper[4821]: I1125 10:43:46.846791 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-644d88f8f5-gtq55" Nov 25 10:43:48 crc kubenswrapper[4821]: I1125 10:43:48.856761 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-666cb5dd59-l9nfz" event={"ID":"83c8e5d6-48b9-47ff-8c0a-aad92bbe511a","Type":"ContainerStarted","Data":"2623f30f90834f348d72ad183e4008d5e45a0a704020779bbf35375db72c8d89"} Nov 25 10:43:48 crc kubenswrapper[4821]: I1125 10:43:48.857197 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-666cb5dd59-l9nfz" Nov 25 10:43:48 crc kubenswrapper[4821]: I1125 10:43:48.871876 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-666cb5dd59-l9nfz" podStartSLOduration=1.617318285 podStartE2EDuration="5.871858088s" podCreationTimestamp="2025-11-25 10:43:43 +0000 UTC" firstStartedPulling="2025-11-25 10:43:43.688082734 +0000 UTC m=+694.224402581" lastFinishedPulling="2025-11-25 10:43:47.942622537 +0000 UTC m=+698.478942384" observedRunningTime="2025-11-25 10:43:48.871309255 +0000 UTC m=+699.407629112" watchObservedRunningTime="2025-11-25 10:43:48.871858088 +0000 UTC m=+699.408177935" Nov 25 10:43:48 crc kubenswrapper[4821]: I1125 10:43:48.874395 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-644d88f8f5-gtq55" podStartSLOduration=4.161128668 podStartE2EDuration="6.874383263s" podCreationTimestamp="2025-11-25 10:43:42 +0000 UTC" firstStartedPulling="2025-11-25 10:43:43.273691442 +0000 UTC m=+693.810011289" lastFinishedPulling="2025-11-25 10:43:45.986946037 +0000 UTC m=+696.523265884" observedRunningTime="2025-11-25 10:43:46.870429494 +0000 UTC m=+697.406749371" watchObservedRunningTime="2025-11-25 10:43:48.874383263 +0000 UTC m=+699.410703110" Nov 25 10:44:03 crc kubenswrapper[4821]: I1125 10:44:03.467515 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-666cb5dd59-l9nfz" Nov 25 10:44:22 crc kubenswrapper[4821]: I1125 10:44:22.895439 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-644d88f8f5-gtq55" Nov 25 10:44:23 crc kubenswrapper[4821]: I1125 10:44:23.687278 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-7knkm"] Nov 25 10:44:23 crc kubenswrapper[4821]: I1125 10:44:23.691060 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-7knkm" Nov 25 10:44:23 crc kubenswrapper[4821]: I1125 10:44:23.693230 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Nov 25 10:44:23 crc kubenswrapper[4821]: I1125 10:44:23.693740 4821 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Nov 25 10:44:23 crc kubenswrapper[4821]: I1125 10:44:23.694512 4821 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-vpmlw" Nov 25 10:44:23 crc kubenswrapper[4821]: I1125 10:44:23.698418 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-bclxv"] Nov 25 10:44:23 crc kubenswrapper[4821]: I1125 10:44:23.699085 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-6998585d5-bclxv" Nov 25 10:44:23 crc kubenswrapper[4821]: I1125 10:44:23.701136 4821 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Nov 25 10:44:23 crc kubenswrapper[4821]: I1125 10:44:23.713372 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-bclxv"] Nov 25 10:44:23 crc kubenswrapper[4821]: I1125 10:44:23.774312 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/0a268005-d45e-4cb1-b04a-5bedb97ea5ee-frr-conf\") pod \"frr-k8s-7knkm\" (UID: \"0a268005-d45e-4cb1-b04a-5bedb97ea5ee\") " pod="metallb-system/frr-k8s-7knkm" Nov 25 10:44:23 crc kubenswrapper[4821]: I1125 10:44:23.774352 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2fbs\" (UniqueName: \"kubernetes.io/projected/1144ecf2-7246-4c05-a7ca-f7b7213092fd-kube-api-access-p2fbs\") pod \"frr-k8s-webhook-server-6998585d5-bclxv\" (UID: \"1144ecf2-7246-4c05-a7ca-f7b7213092fd\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-bclxv" Nov 25 10:44:23 crc kubenswrapper[4821]: I1125 10:44:23.774374 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/0a268005-d45e-4cb1-b04a-5bedb97ea5ee-frr-startup\") pod \"frr-k8s-7knkm\" (UID: \"0a268005-d45e-4cb1-b04a-5bedb97ea5ee\") " pod="metallb-system/frr-k8s-7knkm" Nov 25 10:44:23 crc kubenswrapper[4821]: I1125 10:44:23.774390 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1144ecf2-7246-4c05-a7ca-f7b7213092fd-cert\") pod \"frr-k8s-webhook-server-6998585d5-bclxv\" (UID: \"1144ecf2-7246-4c05-a7ca-f7b7213092fd\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-bclxv" Nov 25 10:44:23 crc kubenswrapper[4821]: I1125 10:44:23.774408 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0a268005-d45e-4cb1-b04a-5bedb97ea5ee-metrics-certs\") pod \"frr-k8s-7knkm\" (UID: \"0a268005-d45e-4cb1-b04a-5bedb97ea5ee\") " pod="metallb-system/frr-k8s-7knkm" Nov 25 10:44:23 crc kubenswrapper[4821]: I1125 10:44:23.774438 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/0a268005-d45e-4cb1-b04a-5bedb97ea5ee-reloader\") pod \"frr-k8s-7knkm\" (UID: \"0a268005-d45e-4cb1-b04a-5bedb97ea5ee\") " pod="metallb-system/frr-k8s-7knkm" Nov 25 10:44:23 crc kubenswrapper[4821]: I1125 10:44:23.774666 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/0a268005-d45e-4cb1-b04a-5bedb97ea5ee-frr-sockets\") pod \"frr-k8s-7knkm\" (UID: \"0a268005-d45e-4cb1-b04a-5bedb97ea5ee\") " pod="metallb-system/frr-k8s-7knkm" Nov 25 10:44:23 crc kubenswrapper[4821]: I1125 10:44:23.774765 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/0a268005-d45e-4cb1-b04a-5bedb97ea5ee-metrics\") pod \"frr-k8s-7knkm\" (UID: \"0a268005-d45e-4cb1-b04a-5bedb97ea5ee\") " pod="metallb-system/frr-k8s-7knkm" Nov 25 10:44:23 crc kubenswrapper[4821]: I1125 10:44:23.774798 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjjtv\" (UniqueName: \"kubernetes.io/projected/0a268005-d45e-4cb1-b04a-5bedb97ea5ee-kube-api-access-pjjtv\") pod \"frr-k8s-7knkm\" (UID: \"0a268005-d45e-4cb1-b04a-5bedb97ea5ee\") " pod="metallb-system/frr-k8s-7knkm" Nov 25 10:44:23 crc kubenswrapper[4821]: I1125 10:44:23.785771 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-5qbb4"] Nov 25 10:44:23 crc kubenswrapper[4821]: I1125 10:44:23.787077 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-5qbb4" Nov 25 10:44:23 crc kubenswrapper[4821]: I1125 10:44:23.789373 4821 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Nov 25 10:44:23 crc kubenswrapper[4821]: I1125 10:44:23.789621 4821 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Nov 25 10:44:23 crc kubenswrapper[4821]: I1125 10:44:23.789662 4821 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-xkp2x" Nov 25 10:44:23 crc kubenswrapper[4821]: I1125 10:44:23.789773 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Nov 25 10:44:23 crc kubenswrapper[4821]: I1125 10:44:23.811519 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-6c7b4b5f48-5knnl"] Nov 25 10:44:23 crc kubenswrapper[4821]: I1125 10:44:23.812667 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6c7b4b5f48-5knnl" Nov 25 10:44:23 crc kubenswrapper[4821]: I1125 10:44:23.814722 4821 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Nov 25 10:44:23 crc kubenswrapper[4821]: I1125 10:44:23.828684 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6c7b4b5f48-5knnl"] Nov 25 10:44:23 crc kubenswrapper[4821]: I1125 10:44:23.876343 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0a268005-d45e-4cb1-b04a-5bedb97ea5ee-metrics-certs\") pod \"frr-k8s-7knkm\" (UID: \"0a268005-d45e-4cb1-b04a-5bedb97ea5ee\") " pod="metallb-system/frr-k8s-7knkm" Nov 25 10:44:23 crc kubenswrapper[4821]: I1125 10:44:23.876401 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/0a268005-d45e-4cb1-b04a-5bedb97ea5ee-reloader\") pod \"frr-k8s-7knkm\" (UID: \"0a268005-d45e-4cb1-b04a-5bedb97ea5ee\") " pod="metallb-system/frr-k8s-7knkm" Nov 25 10:44:23 crc kubenswrapper[4821]: I1125 10:44:23.876430 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/0a268005-d45e-4cb1-b04a-5bedb97ea5ee-frr-sockets\") pod \"frr-k8s-7knkm\" (UID: \"0a268005-d45e-4cb1-b04a-5bedb97ea5ee\") " pod="metallb-system/frr-k8s-7knkm" Nov 25 10:44:23 crc kubenswrapper[4821]: I1125 10:44:23.876450 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/0a268005-d45e-4cb1-b04a-5bedb97ea5ee-metrics\") pod \"frr-k8s-7knkm\" (UID: \"0a268005-d45e-4cb1-b04a-5bedb97ea5ee\") " pod="metallb-system/frr-k8s-7knkm" Nov 25 10:44:23 crc kubenswrapper[4821]: I1125 10:44:23.876464 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjjtv\" (UniqueName: \"kubernetes.io/projected/0a268005-d45e-4cb1-b04a-5bedb97ea5ee-kube-api-access-pjjtv\") pod \"frr-k8s-7knkm\" (UID: \"0a268005-d45e-4cb1-b04a-5bedb97ea5ee\") " pod="metallb-system/frr-k8s-7knkm" Nov 25 10:44:23 crc kubenswrapper[4821]: I1125 10:44:23.876496 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/c8b12b33-c9ca-48dd-9f7f-96b5f1c40f86-metallb-excludel2\") pod \"speaker-5qbb4\" (UID: \"c8b12b33-c9ca-48dd-9f7f-96b5f1c40f86\") " pod="metallb-system/speaker-5qbb4" Nov 25 10:44:23 crc kubenswrapper[4821]: E1125 10:44:23.876509 4821 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Nov 25 10:44:23 crc kubenswrapper[4821]: I1125 10:44:23.876516 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/c8b12b33-c9ca-48dd-9f7f-96b5f1c40f86-memberlist\") pod \"speaker-5qbb4\" (UID: \"c8b12b33-c9ca-48dd-9f7f-96b5f1c40f86\") " pod="metallb-system/speaker-5qbb4" Nov 25 10:44:23 crc kubenswrapper[4821]: E1125 10:44:23.876595 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0a268005-d45e-4cb1-b04a-5bedb97ea5ee-metrics-certs podName:0a268005-d45e-4cb1-b04a-5bedb97ea5ee nodeName:}" failed. No retries permitted until 2025-11-25 10:44:24.376572717 +0000 UTC m=+734.912892624 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0a268005-d45e-4cb1-b04a-5bedb97ea5ee-metrics-certs") pod "frr-k8s-7knkm" (UID: "0a268005-d45e-4cb1-b04a-5bedb97ea5ee") : secret "frr-k8s-certs-secret" not found Nov 25 10:44:23 crc kubenswrapper[4821]: I1125 10:44:23.876714 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xx8x5\" (UniqueName: \"kubernetes.io/projected/c8b12b33-c9ca-48dd-9f7f-96b5f1c40f86-kube-api-access-xx8x5\") pod \"speaker-5qbb4\" (UID: \"c8b12b33-c9ca-48dd-9f7f-96b5f1c40f86\") " pod="metallb-system/speaker-5qbb4" Nov 25 10:44:23 crc kubenswrapper[4821]: I1125 10:44:23.876873 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/0a268005-d45e-4cb1-b04a-5bedb97ea5ee-frr-conf\") pod \"frr-k8s-7knkm\" (UID: \"0a268005-d45e-4cb1-b04a-5bedb97ea5ee\") " pod="metallb-system/frr-k8s-7knkm" Nov 25 10:44:23 crc kubenswrapper[4821]: I1125 10:44:23.876901 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p2fbs\" (UniqueName: \"kubernetes.io/projected/1144ecf2-7246-4c05-a7ca-f7b7213092fd-kube-api-access-p2fbs\") pod \"frr-k8s-webhook-server-6998585d5-bclxv\" (UID: \"1144ecf2-7246-4c05-a7ca-f7b7213092fd\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-bclxv" Nov 25 10:44:23 crc kubenswrapper[4821]: I1125 10:44:23.876927 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c8b12b33-c9ca-48dd-9f7f-96b5f1c40f86-metrics-certs\") pod \"speaker-5qbb4\" (UID: \"c8b12b33-c9ca-48dd-9f7f-96b5f1c40f86\") " pod="metallb-system/speaker-5qbb4" Nov 25 10:44:23 crc kubenswrapper[4821]: I1125 10:44:23.876948 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/0a268005-d45e-4cb1-b04a-5bedb97ea5ee-frr-startup\") pod \"frr-k8s-7knkm\" (UID: \"0a268005-d45e-4cb1-b04a-5bedb97ea5ee\") " pod="metallb-system/frr-k8s-7knkm" Nov 25 10:44:23 crc kubenswrapper[4821]: I1125 10:44:23.876969 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1144ecf2-7246-4c05-a7ca-f7b7213092fd-cert\") pod \"frr-k8s-webhook-server-6998585d5-bclxv\" (UID: \"1144ecf2-7246-4c05-a7ca-f7b7213092fd\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-bclxv" Nov 25 10:44:23 crc kubenswrapper[4821]: I1125 10:44:23.876974 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/0a268005-d45e-4cb1-b04a-5bedb97ea5ee-frr-sockets\") pod \"frr-k8s-7knkm\" (UID: \"0a268005-d45e-4cb1-b04a-5bedb97ea5ee\") " pod="metallb-system/frr-k8s-7knkm" Nov 25 10:44:23 crc kubenswrapper[4821]: I1125 10:44:23.877027 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/0a268005-d45e-4cb1-b04a-5bedb97ea5ee-reloader\") pod \"frr-k8s-7knkm\" (UID: \"0a268005-d45e-4cb1-b04a-5bedb97ea5ee\") " pod="metallb-system/frr-k8s-7knkm" Nov 25 10:44:23 crc kubenswrapper[4821]: I1125 10:44:23.877154 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/0a268005-d45e-4cb1-b04a-5bedb97ea5ee-frr-conf\") pod \"frr-k8s-7knkm\" (UID: \"0a268005-d45e-4cb1-b04a-5bedb97ea5ee\") " pod="metallb-system/frr-k8s-7knkm" Nov 25 10:44:23 crc kubenswrapper[4821]: I1125 10:44:23.877981 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/0a268005-d45e-4cb1-b04a-5bedb97ea5ee-frr-startup\") pod \"frr-k8s-7knkm\" (UID: \"0a268005-d45e-4cb1-b04a-5bedb97ea5ee\") " pod="metallb-system/frr-k8s-7knkm" Nov 25 10:44:23 crc kubenswrapper[4821]: I1125 10:44:23.878192 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/0a268005-d45e-4cb1-b04a-5bedb97ea5ee-metrics\") pod \"frr-k8s-7knkm\" (UID: \"0a268005-d45e-4cb1-b04a-5bedb97ea5ee\") " pod="metallb-system/frr-k8s-7knkm" Nov 25 10:44:23 crc kubenswrapper[4821]: I1125 10:44:23.884982 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1144ecf2-7246-4c05-a7ca-f7b7213092fd-cert\") pod \"frr-k8s-webhook-server-6998585d5-bclxv\" (UID: \"1144ecf2-7246-4c05-a7ca-f7b7213092fd\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-bclxv" Nov 25 10:44:23 crc kubenswrapper[4821]: I1125 10:44:23.897740 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjjtv\" (UniqueName: \"kubernetes.io/projected/0a268005-d45e-4cb1-b04a-5bedb97ea5ee-kube-api-access-pjjtv\") pod \"frr-k8s-7knkm\" (UID: \"0a268005-d45e-4cb1-b04a-5bedb97ea5ee\") " pod="metallb-system/frr-k8s-7knkm" Nov 25 10:44:23 crc kubenswrapper[4821]: I1125 10:44:23.898013 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p2fbs\" (UniqueName: \"kubernetes.io/projected/1144ecf2-7246-4c05-a7ca-f7b7213092fd-kube-api-access-p2fbs\") pod \"frr-k8s-webhook-server-6998585d5-bclxv\" (UID: \"1144ecf2-7246-4c05-a7ca-f7b7213092fd\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-bclxv" Nov 25 10:44:23 crc kubenswrapper[4821]: I1125 10:44:23.978576 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5sthc\" (UniqueName: \"kubernetes.io/projected/902ced61-7b2c-4a77-b8fc-ea9a226e05ba-kube-api-access-5sthc\") pod \"controller-6c7b4b5f48-5knnl\" (UID: \"902ced61-7b2c-4a77-b8fc-ea9a226e05ba\") " pod="metallb-system/controller-6c7b4b5f48-5knnl" Nov 25 10:44:23 crc kubenswrapper[4821]: I1125 10:44:23.978637 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/902ced61-7b2c-4a77-b8fc-ea9a226e05ba-cert\") pod \"controller-6c7b4b5f48-5knnl\" (UID: \"902ced61-7b2c-4a77-b8fc-ea9a226e05ba\") " pod="metallb-system/controller-6c7b4b5f48-5knnl" Nov 25 10:44:23 crc kubenswrapper[4821]: I1125 10:44:23.978674 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c8b12b33-c9ca-48dd-9f7f-96b5f1c40f86-metrics-certs\") pod \"speaker-5qbb4\" (UID: \"c8b12b33-c9ca-48dd-9f7f-96b5f1c40f86\") " pod="metallb-system/speaker-5qbb4" Nov 25 10:44:23 crc kubenswrapper[4821]: I1125 10:44:23.978736 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/902ced61-7b2c-4a77-b8fc-ea9a226e05ba-metrics-certs\") pod \"controller-6c7b4b5f48-5knnl\" (UID: \"902ced61-7b2c-4a77-b8fc-ea9a226e05ba\") " pod="metallb-system/controller-6c7b4b5f48-5knnl" Nov 25 10:44:23 crc kubenswrapper[4821]: I1125 10:44:23.978774 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/c8b12b33-c9ca-48dd-9f7f-96b5f1c40f86-metallb-excludel2\") pod \"speaker-5qbb4\" (UID: \"c8b12b33-c9ca-48dd-9f7f-96b5f1c40f86\") " pod="metallb-system/speaker-5qbb4" Nov 25 10:44:23 crc kubenswrapper[4821]: I1125 10:44:23.978794 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/c8b12b33-c9ca-48dd-9f7f-96b5f1c40f86-memberlist\") pod \"speaker-5qbb4\" (UID: \"c8b12b33-c9ca-48dd-9f7f-96b5f1c40f86\") " pod="metallb-system/speaker-5qbb4" Nov 25 10:44:23 crc kubenswrapper[4821]: I1125 10:44:23.978816 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xx8x5\" (UniqueName: \"kubernetes.io/projected/c8b12b33-c9ca-48dd-9f7f-96b5f1c40f86-kube-api-access-xx8x5\") pod \"speaker-5qbb4\" (UID: \"c8b12b33-c9ca-48dd-9f7f-96b5f1c40f86\") " pod="metallb-system/speaker-5qbb4" Nov 25 10:44:23 crc kubenswrapper[4821]: E1125 10:44:23.978905 4821 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Nov 25 10:44:23 crc kubenswrapper[4821]: E1125 10:44:23.978930 4821 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 25 10:44:23 crc kubenswrapper[4821]: E1125 10:44:23.978988 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c8b12b33-c9ca-48dd-9f7f-96b5f1c40f86-metrics-certs podName:c8b12b33-c9ca-48dd-9f7f-96b5f1c40f86 nodeName:}" failed. No retries permitted until 2025-11-25 10:44:24.478965435 +0000 UTC m=+735.015285292 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c8b12b33-c9ca-48dd-9f7f-96b5f1c40f86-metrics-certs") pod "speaker-5qbb4" (UID: "c8b12b33-c9ca-48dd-9f7f-96b5f1c40f86") : secret "speaker-certs-secret" not found Nov 25 10:44:23 crc kubenswrapper[4821]: E1125 10:44:23.979007 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c8b12b33-c9ca-48dd-9f7f-96b5f1c40f86-memberlist podName:c8b12b33-c9ca-48dd-9f7f-96b5f1c40f86 nodeName:}" failed. No retries permitted until 2025-11-25 10:44:24.478998536 +0000 UTC m=+735.015318393 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/c8b12b33-c9ca-48dd-9f7f-96b5f1c40f86-memberlist") pod "speaker-5qbb4" (UID: "c8b12b33-c9ca-48dd-9f7f-96b5f1c40f86") : secret "metallb-memberlist" not found Nov 25 10:44:23 crc kubenswrapper[4821]: I1125 10:44:23.979525 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/c8b12b33-c9ca-48dd-9f7f-96b5f1c40f86-metallb-excludel2\") pod \"speaker-5qbb4\" (UID: \"c8b12b33-c9ca-48dd-9f7f-96b5f1c40f86\") " pod="metallb-system/speaker-5qbb4" Nov 25 10:44:23 crc kubenswrapper[4821]: I1125 10:44:23.994108 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xx8x5\" (UniqueName: \"kubernetes.io/projected/c8b12b33-c9ca-48dd-9f7f-96b5f1c40f86-kube-api-access-xx8x5\") pod \"speaker-5qbb4\" (UID: \"c8b12b33-c9ca-48dd-9f7f-96b5f1c40f86\") " pod="metallb-system/speaker-5qbb4" Nov 25 10:44:24 crc kubenswrapper[4821]: I1125 10:44:24.020591 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-6998585d5-bclxv" Nov 25 10:44:24 crc kubenswrapper[4821]: I1125 10:44:24.080437 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/902ced61-7b2c-4a77-b8fc-ea9a226e05ba-cert\") pod \"controller-6c7b4b5f48-5knnl\" (UID: \"902ced61-7b2c-4a77-b8fc-ea9a226e05ba\") " pod="metallb-system/controller-6c7b4b5f48-5knnl" Nov 25 10:44:24 crc kubenswrapper[4821]: I1125 10:44:24.080563 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/902ced61-7b2c-4a77-b8fc-ea9a226e05ba-metrics-certs\") pod \"controller-6c7b4b5f48-5knnl\" (UID: \"902ced61-7b2c-4a77-b8fc-ea9a226e05ba\") " pod="metallb-system/controller-6c7b4b5f48-5knnl" Nov 25 10:44:24 crc kubenswrapper[4821]: I1125 10:44:24.080652 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5sthc\" (UniqueName: \"kubernetes.io/projected/902ced61-7b2c-4a77-b8fc-ea9a226e05ba-kube-api-access-5sthc\") pod \"controller-6c7b4b5f48-5knnl\" (UID: \"902ced61-7b2c-4a77-b8fc-ea9a226e05ba\") " pod="metallb-system/controller-6c7b4b5f48-5knnl" Nov 25 10:44:24 crc kubenswrapper[4821]: I1125 10:44:24.082832 4821 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 25 10:44:24 crc kubenswrapper[4821]: I1125 10:44:24.083405 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/902ced61-7b2c-4a77-b8fc-ea9a226e05ba-metrics-certs\") pod \"controller-6c7b4b5f48-5knnl\" (UID: \"902ced61-7b2c-4a77-b8fc-ea9a226e05ba\") " pod="metallb-system/controller-6c7b4b5f48-5knnl" Nov 25 10:44:24 crc kubenswrapper[4821]: I1125 10:44:24.095218 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5sthc\" (UniqueName: \"kubernetes.io/projected/902ced61-7b2c-4a77-b8fc-ea9a226e05ba-kube-api-access-5sthc\") pod \"controller-6c7b4b5f48-5knnl\" (UID: \"902ced61-7b2c-4a77-b8fc-ea9a226e05ba\") " pod="metallb-system/controller-6c7b4b5f48-5knnl" Nov 25 10:44:24 crc kubenswrapper[4821]: I1125 10:44:24.095675 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/902ced61-7b2c-4a77-b8fc-ea9a226e05ba-cert\") pod \"controller-6c7b4b5f48-5knnl\" (UID: \"902ced61-7b2c-4a77-b8fc-ea9a226e05ba\") " pod="metallb-system/controller-6c7b4b5f48-5knnl" Nov 25 10:44:24 crc kubenswrapper[4821]: I1125 10:44:24.126762 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6c7b4b5f48-5knnl" Nov 25 10:44:24 crc kubenswrapper[4821]: I1125 10:44:24.211407 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-bclxv"] Nov 25 10:44:24 crc kubenswrapper[4821]: I1125 10:44:24.385853 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0a268005-d45e-4cb1-b04a-5bedb97ea5ee-metrics-certs\") pod \"frr-k8s-7knkm\" (UID: \"0a268005-d45e-4cb1-b04a-5bedb97ea5ee\") " pod="metallb-system/frr-k8s-7knkm" Nov 25 10:44:24 crc kubenswrapper[4821]: I1125 10:44:24.390547 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0a268005-d45e-4cb1-b04a-5bedb97ea5ee-metrics-certs\") pod \"frr-k8s-7knkm\" (UID: \"0a268005-d45e-4cb1-b04a-5bedb97ea5ee\") " pod="metallb-system/frr-k8s-7knkm" Nov 25 10:44:24 crc kubenswrapper[4821]: I1125 10:44:24.422930 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6c7b4b5f48-5knnl"] Nov 25 10:44:24 crc kubenswrapper[4821]: W1125 10:44:24.433120 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod902ced61_7b2c_4a77_b8fc_ea9a226e05ba.slice/crio-d3da678d4ae943124099ca826ac484ddc67d00ffd1ccfc8438c741cf0337486a WatchSource:0}: Error finding container d3da678d4ae943124099ca826ac484ddc67d00ffd1ccfc8438c741cf0337486a: Status 404 returned error can't find the container with id d3da678d4ae943124099ca826ac484ddc67d00ffd1ccfc8438c741cf0337486a Nov 25 10:44:24 crc kubenswrapper[4821]: I1125 10:44:24.487571 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/c8b12b33-c9ca-48dd-9f7f-96b5f1c40f86-memberlist\") pod \"speaker-5qbb4\" (UID: \"c8b12b33-c9ca-48dd-9f7f-96b5f1c40f86\") " pod="metallb-system/speaker-5qbb4" Nov 25 10:44:24 crc kubenswrapper[4821]: I1125 10:44:24.487860 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c8b12b33-c9ca-48dd-9f7f-96b5f1c40f86-metrics-certs\") pod \"speaker-5qbb4\" (UID: \"c8b12b33-c9ca-48dd-9f7f-96b5f1c40f86\") " pod="metallb-system/speaker-5qbb4" Nov 25 10:44:24 crc kubenswrapper[4821]: E1125 10:44:24.487751 4821 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 25 10:44:24 crc kubenswrapper[4821]: E1125 10:44:24.487966 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c8b12b33-c9ca-48dd-9f7f-96b5f1c40f86-memberlist podName:c8b12b33-c9ca-48dd-9f7f-96b5f1c40f86 nodeName:}" failed. No retries permitted until 2025-11-25 10:44:25.487944981 +0000 UTC m=+736.024264828 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/c8b12b33-c9ca-48dd-9f7f-96b5f1c40f86-memberlist") pod "speaker-5qbb4" (UID: "c8b12b33-c9ca-48dd-9f7f-96b5f1c40f86") : secret "metallb-memberlist" not found Nov 25 10:44:24 crc kubenswrapper[4821]: I1125 10:44:24.495699 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c8b12b33-c9ca-48dd-9f7f-96b5f1c40f86-metrics-certs\") pod \"speaker-5qbb4\" (UID: \"c8b12b33-c9ca-48dd-9f7f-96b5f1c40f86\") " pod="metallb-system/speaker-5qbb4" Nov 25 10:44:24 crc kubenswrapper[4821]: I1125 10:44:24.612362 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-7knkm" Nov 25 10:44:25 crc kubenswrapper[4821]: I1125 10:44:25.044585 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-5knnl" event={"ID":"902ced61-7b2c-4a77-b8fc-ea9a226e05ba","Type":"ContainerStarted","Data":"76b54c2f4e1e96faa266560be77b68391546ca7609d8533f14120965bcb4c545"} Nov 25 10:44:25 crc kubenswrapper[4821]: I1125 10:44:25.044631 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-5knnl" event={"ID":"902ced61-7b2c-4a77-b8fc-ea9a226e05ba","Type":"ContainerStarted","Data":"5b8dd28b836de6b1799f12254a81247c59fe4cb83e46e8e662a5b43866f101f3"} Nov 25 10:44:25 crc kubenswrapper[4821]: I1125 10:44:25.044646 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-5knnl" event={"ID":"902ced61-7b2c-4a77-b8fc-ea9a226e05ba","Type":"ContainerStarted","Data":"d3da678d4ae943124099ca826ac484ddc67d00ffd1ccfc8438c741cf0337486a"} Nov 25 10:44:25 crc kubenswrapper[4821]: I1125 10:44:25.044665 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-6c7b4b5f48-5knnl" Nov 25 10:44:25 crc kubenswrapper[4821]: I1125 10:44:25.045866 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-7knkm" event={"ID":"0a268005-d45e-4cb1-b04a-5bedb97ea5ee","Type":"ContainerStarted","Data":"3c8449e334467c9d7408193c0dd034b51ec3ea126aaa8536889fba676b05e60e"} Nov 25 10:44:25 crc kubenswrapper[4821]: I1125 10:44:25.046990 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-6998585d5-bclxv" event={"ID":"1144ecf2-7246-4c05-a7ca-f7b7213092fd","Type":"ContainerStarted","Data":"bb15ca5822e765389f72b914098e59806ee4fc735eab5fc8d849ccda93839336"} Nov 25 10:44:25 crc kubenswrapper[4821]: I1125 10:44:25.065856 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-6c7b4b5f48-5knnl" podStartSLOduration=2.065837092 podStartE2EDuration="2.065837092s" podCreationTimestamp="2025-11-25 10:44:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:44:25.058705918 +0000 UTC m=+735.595025765" watchObservedRunningTime="2025-11-25 10:44:25.065837092 +0000 UTC m=+735.602156939" Nov 25 10:44:25 crc kubenswrapper[4821]: I1125 10:44:25.503020 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/c8b12b33-c9ca-48dd-9f7f-96b5f1c40f86-memberlist\") pod \"speaker-5qbb4\" (UID: \"c8b12b33-c9ca-48dd-9f7f-96b5f1c40f86\") " pod="metallb-system/speaker-5qbb4" Nov 25 10:44:25 crc kubenswrapper[4821]: I1125 10:44:25.510320 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/c8b12b33-c9ca-48dd-9f7f-96b5f1c40f86-memberlist\") pod \"speaker-5qbb4\" (UID: \"c8b12b33-c9ca-48dd-9f7f-96b5f1c40f86\") " pod="metallb-system/speaker-5qbb4" Nov 25 10:44:25 crc kubenswrapper[4821]: I1125 10:44:25.600745 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-5qbb4" Nov 25 10:44:25 crc kubenswrapper[4821]: W1125 10:44:25.648279 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc8b12b33_c9ca_48dd_9f7f_96b5f1c40f86.slice/crio-5aa4d8968be62140cf46d2079acb901342d1881296f0763adfe3f74d8c876231 WatchSource:0}: Error finding container 5aa4d8968be62140cf46d2079acb901342d1881296f0763adfe3f74d8c876231: Status 404 returned error can't find the container with id 5aa4d8968be62140cf46d2079acb901342d1881296f0763adfe3f74d8c876231 Nov 25 10:44:26 crc kubenswrapper[4821]: I1125 10:44:26.055534 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-5qbb4" event={"ID":"c8b12b33-c9ca-48dd-9f7f-96b5f1c40f86","Type":"ContainerStarted","Data":"b1e92f04078b3219293a40f3ffbeb3912702083d4091add53766e36afe71a6f7"} Nov 25 10:44:26 crc kubenswrapper[4821]: I1125 10:44:26.055586 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-5qbb4" event={"ID":"c8b12b33-c9ca-48dd-9f7f-96b5f1c40f86","Type":"ContainerStarted","Data":"5aa4d8968be62140cf46d2079acb901342d1881296f0763adfe3f74d8c876231"} Nov 25 10:44:27 crc kubenswrapper[4821]: I1125 10:44:27.066007 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-5qbb4" event={"ID":"c8b12b33-c9ca-48dd-9f7f-96b5f1c40f86","Type":"ContainerStarted","Data":"be26bec70c76252188668bc31612d9a7ec5df3feb9df86368d6f334ae7e8dc42"} Nov 25 10:44:27 crc kubenswrapper[4821]: I1125 10:44:27.066355 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-5qbb4" Nov 25 10:44:27 crc kubenswrapper[4821]: I1125 10:44:27.085283 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-5qbb4" podStartSLOduration=4.085265319 podStartE2EDuration="4.085265319s" podCreationTimestamp="2025-11-25 10:44:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:44:27.083437452 +0000 UTC m=+737.619757299" watchObservedRunningTime="2025-11-25 10:44:27.085265319 +0000 UTC m=+737.621585166" Nov 25 10:44:32 crc kubenswrapper[4821]: I1125 10:44:32.094048 4821 generic.go:334] "Generic (PLEG): container finished" podID="0a268005-d45e-4cb1-b04a-5bedb97ea5ee" containerID="33271449b964ee06be2c7b774f3fbc32a1b9571d4d8e46558755bae49e87dd5b" exitCode=0 Nov 25 10:44:32 crc kubenswrapper[4821]: I1125 10:44:32.094205 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-7knkm" event={"ID":"0a268005-d45e-4cb1-b04a-5bedb97ea5ee","Type":"ContainerDied","Data":"33271449b964ee06be2c7b774f3fbc32a1b9571d4d8e46558755bae49e87dd5b"} Nov 25 10:44:32 crc kubenswrapper[4821]: I1125 10:44:32.096524 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-6998585d5-bclxv" event={"ID":"1144ecf2-7246-4c05-a7ca-f7b7213092fd","Type":"ContainerStarted","Data":"43806cf1a03dc5af79af612bfc028a43f4dfab1f31523b0ad51e1d3c836c2ef3"} Nov 25 10:44:32 crc kubenswrapper[4821]: I1125 10:44:32.096760 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-6998585d5-bclxv" Nov 25 10:44:32 crc kubenswrapper[4821]: I1125 10:44:32.165307 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-6998585d5-bclxv" podStartSLOduration=2.064326313 podStartE2EDuration="9.165290642s" podCreationTimestamp="2025-11-25 10:44:23 +0000 UTC" firstStartedPulling="2025-11-25 10:44:24.294402524 +0000 UTC m=+734.830722381" lastFinishedPulling="2025-11-25 10:44:31.395366863 +0000 UTC m=+741.931686710" observedRunningTime="2025-11-25 10:44:32.161852304 +0000 UTC m=+742.698172151" watchObservedRunningTime="2025-11-25 10:44:32.165290642 +0000 UTC m=+742.701610479" Nov 25 10:44:33 crc kubenswrapper[4821]: I1125 10:44:33.106918 4821 generic.go:334] "Generic (PLEG): container finished" podID="0a268005-d45e-4cb1-b04a-5bedb97ea5ee" containerID="ce743654d2d906e9487da94f468b24693fc6aa54f78da47d93c923750e549acc" exitCode=0 Nov 25 10:44:33 crc kubenswrapper[4821]: I1125 10:44:33.107701 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-7knkm" event={"ID":"0a268005-d45e-4cb1-b04a-5bedb97ea5ee","Type":"ContainerDied","Data":"ce743654d2d906e9487da94f468b24693fc6aa54f78da47d93c923750e549acc"} Nov 25 10:44:34 crc kubenswrapper[4821]: I1125 10:44:34.120029 4821 generic.go:334] "Generic (PLEG): container finished" podID="0a268005-d45e-4cb1-b04a-5bedb97ea5ee" containerID="1edee5ec4dc9918d64aceab7adeafa08e93afbb12bcf4d56f014606a6239c928" exitCode=0 Nov 25 10:44:34 crc kubenswrapper[4821]: I1125 10:44:34.126605 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-7knkm" event={"ID":"0a268005-d45e-4cb1-b04a-5bedb97ea5ee","Type":"ContainerDied","Data":"1edee5ec4dc9918d64aceab7adeafa08e93afbb12bcf4d56f014606a6239c928"} Nov 25 10:44:34 crc kubenswrapper[4821]: I1125 10:44:34.133776 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-6c7b4b5f48-5knnl" Nov 25 10:44:34 crc kubenswrapper[4821]: I1125 10:44:34.714855 4821 patch_prober.go:28] interesting pod/machine-config-daemon-2krbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:44:34 crc kubenswrapper[4821]: I1125 10:44:34.715416 4821 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:44:35 crc kubenswrapper[4821]: I1125 10:44:35.128272 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-7knkm" event={"ID":"0a268005-d45e-4cb1-b04a-5bedb97ea5ee","Type":"ContainerStarted","Data":"e4aa6d32dbd7993b234c5b3cc53dcd6e40eef8050c638013524fa5d1bf249a1c"} Nov 25 10:44:35 crc kubenswrapper[4821]: I1125 10:44:35.128306 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-7knkm" event={"ID":"0a268005-d45e-4cb1-b04a-5bedb97ea5ee","Type":"ContainerStarted","Data":"4719c7f0d224d7ccb2c48446d3c31e87254cb2d7e7fb67952dc435aeaf4721eb"} Nov 25 10:44:35 crc kubenswrapper[4821]: I1125 10:44:35.128315 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-7knkm" event={"ID":"0a268005-d45e-4cb1-b04a-5bedb97ea5ee","Type":"ContainerStarted","Data":"24e1e23775db95d7a4ee30970c53fe1728c06110a8e0cbfe9f838e68236c3678"} Nov 25 10:44:35 crc kubenswrapper[4821]: I1125 10:44:35.128323 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-7knkm" event={"ID":"0a268005-d45e-4cb1-b04a-5bedb97ea5ee","Type":"ContainerStarted","Data":"3d5765886f554eb817f7aec6bc40731e70d57df8aa6aaec8c5b09ab0d63593a1"} Nov 25 10:44:35 crc kubenswrapper[4821]: I1125 10:44:35.128331 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-7knkm" event={"ID":"0a268005-d45e-4cb1-b04a-5bedb97ea5ee","Type":"ContainerStarted","Data":"ce79fc43fd5aaf7ee621530a9212bbc55d27a4b8fc4bcaaea8a5c1492272da20"} Nov 25 10:44:35 crc kubenswrapper[4821]: I1125 10:44:35.605819 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-5qbb4" Nov 25 10:44:36 crc kubenswrapper[4821]: I1125 10:44:36.137469 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-7knkm" event={"ID":"0a268005-d45e-4cb1-b04a-5bedb97ea5ee","Type":"ContainerStarted","Data":"4da87a2cd14e04135ff28a29677253d2de1bb8261d0de31c6d4940a009aea375"} Nov 25 10:44:36 crc kubenswrapper[4821]: I1125 10:44:36.137664 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-7knkm" Nov 25 10:44:36 crc kubenswrapper[4821]: I1125 10:44:36.159908 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-7knkm" podStartSLOduration=6.498108304 podStartE2EDuration="13.159890647s" podCreationTimestamp="2025-11-25 10:44:23 +0000 UTC" firstStartedPulling="2025-11-25 10:44:24.710458145 +0000 UTC m=+735.246777992" lastFinishedPulling="2025-11-25 10:44:31.372240468 +0000 UTC m=+741.908560335" observedRunningTime="2025-11-25 10:44:36.158114141 +0000 UTC m=+746.694433988" watchObservedRunningTime="2025-11-25 10:44:36.159890647 +0000 UTC m=+746.696210494" Nov 25 10:44:37 crc kubenswrapper[4821]: I1125 10:44:37.450728 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-58htt"] Nov 25 10:44:37 crc kubenswrapper[4821]: I1125 10:44:37.450976 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-58htt" podUID="518db853-3bbe-421b-a2dd-6d1b2e965120" containerName="controller-manager" containerID="cri-o://717e41ba50a71a53a80e70255a9aed3a28af1f2d24eb9efc4567403bbd4385f4" gracePeriod=30 Nov 25 10:44:37 crc kubenswrapper[4821]: I1125 10:44:37.582731 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-dv9nt"] Nov 25 10:44:37 crc kubenswrapper[4821]: I1125 10:44:37.583082 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dv9nt" podUID="ba27af47-1944-4970-a015-e96c7cee4862" containerName="route-controller-manager" containerID="cri-o://afa3b5464872e57b7a8921ecdae57338715ab15ac3e01f3eb570907c98771328" gracePeriod=30 Nov 25 10:44:37 crc kubenswrapper[4821]: I1125 10:44:37.978450 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-58htt" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.042474 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dv9nt" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.070000 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/518db853-3bbe-421b-a2dd-6d1b2e965120-client-ca\") pod \"518db853-3bbe-421b-a2dd-6d1b2e965120\" (UID: \"518db853-3bbe-421b-a2dd-6d1b2e965120\") " Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.070038 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/518db853-3bbe-421b-a2dd-6d1b2e965120-proxy-ca-bundles\") pod \"518db853-3bbe-421b-a2dd-6d1b2e965120\" (UID: \"518db853-3bbe-421b-a2dd-6d1b2e965120\") " Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.070105 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/518db853-3bbe-421b-a2dd-6d1b2e965120-config\") pod \"518db853-3bbe-421b-a2dd-6d1b2e965120\" (UID: \"518db853-3bbe-421b-a2dd-6d1b2e965120\") " Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.070138 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-trdqk\" (UniqueName: \"kubernetes.io/projected/518db853-3bbe-421b-a2dd-6d1b2e965120-kube-api-access-trdqk\") pod \"518db853-3bbe-421b-a2dd-6d1b2e965120\" (UID: \"518db853-3bbe-421b-a2dd-6d1b2e965120\") " Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.070268 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/518db853-3bbe-421b-a2dd-6d1b2e965120-serving-cert\") pod \"518db853-3bbe-421b-a2dd-6d1b2e965120\" (UID: \"518db853-3bbe-421b-a2dd-6d1b2e965120\") " Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.070990 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/518db853-3bbe-421b-a2dd-6d1b2e965120-client-ca" (OuterVolumeSpecName: "client-ca") pod "518db853-3bbe-421b-a2dd-6d1b2e965120" (UID: "518db853-3bbe-421b-a2dd-6d1b2e965120"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.070998 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/518db853-3bbe-421b-a2dd-6d1b2e965120-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "518db853-3bbe-421b-a2dd-6d1b2e965120" (UID: "518db853-3bbe-421b-a2dd-6d1b2e965120"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.071116 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/518db853-3bbe-421b-a2dd-6d1b2e965120-config" (OuterVolumeSpecName: "config") pod "518db853-3bbe-421b-a2dd-6d1b2e965120" (UID: "518db853-3bbe-421b-a2dd-6d1b2e965120"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.078449 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/518db853-3bbe-421b-a2dd-6d1b2e965120-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "518db853-3bbe-421b-a2dd-6d1b2e965120" (UID: "518db853-3bbe-421b-a2dd-6d1b2e965120"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.079517 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/518db853-3bbe-421b-a2dd-6d1b2e965120-kube-api-access-trdqk" (OuterVolumeSpecName: "kube-api-access-trdqk") pod "518db853-3bbe-421b-a2dd-6d1b2e965120" (UID: "518db853-3bbe-421b-a2dd-6d1b2e965120"). InnerVolumeSpecName "kube-api-access-trdqk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.147539 4821 generic.go:334] "Generic (PLEG): container finished" podID="518db853-3bbe-421b-a2dd-6d1b2e965120" containerID="717e41ba50a71a53a80e70255a9aed3a28af1f2d24eb9efc4567403bbd4385f4" exitCode=0 Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.147604 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-58htt" event={"ID":"518db853-3bbe-421b-a2dd-6d1b2e965120","Type":"ContainerDied","Data":"717e41ba50a71a53a80e70255a9aed3a28af1f2d24eb9efc4567403bbd4385f4"} Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.147609 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-58htt" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.147630 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-58htt" event={"ID":"518db853-3bbe-421b-a2dd-6d1b2e965120","Type":"ContainerDied","Data":"903cb8e95459b26e7c1e4806fc6c432eac8bea8cdc1e84013d939af20b308de5"} Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.147646 4821 scope.go:117] "RemoveContainer" containerID="717e41ba50a71a53a80e70255a9aed3a28af1f2d24eb9efc4567403bbd4385f4" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.149440 4821 generic.go:334] "Generic (PLEG): container finished" podID="ba27af47-1944-4970-a015-e96c7cee4862" containerID="afa3b5464872e57b7a8921ecdae57338715ab15ac3e01f3eb570907c98771328" exitCode=0 Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.149472 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dv9nt" event={"ID":"ba27af47-1944-4970-a015-e96c7cee4862","Type":"ContainerDied","Data":"afa3b5464872e57b7a8921ecdae57338715ab15ac3e01f3eb570907c98771328"} Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.149495 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dv9nt" event={"ID":"ba27af47-1944-4970-a015-e96c7cee4862","Type":"ContainerDied","Data":"21ee34f3e4c08e218c5e2422993cd523869563687b4357d81391e29fc85026c3"} Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.149527 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dv9nt" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.164344 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-58htt"] Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.165465 4821 scope.go:117] "RemoveContainer" containerID="717e41ba50a71a53a80e70255a9aed3a28af1f2d24eb9efc4567403bbd4385f4" Nov 25 10:44:38 crc kubenswrapper[4821]: E1125 10:44:38.165828 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"717e41ba50a71a53a80e70255a9aed3a28af1f2d24eb9efc4567403bbd4385f4\": container with ID starting with 717e41ba50a71a53a80e70255a9aed3a28af1f2d24eb9efc4567403bbd4385f4 not found: ID does not exist" containerID="717e41ba50a71a53a80e70255a9aed3a28af1f2d24eb9efc4567403bbd4385f4" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.165859 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"717e41ba50a71a53a80e70255a9aed3a28af1f2d24eb9efc4567403bbd4385f4"} err="failed to get container status \"717e41ba50a71a53a80e70255a9aed3a28af1f2d24eb9efc4567403bbd4385f4\": rpc error: code = NotFound desc = could not find container \"717e41ba50a71a53a80e70255a9aed3a28af1f2d24eb9efc4567403bbd4385f4\": container with ID starting with 717e41ba50a71a53a80e70255a9aed3a28af1f2d24eb9efc4567403bbd4385f4 not found: ID does not exist" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.165879 4821 scope.go:117] "RemoveContainer" containerID="afa3b5464872e57b7a8921ecdae57338715ab15ac3e01f3eb570907c98771328" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.168433 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-58htt"] Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.171283 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gzl9t\" (UniqueName: \"kubernetes.io/projected/ba27af47-1944-4970-a015-e96c7cee4862-kube-api-access-gzl9t\") pod \"ba27af47-1944-4970-a015-e96c7cee4862\" (UID: \"ba27af47-1944-4970-a015-e96c7cee4862\") " Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.171342 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba27af47-1944-4970-a015-e96c7cee4862-config\") pod \"ba27af47-1944-4970-a015-e96c7cee4862\" (UID: \"ba27af47-1944-4970-a015-e96c7cee4862\") " Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.171454 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ba27af47-1944-4970-a015-e96c7cee4862-serving-cert\") pod \"ba27af47-1944-4970-a015-e96c7cee4862\" (UID: \"ba27af47-1944-4970-a015-e96c7cee4862\") " Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.171486 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ba27af47-1944-4970-a015-e96c7cee4862-client-ca\") pod \"ba27af47-1944-4970-a015-e96c7cee4862\" (UID: \"ba27af47-1944-4970-a015-e96c7cee4862\") " Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.171779 4821 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/518db853-3bbe-421b-a2dd-6d1b2e965120-client-ca\") on node \"crc\" DevicePath \"\"" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.171797 4821 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/518db853-3bbe-421b-a2dd-6d1b2e965120-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.171807 4821 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/518db853-3bbe-421b-a2dd-6d1b2e965120-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.171817 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-trdqk\" (UniqueName: \"kubernetes.io/projected/518db853-3bbe-421b-a2dd-6d1b2e965120-kube-api-access-trdqk\") on node \"crc\" DevicePath \"\"" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.171826 4821 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/518db853-3bbe-421b-a2dd-6d1b2e965120-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.173153 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ba27af47-1944-4970-a015-e96c7cee4862-client-ca" (OuterVolumeSpecName: "client-ca") pod "ba27af47-1944-4970-a015-e96c7cee4862" (UID: "ba27af47-1944-4970-a015-e96c7cee4862"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.173204 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ba27af47-1944-4970-a015-e96c7cee4862-config" (OuterVolumeSpecName: "config") pod "ba27af47-1944-4970-a015-e96c7cee4862" (UID: "ba27af47-1944-4970-a015-e96c7cee4862"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.174842 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba27af47-1944-4970-a015-e96c7cee4862-kube-api-access-gzl9t" (OuterVolumeSpecName: "kube-api-access-gzl9t") pod "ba27af47-1944-4970-a015-e96c7cee4862" (UID: "ba27af47-1944-4970-a015-e96c7cee4862"). InnerVolumeSpecName "kube-api-access-gzl9t". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.182537 4821 scope.go:117] "RemoveContainer" containerID="afa3b5464872e57b7a8921ecdae57338715ab15ac3e01f3eb570907c98771328" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.183627 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba27af47-1944-4970-a015-e96c7cee4862-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "ba27af47-1944-4970-a015-e96c7cee4862" (UID: "ba27af47-1944-4970-a015-e96c7cee4862"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:44:38 crc kubenswrapper[4821]: E1125 10:44:38.183663 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"afa3b5464872e57b7a8921ecdae57338715ab15ac3e01f3eb570907c98771328\": container with ID starting with afa3b5464872e57b7a8921ecdae57338715ab15ac3e01f3eb570907c98771328 not found: ID does not exist" containerID="afa3b5464872e57b7a8921ecdae57338715ab15ac3e01f3eb570907c98771328" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.183700 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"afa3b5464872e57b7a8921ecdae57338715ab15ac3e01f3eb570907c98771328"} err="failed to get container status \"afa3b5464872e57b7a8921ecdae57338715ab15ac3e01f3eb570907c98771328\": rpc error: code = NotFound desc = could not find container \"afa3b5464872e57b7a8921ecdae57338715ab15ac3e01f3eb570907c98771328\": container with ID starting with afa3b5464872e57b7a8921ecdae57338715ab15ac3e01f3eb570907c98771328 not found: ID does not exist" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.273030 4821 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ba27af47-1944-4970-a015-e96c7cee4862-client-ca\") on node \"crc\" DevicePath \"\"" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.273063 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gzl9t\" (UniqueName: \"kubernetes.io/projected/ba27af47-1944-4970-a015-e96c7cee4862-kube-api-access-gzl9t\") on node \"crc\" DevicePath \"\"" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.273072 4821 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba27af47-1944-4970-a015-e96c7cee4862-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.273084 4821 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ba27af47-1944-4970-a015-e96c7cee4862-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.473388 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-dv9nt"] Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.483062 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-dv9nt"] Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.767332 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-xxx9j"] Nov 25 10:44:38 crc kubenswrapper[4821]: E1125 10:44:38.767676 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="518db853-3bbe-421b-a2dd-6d1b2e965120" containerName="controller-manager" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.767704 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="518db853-3bbe-421b-a2dd-6d1b2e965120" containerName="controller-manager" Nov 25 10:44:38 crc kubenswrapper[4821]: E1125 10:44:38.767721 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba27af47-1944-4970-a015-e96c7cee4862" containerName="route-controller-manager" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.767733 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba27af47-1944-4970-a015-e96c7cee4862" containerName="route-controller-manager" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.767939 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba27af47-1944-4970-a015-e96c7cee4862" containerName="route-controller-manager" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.767973 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="518db853-3bbe-421b-a2dd-6d1b2e965120" containerName="controller-manager" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.768554 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-xxx9j" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.770422 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.774489 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.775442 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-85cd86cfbf-mw7kk"] Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.776538 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-85cd86cfbf-mw7kk" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.780637 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.780776 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.782962 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.785322 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.785353 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.785465 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.786284 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-697497c75d-ksxfn"] Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.787199 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-697497c75d-ksxfn" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.793478 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-xxx9j"] Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.801315 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-697497c75d-ksxfn"] Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.805925 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.807948 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.808354 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.808410 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.808715 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.808873 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.809188 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-85cd86cfbf-mw7kk"] Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.809612 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.882100 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/95bdfed1-046c-41a8-a531-d6db7024715b-client-ca\") pod \"controller-manager-85cd86cfbf-mw7kk\" (UID: \"95bdfed1-046c-41a8-a531-d6db7024715b\") " pod="openshift-controller-manager/controller-manager-85cd86cfbf-mw7kk" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.882177 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/95bdfed1-046c-41a8-a531-d6db7024715b-proxy-ca-bundles\") pod \"controller-manager-85cd86cfbf-mw7kk\" (UID: \"95bdfed1-046c-41a8-a531-d6db7024715b\") " pod="openshift-controller-manager/controller-manager-85cd86cfbf-mw7kk" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.882207 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/95bdfed1-046c-41a8-a531-d6db7024715b-config\") pod \"controller-manager-85cd86cfbf-mw7kk\" (UID: \"95bdfed1-046c-41a8-a531-d6db7024715b\") " pod="openshift-controller-manager/controller-manager-85cd86cfbf-mw7kk" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.882237 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7s87l\" (UniqueName: \"kubernetes.io/projected/8add6d36-8ff3-41aa-8383-a9e86b6ab629-kube-api-access-7s87l\") pod \"openstack-operator-index-xxx9j\" (UID: \"8add6d36-8ff3-41aa-8383-a9e86b6ab629\") " pod="openstack-operators/openstack-operator-index-xxx9j" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.882277 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/95bdfed1-046c-41a8-a531-d6db7024715b-serving-cert\") pod \"controller-manager-85cd86cfbf-mw7kk\" (UID: \"95bdfed1-046c-41a8-a531-d6db7024715b\") " pod="openshift-controller-manager/controller-manager-85cd86cfbf-mw7kk" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.882316 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdwzq\" (UniqueName: \"kubernetes.io/projected/95bdfed1-046c-41a8-a531-d6db7024715b-kube-api-access-bdwzq\") pod \"controller-manager-85cd86cfbf-mw7kk\" (UID: \"95bdfed1-046c-41a8-a531-d6db7024715b\") " pod="openshift-controller-manager/controller-manager-85cd86cfbf-mw7kk" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.882338 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e66f3c98-1195-4990-959f-eb709fd72b76-config\") pod \"route-controller-manager-697497c75d-ksxfn\" (UID: \"e66f3c98-1195-4990-959f-eb709fd72b76\") " pod="openshift-route-controller-manager/route-controller-manager-697497c75d-ksxfn" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.882361 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e66f3c98-1195-4990-959f-eb709fd72b76-client-ca\") pod \"route-controller-manager-697497c75d-ksxfn\" (UID: \"e66f3c98-1195-4990-959f-eb709fd72b76\") " pod="openshift-route-controller-manager/route-controller-manager-697497c75d-ksxfn" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.882397 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnv2n\" (UniqueName: \"kubernetes.io/projected/e66f3c98-1195-4990-959f-eb709fd72b76-kube-api-access-lnv2n\") pod \"route-controller-manager-697497c75d-ksxfn\" (UID: \"e66f3c98-1195-4990-959f-eb709fd72b76\") " pod="openshift-route-controller-manager/route-controller-manager-697497c75d-ksxfn" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.882421 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e66f3c98-1195-4990-959f-eb709fd72b76-serving-cert\") pod \"route-controller-manager-697497c75d-ksxfn\" (UID: \"e66f3c98-1195-4990-959f-eb709fd72b76\") " pod="openshift-route-controller-manager/route-controller-manager-697497c75d-ksxfn" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.984143 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/95bdfed1-046c-41a8-a531-d6db7024715b-client-ca\") pod \"controller-manager-85cd86cfbf-mw7kk\" (UID: \"95bdfed1-046c-41a8-a531-d6db7024715b\") " pod="openshift-controller-manager/controller-manager-85cd86cfbf-mw7kk" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.984231 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/95bdfed1-046c-41a8-a531-d6db7024715b-proxy-ca-bundles\") pod \"controller-manager-85cd86cfbf-mw7kk\" (UID: \"95bdfed1-046c-41a8-a531-d6db7024715b\") " pod="openshift-controller-manager/controller-manager-85cd86cfbf-mw7kk" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.984256 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/95bdfed1-046c-41a8-a531-d6db7024715b-config\") pod \"controller-manager-85cd86cfbf-mw7kk\" (UID: \"95bdfed1-046c-41a8-a531-d6db7024715b\") " pod="openshift-controller-manager/controller-manager-85cd86cfbf-mw7kk" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.984288 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7s87l\" (UniqueName: \"kubernetes.io/projected/8add6d36-8ff3-41aa-8383-a9e86b6ab629-kube-api-access-7s87l\") pod \"openstack-operator-index-xxx9j\" (UID: \"8add6d36-8ff3-41aa-8383-a9e86b6ab629\") " pod="openstack-operators/openstack-operator-index-xxx9j" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.984327 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/95bdfed1-046c-41a8-a531-d6db7024715b-serving-cert\") pod \"controller-manager-85cd86cfbf-mw7kk\" (UID: \"95bdfed1-046c-41a8-a531-d6db7024715b\") " pod="openshift-controller-manager/controller-manager-85cd86cfbf-mw7kk" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.984410 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdwzq\" (UniqueName: \"kubernetes.io/projected/95bdfed1-046c-41a8-a531-d6db7024715b-kube-api-access-bdwzq\") pod \"controller-manager-85cd86cfbf-mw7kk\" (UID: \"95bdfed1-046c-41a8-a531-d6db7024715b\") " pod="openshift-controller-manager/controller-manager-85cd86cfbf-mw7kk" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.984440 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e66f3c98-1195-4990-959f-eb709fd72b76-config\") pod \"route-controller-manager-697497c75d-ksxfn\" (UID: \"e66f3c98-1195-4990-959f-eb709fd72b76\") " pod="openshift-route-controller-manager/route-controller-manager-697497c75d-ksxfn" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.984462 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e66f3c98-1195-4990-959f-eb709fd72b76-client-ca\") pod \"route-controller-manager-697497c75d-ksxfn\" (UID: \"e66f3c98-1195-4990-959f-eb709fd72b76\") " pod="openshift-route-controller-manager/route-controller-manager-697497c75d-ksxfn" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.984508 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lnv2n\" (UniqueName: \"kubernetes.io/projected/e66f3c98-1195-4990-959f-eb709fd72b76-kube-api-access-lnv2n\") pod \"route-controller-manager-697497c75d-ksxfn\" (UID: \"e66f3c98-1195-4990-959f-eb709fd72b76\") " pod="openshift-route-controller-manager/route-controller-manager-697497c75d-ksxfn" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.984533 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e66f3c98-1195-4990-959f-eb709fd72b76-serving-cert\") pod \"route-controller-manager-697497c75d-ksxfn\" (UID: \"e66f3c98-1195-4990-959f-eb709fd72b76\") " pod="openshift-route-controller-manager/route-controller-manager-697497c75d-ksxfn" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.985335 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/95bdfed1-046c-41a8-a531-d6db7024715b-client-ca\") pod \"controller-manager-85cd86cfbf-mw7kk\" (UID: \"95bdfed1-046c-41a8-a531-d6db7024715b\") " pod="openshift-controller-manager/controller-manager-85cd86cfbf-mw7kk" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.985465 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e66f3c98-1195-4990-959f-eb709fd72b76-client-ca\") pod \"route-controller-manager-697497c75d-ksxfn\" (UID: \"e66f3c98-1195-4990-959f-eb709fd72b76\") " pod="openshift-route-controller-manager/route-controller-manager-697497c75d-ksxfn" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.985878 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e66f3c98-1195-4990-959f-eb709fd72b76-config\") pod \"route-controller-manager-697497c75d-ksxfn\" (UID: \"e66f3c98-1195-4990-959f-eb709fd72b76\") " pod="openshift-route-controller-manager/route-controller-manager-697497c75d-ksxfn" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.985938 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/95bdfed1-046c-41a8-a531-d6db7024715b-config\") pod \"controller-manager-85cd86cfbf-mw7kk\" (UID: \"95bdfed1-046c-41a8-a531-d6db7024715b\") " pod="openshift-controller-manager/controller-manager-85cd86cfbf-mw7kk" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.986335 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/95bdfed1-046c-41a8-a531-d6db7024715b-proxy-ca-bundles\") pod \"controller-manager-85cd86cfbf-mw7kk\" (UID: \"95bdfed1-046c-41a8-a531-d6db7024715b\") " pod="openshift-controller-manager/controller-manager-85cd86cfbf-mw7kk" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.988088 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/95bdfed1-046c-41a8-a531-d6db7024715b-serving-cert\") pod \"controller-manager-85cd86cfbf-mw7kk\" (UID: \"95bdfed1-046c-41a8-a531-d6db7024715b\") " pod="openshift-controller-manager/controller-manager-85cd86cfbf-mw7kk" Nov 25 10:44:38 crc kubenswrapper[4821]: I1125 10:44:38.996879 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e66f3c98-1195-4990-959f-eb709fd72b76-serving-cert\") pod \"route-controller-manager-697497c75d-ksxfn\" (UID: \"e66f3c98-1195-4990-959f-eb709fd72b76\") " pod="openshift-route-controller-manager/route-controller-manager-697497c75d-ksxfn" Nov 25 10:44:39 crc kubenswrapper[4821]: I1125 10:44:39.004650 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdwzq\" (UniqueName: \"kubernetes.io/projected/95bdfed1-046c-41a8-a531-d6db7024715b-kube-api-access-bdwzq\") pod \"controller-manager-85cd86cfbf-mw7kk\" (UID: \"95bdfed1-046c-41a8-a531-d6db7024715b\") " pod="openshift-controller-manager/controller-manager-85cd86cfbf-mw7kk" Nov 25 10:44:39 crc kubenswrapper[4821]: I1125 10:44:39.005452 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnv2n\" (UniqueName: \"kubernetes.io/projected/e66f3c98-1195-4990-959f-eb709fd72b76-kube-api-access-lnv2n\") pod \"route-controller-manager-697497c75d-ksxfn\" (UID: \"e66f3c98-1195-4990-959f-eb709fd72b76\") " pod="openshift-route-controller-manager/route-controller-manager-697497c75d-ksxfn" Nov 25 10:44:39 crc kubenswrapper[4821]: I1125 10:44:39.005694 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7s87l\" (UniqueName: \"kubernetes.io/projected/8add6d36-8ff3-41aa-8383-a9e86b6ab629-kube-api-access-7s87l\") pod \"openstack-operator-index-xxx9j\" (UID: \"8add6d36-8ff3-41aa-8383-a9e86b6ab629\") " pod="openstack-operators/openstack-operator-index-xxx9j" Nov 25 10:44:39 crc kubenswrapper[4821]: I1125 10:44:39.086282 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-xxx9j" Nov 25 10:44:39 crc kubenswrapper[4821]: I1125 10:44:39.098700 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-85cd86cfbf-mw7kk" Nov 25 10:44:39 crc kubenswrapper[4821]: I1125 10:44:39.107259 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-697497c75d-ksxfn" Nov 25 10:44:39 crc kubenswrapper[4821]: W1125 10:44:39.547348 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8add6d36_8ff3_41aa_8383_a9e86b6ab629.slice/crio-2e5e993c4c9876b965e91137109161296455ee564f701df688bb3a21d1d99bee WatchSource:0}: Error finding container 2e5e993c4c9876b965e91137109161296455ee564f701df688bb3a21d1d99bee: Status 404 returned error can't find the container with id 2e5e993c4c9876b965e91137109161296455ee564f701df688bb3a21d1d99bee Nov 25 10:44:39 crc kubenswrapper[4821]: I1125 10:44:39.549616 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-xxx9j"] Nov 25 10:44:39 crc kubenswrapper[4821]: I1125 10:44:39.602947 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-697497c75d-ksxfn"] Nov 25 10:44:39 crc kubenswrapper[4821]: W1125 10:44:39.606145 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode66f3c98_1195_4990_959f_eb709fd72b76.slice/crio-e8136e7440e9c4d6a9406dd586b6ea672d5e82a1ee0c6b7920af946ec31bb4af WatchSource:0}: Error finding container e8136e7440e9c4d6a9406dd586b6ea672d5e82a1ee0c6b7920af946ec31bb4af: Status 404 returned error can't find the container with id e8136e7440e9c4d6a9406dd586b6ea672d5e82a1ee0c6b7920af946ec31bb4af Nov 25 10:44:39 crc kubenswrapper[4821]: I1125 10:44:39.610476 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-85cd86cfbf-mw7kk"] Nov 25 10:44:39 crc kubenswrapper[4821]: I1125 10:44:39.613184 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-7knkm" Nov 25 10:44:39 crc kubenswrapper[4821]: I1125 10:44:39.687789 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-7knkm" Nov 25 10:44:40 crc kubenswrapper[4821]: I1125 10:44:40.122432 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="518db853-3bbe-421b-a2dd-6d1b2e965120" path="/var/lib/kubelet/pods/518db853-3bbe-421b-a2dd-6d1b2e965120/volumes" Nov 25 10:44:40 crc kubenswrapper[4821]: I1125 10:44:40.123395 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba27af47-1944-4970-a015-e96c7cee4862" path="/var/lib/kubelet/pods/ba27af47-1944-4970-a015-e96c7cee4862/volumes" Nov 25 10:44:40 crc kubenswrapper[4821]: I1125 10:44:40.172081 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-85cd86cfbf-mw7kk" event={"ID":"95bdfed1-046c-41a8-a531-d6db7024715b","Type":"ContainerStarted","Data":"fcc16ac1d079c951f25c8016eb74b737a9daba250c452754f3a447e37229ceaf"} Nov 25 10:44:40 crc kubenswrapper[4821]: I1125 10:44:40.172145 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-85cd86cfbf-mw7kk" event={"ID":"95bdfed1-046c-41a8-a531-d6db7024715b","Type":"ContainerStarted","Data":"847dd2259313668bdad7897f62f90253f710ac68f80eb79df51916dbbff955f6"} Nov 25 10:44:40 crc kubenswrapper[4821]: I1125 10:44:40.173295 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-85cd86cfbf-mw7kk" Nov 25 10:44:40 crc kubenswrapper[4821]: I1125 10:44:40.175385 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-697497c75d-ksxfn" event={"ID":"e66f3c98-1195-4990-959f-eb709fd72b76","Type":"ContainerStarted","Data":"07367166608c45577c55d798ea24f3f9af177a9543dd36eda3e5ee978090b76a"} Nov 25 10:44:40 crc kubenswrapper[4821]: I1125 10:44:40.175437 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-697497c75d-ksxfn" event={"ID":"e66f3c98-1195-4990-959f-eb709fd72b76","Type":"ContainerStarted","Data":"e8136e7440e9c4d6a9406dd586b6ea672d5e82a1ee0c6b7920af946ec31bb4af"} Nov 25 10:44:40 crc kubenswrapper[4821]: I1125 10:44:40.175712 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-697497c75d-ksxfn" Nov 25 10:44:40 crc kubenswrapper[4821]: I1125 10:44:40.179688 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-xxx9j" event={"ID":"8add6d36-8ff3-41aa-8383-a9e86b6ab629","Type":"ContainerStarted","Data":"2e5e993c4c9876b965e91137109161296455ee564f701df688bb3a21d1d99bee"} Nov 25 10:44:40 crc kubenswrapper[4821]: I1125 10:44:40.199590 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-85cd86cfbf-mw7kk" Nov 25 10:44:40 crc kubenswrapper[4821]: I1125 10:44:40.275751 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-85cd86cfbf-mw7kk" podStartSLOduration=3.275719495 podStartE2EDuration="3.275719495s" podCreationTimestamp="2025-11-25 10:44:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:44:40.272179483 +0000 UTC m=+750.808499340" watchObservedRunningTime="2025-11-25 10:44:40.275719495 +0000 UTC m=+750.812039342" Nov 25 10:44:40 crc kubenswrapper[4821]: I1125 10:44:40.291709 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-697497c75d-ksxfn" podStartSLOduration=3.291688346 podStartE2EDuration="3.291688346s" podCreationTimestamp="2025-11-25 10:44:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:44:40.291131562 +0000 UTC m=+750.827451409" watchObservedRunningTime="2025-11-25 10:44:40.291688346 +0000 UTC m=+750.828008203" Nov 25 10:44:40 crc kubenswrapper[4821]: I1125 10:44:40.522089 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-697497c75d-ksxfn" Nov 25 10:44:43 crc kubenswrapper[4821]: I1125 10:44:43.197599 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-xxx9j" event={"ID":"8add6d36-8ff3-41aa-8383-a9e86b6ab629","Type":"ContainerStarted","Data":"df7807b1909a9cd5299c19adbc7b4945f07f73e506859745c3ca9279903fd1e7"} Nov 25 10:44:43 crc kubenswrapper[4821]: I1125 10:44:43.210260 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-xxx9j" podStartSLOduration=1.820155196 podStartE2EDuration="5.210242922s" podCreationTimestamp="2025-11-25 10:44:38 +0000 UTC" firstStartedPulling="2025-11-25 10:44:39.54945915 +0000 UTC m=+750.085778997" lastFinishedPulling="2025-11-25 10:44:42.939546876 +0000 UTC m=+753.475866723" observedRunningTime="2025-11-25 10:44:43.208361633 +0000 UTC m=+753.744681480" watchObservedRunningTime="2025-11-25 10:44:43.210242922 +0000 UTC m=+753.746562769" Nov 25 10:44:44 crc kubenswrapper[4821]: I1125 10:44:44.028286 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-6998585d5-bclxv" Nov 25 10:44:44 crc kubenswrapper[4821]: I1125 10:44:44.545153 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-xxx9j"] Nov 25 10:44:44 crc kubenswrapper[4821]: I1125 10:44:44.615649 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-7knkm" Nov 25 10:44:44 crc kubenswrapper[4821]: I1125 10:44:44.952477 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-tzjrl"] Nov 25 10:44:44 crc kubenswrapper[4821]: I1125 10:44:44.953429 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-tzjrl" Nov 25 10:44:44 crc kubenswrapper[4821]: I1125 10:44:44.955537 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-4g8td" Nov 25 10:44:44 crc kubenswrapper[4821]: I1125 10:44:44.972197 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-tzjrl"] Nov 25 10:44:45 crc kubenswrapper[4821]: I1125 10:44:45.086788 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xwd86\" (UniqueName: \"kubernetes.io/projected/afdcf8ff-544a-4625-b72e-d1a880434014-kube-api-access-xwd86\") pod \"openstack-operator-index-tzjrl\" (UID: \"afdcf8ff-544a-4625-b72e-d1a880434014\") " pod="openstack-operators/openstack-operator-index-tzjrl" Nov 25 10:44:45 crc kubenswrapper[4821]: I1125 10:44:45.188674 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xwd86\" (UniqueName: \"kubernetes.io/projected/afdcf8ff-544a-4625-b72e-d1a880434014-kube-api-access-xwd86\") pod \"openstack-operator-index-tzjrl\" (UID: \"afdcf8ff-544a-4625-b72e-d1a880434014\") " pod="openstack-operators/openstack-operator-index-tzjrl" Nov 25 10:44:45 crc kubenswrapper[4821]: I1125 10:44:45.208567 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xwd86\" (UniqueName: \"kubernetes.io/projected/afdcf8ff-544a-4625-b72e-d1a880434014-kube-api-access-xwd86\") pod \"openstack-operator-index-tzjrl\" (UID: \"afdcf8ff-544a-4625-b72e-d1a880434014\") " pod="openstack-operators/openstack-operator-index-tzjrl" Nov 25 10:44:45 crc kubenswrapper[4821]: I1125 10:44:45.209467 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-xxx9j" podUID="8add6d36-8ff3-41aa-8383-a9e86b6ab629" containerName="registry-server" containerID="cri-o://df7807b1909a9cd5299c19adbc7b4945f07f73e506859745c3ca9279903fd1e7" gracePeriod=2 Nov 25 10:44:45 crc kubenswrapper[4821]: I1125 10:44:45.286935 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-tzjrl" Nov 25 10:44:45 crc kubenswrapper[4821]: W1125 10:44:45.706652 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podafdcf8ff_544a_4625_b72e_d1a880434014.slice/crio-8a065f43e0f3bc77c2c47cbdbd1a7e2a86f64e21186d9de86fdbd7fa7da17cb6 WatchSource:0}: Error finding container 8a065f43e0f3bc77c2c47cbdbd1a7e2a86f64e21186d9de86fdbd7fa7da17cb6: Status 404 returned error can't find the container with id 8a065f43e0f3bc77c2c47cbdbd1a7e2a86f64e21186d9de86fdbd7fa7da17cb6 Nov 25 10:44:45 crc kubenswrapper[4821]: I1125 10:44:45.707124 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-tzjrl"] Nov 25 10:44:45 crc kubenswrapper[4821]: I1125 10:44:45.713565 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-xxx9j" Nov 25 10:44:45 crc kubenswrapper[4821]: I1125 10:44:45.795175 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7s87l\" (UniqueName: \"kubernetes.io/projected/8add6d36-8ff3-41aa-8383-a9e86b6ab629-kube-api-access-7s87l\") pod \"8add6d36-8ff3-41aa-8383-a9e86b6ab629\" (UID: \"8add6d36-8ff3-41aa-8383-a9e86b6ab629\") " Nov 25 10:44:45 crc kubenswrapper[4821]: I1125 10:44:45.800295 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8add6d36-8ff3-41aa-8383-a9e86b6ab629-kube-api-access-7s87l" (OuterVolumeSpecName: "kube-api-access-7s87l") pod "8add6d36-8ff3-41aa-8383-a9e86b6ab629" (UID: "8add6d36-8ff3-41aa-8383-a9e86b6ab629"). InnerVolumeSpecName "kube-api-access-7s87l". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:44:45 crc kubenswrapper[4821]: I1125 10:44:45.896518 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7s87l\" (UniqueName: \"kubernetes.io/projected/8add6d36-8ff3-41aa-8383-a9e86b6ab629-kube-api-access-7s87l\") on node \"crc\" DevicePath \"\"" Nov 25 10:44:46 crc kubenswrapper[4821]: I1125 10:44:46.222668 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-tzjrl" event={"ID":"afdcf8ff-544a-4625-b72e-d1a880434014","Type":"ContainerStarted","Data":"f2327cc56b7386d0d49bf7244062c000e9e04d9791ea06ea12670c841346d185"} Nov 25 10:44:46 crc kubenswrapper[4821]: I1125 10:44:46.224052 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-tzjrl" event={"ID":"afdcf8ff-544a-4625-b72e-d1a880434014","Type":"ContainerStarted","Data":"8a065f43e0f3bc77c2c47cbdbd1a7e2a86f64e21186d9de86fdbd7fa7da17cb6"} Nov 25 10:44:46 crc kubenswrapper[4821]: I1125 10:44:46.228360 4821 generic.go:334] "Generic (PLEG): container finished" podID="8add6d36-8ff3-41aa-8383-a9e86b6ab629" containerID="df7807b1909a9cd5299c19adbc7b4945f07f73e506859745c3ca9279903fd1e7" exitCode=0 Nov 25 10:44:46 crc kubenswrapper[4821]: I1125 10:44:46.228449 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-xxx9j" event={"ID":"8add6d36-8ff3-41aa-8383-a9e86b6ab629","Type":"ContainerDied","Data":"df7807b1909a9cd5299c19adbc7b4945f07f73e506859745c3ca9279903fd1e7"} Nov 25 10:44:46 crc kubenswrapper[4821]: I1125 10:44:46.228511 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-xxx9j" event={"ID":"8add6d36-8ff3-41aa-8383-a9e86b6ab629","Type":"ContainerDied","Data":"2e5e993c4c9876b965e91137109161296455ee564f701df688bb3a21d1d99bee"} Nov 25 10:44:46 crc kubenswrapper[4821]: I1125 10:44:46.228541 4821 scope.go:117] "RemoveContainer" containerID="df7807b1909a9cd5299c19adbc7b4945f07f73e506859745c3ca9279903fd1e7" Nov 25 10:44:46 crc kubenswrapper[4821]: I1125 10:44:46.228447 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-xxx9j" Nov 25 10:44:46 crc kubenswrapper[4821]: I1125 10:44:46.248222 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-tzjrl" podStartSLOduration=2.1961335220000002 podStartE2EDuration="2.248157483s" podCreationTimestamp="2025-11-25 10:44:44 +0000 UTC" firstStartedPulling="2025-11-25 10:44:45.711436743 +0000 UTC m=+756.247756590" lastFinishedPulling="2025-11-25 10:44:45.763460704 +0000 UTC m=+756.299780551" observedRunningTime="2025-11-25 10:44:46.242405645 +0000 UTC m=+756.778725642" watchObservedRunningTime="2025-11-25 10:44:46.248157483 +0000 UTC m=+756.784477371" Nov 25 10:44:46 crc kubenswrapper[4821]: I1125 10:44:46.257516 4821 scope.go:117] "RemoveContainer" containerID="df7807b1909a9cd5299c19adbc7b4945f07f73e506859745c3ca9279903fd1e7" Nov 25 10:44:46 crc kubenswrapper[4821]: E1125 10:44:46.259048 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df7807b1909a9cd5299c19adbc7b4945f07f73e506859745c3ca9279903fd1e7\": container with ID starting with df7807b1909a9cd5299c19adbc7b4945f07f73e506859745c3ca9279903fd1e7 not found: ID does not exist" containerID="df7807b1909a9cd5299c19adbc7b4945f07f73e506859745c3ca9279903fd1e7" Nov 25 10:44:46 crc kubenswrapper[4821]: I1125 10:44:46.259087 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df7807b1909a9cd5299c19adbc7b4945f07f73e506859745c3ca9279903fd1e7"} err="failed to get container status \"df7807b1909a9cd5299c19adbc7b4945f07f73e506859745c3ca9279903fd1e7\": rpc error: code = NotFound desc = could not find container \"df7807b1909a9cd5299c19adbc7b4945f07f73e506859745c3ca9279903fd1e7\": container with ID starting with df7807b1909a9cd5299c19adbc7b4945f07f73e506859745c3ca9279903fd1e7 not found: ID does not exist" Nov 25 10:44:46 crc kubenswrapper[4821]: I1125 10:44:46.265536 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-xxx9j"] Nov 25 10:44:46 crc kubenswrapper[4821]: I1125 10:44:46.272170 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-xxx9j"] Nov 25 10:44:47 crc kubenswrapper[4821]: I1125 10:44:47.005374 4821 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 25 10:44:48 crc kubenswrapper[4821]: I1125 10:44:48.122998 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8add6d36-8ff3-41aa-8383-a9e86b6ab629" path="/var/lib/kubelet/pods/8add6d36-8ff3-41aa-8383-a9e86b6ab629/volumes" Nov 25 10:44:55 crc kubenswrapper[4821]: I1125 10:44:55.287442 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-tzjrl" Nov 25 10:44:55 crc kubenswrapper[4821]: I1125 10:44:55.287748 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-tzjrl" Nov 25 10:44:55 crc kubenswrapper[4821]: I1125 10:44:55.330150 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-tzjrl" Nov 25 10:44:56 crc kubenswrapper[4821]: I1125 10:44:56.317097 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-tzjrl" Nov 25 10:45:00 crc kubenswrapper[4821]: I1125 10:45:00.151527 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401125-5p66d"] Nov 25 10:45:00 crc kubenswrapper[4821]: E1125 10:45:00.152014 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8add6d36-8ff3-41aa-8383-a9e86b6ab629" containerName="registry-server" Nov 25 10:45:00 crc kubenswrapper[4821]: I1125 10:45:00.152025 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="8add6d36-8ff3-41aa-8383-a9e86b6ab629" containerName="registry-server" Nov 25 10:45:00 crc kubenswrapper[4821]: I1125 10:45:00.152138 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="8add6d36-8ff3-41aa-8383-a9e86b6ab629" containerName="registry-server" Nov 25 10:45:00 crc kubenswrapper[4821]: I1125 10:45:00.152532 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401125-5p66d" Nov 25 10:45:00 crc kubenswrapper[4821]: I1125 10:45:00.154134 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 25 10:45:00 crc kubenswrapper[4821]: I1125 10:45:00.156021 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 25 10:45:00 crc kubenswrapper[4821]: I1125 10:45:00.161409 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401125-5p66d"] Nov 25 10:45:00 crc kubenswrapper[4821]: I1125 10:45:00.290805 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2fl2w\" (UniqueName: \"kubernetes.io/projected/52eea54a-9fe6-40ec-9354-8937bcc5a924-kube-api-access-2fl2w\") pod \"collect-profiles-29401125-5p66d\" (UID: \"52eea54a-9fe6-40ec-9354-8937bcc5a924\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401125-5p66d" Nov 25 10:45:00 crc kubenswrapper[4821]: I1125 10:45:00.290875 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/52eea54a-9fe6-40ec-9354-8937bcc5a924-config-volume\") pod \"collect-profiles-29401125-5p66d\" (UID: \"52eea54a-9fe6-40ec-9354-8937bcc5a924\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401125-5p66d" Nov 25 10:45:00 crc kubenswrapper[4821]: I1125 10:45:00.290972 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/52eea54a-9fe6-40ec-9354-8937bcc5a924-secret-volume\") pod \"collect-profiles-29401125-5p66d\" (UID: \"52eea54a-9fe6-40ec-9354-8937bcc5a924\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401125-5p66d" Nov 25 10:45:00 crc kubenswrapper[4821]: I1125 10:45:00.391930 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2fl2w\" (UniqueName: \"kubernetes.io/projected/52eea54a-9fe6-40ec-9354-8937bcc5a924-kube-api-access-2fl2w\") pod \"collect-profiles-29401125-5p66d\" (UID: \"52eea54a-9fe6-40ec-9354-8937bcc5a924\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401125-5p66d" Nov 25 10:45:00 crc kubenswrapper[4821]: I1125 10:45:00.391992 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/52eea54a-9fe6-40ec-9354-8937bcc5a924-config-volume\") pod \"collect-profiles-29401125-5p66d\" (UID: \"52eea54a-9fe6-40ec-9354-8937bcc5a924\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401125-5p66d" Nov 25 10:45:00 crc kubenswrapper[4821]: I1125 10:45:00.392032 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/52eea54a-9fe6-40ec-9354-8937bcc5a924-secret-volume\") pod \"collect-profiles-29401125-5p66d\" (UID: \"52eea54a-9fe6-40ec-9354-8937bcc5a924\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401125-5p66d" Nov 25 10:45:00 crc kubenswrapper[4821]: I1125 10:45:00.393816 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/52eea54a-9fe6-40ec-9354-8937bcc5a924-config-volume\") pod \"collect-profiles-29401125-5p66d\" (UID: \"52eea54a-9fe6-40ec-9354-8937bcc5a924\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401125-5p66d" Nov 25 10:45:00 crc kubenswrapper[4821]: I1125 10:45:00.396838 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/52eea54a-9fe6-40ec-9354-8937bcc5a924-secret-volume\") pod \"collect-profiles-29401125-5p66d\" (UID: \"52eea54a-9fe6-40ec-9354-8937bcc5a924\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401125-5p66d" Nov 25 10:45:00 crc kubenswrapper[4821]: I1125 10:45:00.408715 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2fl2w\" (UniqueName: \"kubernetes.io/projected/52eea54a-9fe6-40ec-9354-8937bcc5a924-kube-api-access-2fl2w\") pod \"collect-profiles-29401125-5p66d\" (UID: \"52eea54a-9fe6-40ec-9354-8937bcc5a924\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401125-5p66d" Nov 25 10:45:00 crc kubenswrapper[4821]: I1125 10:45:00.479679 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401125-5p66d" Nov 25 10:45:01 crc kubenswrapper[4821]: I1125 10:45:00.910411 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401125-5p66d"] Nov 25 10:45:01 crc kubenswrapper[4821]: W1125 10:45:00.916985 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod52eea54a_9fe6_40ec_9354_8937bcc5a924.slice/crio-674b3a5d1a79a531a18a1d87f79aedd646047dae7ec3b5d0d4f873c43ee21601 WatchSource:0}: Error finding container 674b3a5d1a79a531a18a1d87f79aedd646047dae7ec3b5d0d4f873c43ee21601: Status 404 returned error can't find the container with id 674b3a5d1a79a531a18a1d87f79aedd646047dae7ec3b5d0d4f873c43ee21601 Nov 25 10:45:01 crc kubenswrapper[4821]: I1125 10:45:01.323924 4821 generic.go:334] "Generic (PLEG): container finished" podID="52eea54a-9fe6-40ec-9354-8937bcc5a924" containerID="e6513bd7ff68841deae35d9fd9ce21fe69b3ac3a094be4c9a379dbe841e020dd" exitCode=0 Nov 25 10:45:01 crc kubenswrapper[4821]: I1125 10:45:01.323973 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401125-5p66d" event={"ID":"52eea54a-9fe6-40ec-9354-8937bcc5a924","Type":"ContainerDied","Data":"e6513bd7ff68841deae35d9fd9ce21fe69b3ac3a094be4c9a379dbe841e020dd"} Nov 25 10:45:01 crc kubenswrapper[4821]: I1125 10:45:01.324000 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401125-5p66d" event={"ID":"52eea54a-9fe6-40ec-9354-8937bcc5a924","Type":"ContainerStarted","Data":"674b3a5d1a79a531a18a1d87f79aedd646047dae7ec3b5d0d4f873c43ee21601"} Nov 25 10:45:02 crc kubenswrapper[4821]: I1125 10:45:02.206766 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/f381f3655dcf4f93a6a93fe2921d660ff39f878b90de9b88a6ab8f46514g6tz"] Nov 25 10:45:02 crc kubenswrapper[4821]: I1125 10:45:02.208013 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/f381f3655dcf4f93a6a93fe2921d660ff39f878b90de9b88a6ab8f46514g6tz" Nov 25 10:45:02 crc kubenswrapper[4821]: I1125 10:45:02.209661 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-cpx2z" Nov 25 10:45:02 crc kubenswrapper[4821]: I1125 10:45:02.222509 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/f381f3655dcf4f93a6a93fe2921d660ff39f878b90de9b88a6ab8f46514g6tz"] Nov 25 10:45:02 crc kubenswrapper[4821]: I1125 10:45:02.323973 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bd4ce1de-05e0-49f4-ad3a-7baae72f3e09-bundle\") pod \"f381f3655dcf4f93a6a93fe2921d660ff39f878b90de9b88a6ab8f46514g6tz\" (UID: \"bd4ce1de-05e0-49f4-ad3a-7baae72f3e09\") " pod="openstack-operators/f381f3655dcf4f93a6a93fe2921d660ff39f878b90de9b88a6ab8f46514g6tz" Nov 25 10:45:02 crc kubenswrapper[4821]: I1125 10:45:02.324956 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bd4ce1de-05e0-49f4-ad3a-7baae72f3e09-util\") pod \"f381f3655dcf4f93a6a93fe2921d660ff39f878b90de9b88a6ab8f46514g6tz\" (UID: \"bd4ce1de-05e0-49f4-ad3a-7baae72f3e09\") " pod="openstack-operators/f381f3655dcf4f93a6a93fe2921d660ff39f878b90de9b88a6ab8f46514g6tz" Nov 25 10:45:02 crc kubenswrapper[4821]: I1125 10:45:02.325112 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5jqw7\" (UniqueName: \"kubernetes.io/projected/bd4ce1de-05e0-49f4-ad3a-7baae72f3e09-kube-api-access-5jqw7\") pod \"f381f3655dcf4f93a6a93fe2921d660ff39f878b90de9b88a6ab8f46514g6tz\" (UID: \"bd4ce1de-05e0-49f4-ad3a-7baae72f3e09\") " pod="openstack-operators/f381f3655dcf4f93a6a93fe2921d660ff39f878b90de9b88a6ab8f46514g6tz" Nov 25 10:45:02 crc kubenswrapper[4821]: I1125 10:45:02.427091 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5jqw7\" (UniqueName: \"kubernetes.io/projected/bd4ce1de-05e0-49f4-ad3a-7baae72f3e09-kube-api-access-5jqw7\") pod \"f381f3655dcf4f93a6a93fe2921d660ff39f878b90de9b88a6ab8f46514g6tz\" (UID: \"bd4ce1de-05e0-49f4-ad3a-7baae72f3e09\") " pod="openstack-operators/f381f3655dcf4f93a6a93fe2921d660ff39f878b90de9b88a6ab8f46514g6tz" Nov 25 10:45:02 crc kubenswrapper[4821]: I1125 10:45:02.427227 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bd4ce1de-05e0-49f4-ad3a-7baae72f3e09-bundle\") pod \"f381f3655dcf4f93a6a93fe2921d660ff39f878b90de9b88a6ab8f46514g6tz\" (UID: \"bd4ce1de-05e0-49f4-ad3a-7baae72f3e09\") " pod="openstack-operators/f381f3655dcf4f93a6a93fe2921d660ff39f878b90de9b88a6ab8f46514g6tz" Nov 25 10:45:02 crc kubenswrapper[4821]: I1125 10:45:02.427262 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bd4ce1de-05e0-49f4-ad3a-7baae72f3e09-util\") pod \"f381f3655dcf4f93a6a93fe2921d660ff39f878b90de9b88a6ab8f46514g6tz\" (UID: \"bd4ce1de-05e0-49f4-ad3a-7baae72f3e09\") " pod="openstack-operators/f381f3655dcf4f93a6a93fe2921d660ff39f878b90de9b88a6ab8f46514g6tz" Nov 25 10:45:02 crc kubenswrapper[4821]: I1125 10:45:02.427701 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bd4ce1de-05e0-49f4-ad3a-7baae72f3e09-util\") pod \"f381f3655dcf4f93a6a93fe2921d660ff39f878b90de9b88a6ab8f46514g6tz\" (UID: \"bd4ce1de-05e0-49f4-ad3a-7baae72f3e09\") " pod="openstack-operators/f381f3655dcf4f93a6a93fe2921d660ff39f878b90de9b88a6ab8f46514g6tz" Nov 25 10:45:02 crc kubenswrapper[4821]: I1125 10:45:02.428191 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bd4ce1de-05e0-49f4-ad3a-7baae72f3e09-bundle\") pod \"f381f3655dcf4f93a6a93fe2921d660ff39f878b90de9b88a6ab8f46514g6tz\" (UID: \"bd4ce1de-05e0-49f4-ad3a-7baae72f3e09\") " pod="openstack-operators/f381f3655dcf4f93a6a93fe2921d660ff39f878b90de9b88a6ab8f46514g6tz" Nov 25 10:45:02 crc kubenswrapper[4821]: I1125 10:45:02.444449 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5jqw7\" (UniqueName: \"kubernetes.io/projected/bd4ce1de-05e0-49f4-ad3a-7baae72f3e09-kube-api-access-5jqw7\") pod \"f381f3655dcf4f93a6a93fe2921d660ff39f878b90de9b88a6ab8f46514g6tz\" (UID: \"bd4ce1de-05e0-49f4-ad3a-7baae72f3e09\") " pod="openstack-operators/f381f3655dcf4f93a6a93fe2921d660ff39f878b90de9b88a6ab8f46514g6tz" Nov 25 10:45:02 crc kubenswrapper[4821]: I1125 10:45:02.528388 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/f381f3655dcf4f93a6a93fe2921d660ff39f878b90de9b88a6ab8f46514g6tz" Nov 25 10:45:02 crc kubenswrapper[4821]: I1125 10:45:02.686819 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401125-5p66d" Nov 25 10:45:02 crc kubenswrapper[4821]: I1125 10:45:02.834049 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/52eea54a-9fe6-40ec-9354-8937bcc5a924-config-volume\") pod \"52eea54a-9fe6-40ec-9354-8937bcc5a924\" (UID: \"52eea54a-9fe6-40ec-9354-8937bcc5a924\") " Nov 25 10:45:02 crc kubenswrapper[4821]: I1125 10:45:02.834361 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2fl2w\" (UniqueName: \"kubernetes.io/projected/52eea54a-9fe6-40ec-9354-8937bcc5a924-kube-api-access-2fl2w\") pod \"52eea54a-9fe6-40ec-9354-8937bcc5a924\" (UID: \"52eea54a-9fe6-40ec-9354-8937bcc5a924\") " Nov 25 10:45:02 crc kubenswrapper[4821]: I1125 10:45:02.834480 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/52eea54a-9fe6-40ec-9354-8937bcc5a924-secret-volume\") pod \"52eea54a-9fe6-40ec-9354-8937bcc5a924\" (UID: \"52eea54a-9fe6-40ec-9354-8937bcc5a924\") " Nov 25 10:45:02 crc kubenswrapper[4821]: I1125 10:45:02.835396 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/52eea54a-9fe6-40ec-9354-8937bcc5a924-config-volume" (OuterVolumeSpecName: "config-volume") pod "52eea54a-9fe6-40ec-9354-8937bcc5a924" (UID: "52eea54a-9fe6-40ec-9354-8937bcc5a924"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:45:02 crc kubenswrapper[4821]: I1125 10:45:02.840003 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52eea54a-9fe6-40ec-9354-8937bcc5a924-kube-api-access-2fl2w" (OuterVolumeSpecName: "kube-api-access-2fl2w") pod "52eea54a-9fe6-40ec-9354-8937bcc5a924" (UID: "52eea54a-9fe6-40ec-9354-8937bcc5a924"). InnerVolumeSpecName "kube-api-access-2fl2w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:45:02 crc kubenswrapper[4821]: I1125 10:45:02.840100 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52eea54a-9fe6-40ec-9354-8937bcc5a924-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "52eea54a-9fe6-40ec-9354-8937bcc5a924" (UID: "52eea54a-9fe6-40ec-9354-8937bcc5a924"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:45:02 crc kubenswrapper[4821]: I1125 10:45:02.936188 4821 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/52eea54a-9fe6-40ec-9354-8937bcc5a924-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 25 10:45:02 crc kubenswrapper[4821]: I1125 10:45:02.936230 4821 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/52eea54a-9fe6-40ec-9354-8937bcc5a924-config-volume\") on node \"crc\" DevicePath \"\"" Nov 25 10:45:02 crc kubenswrapper[4821]: I1125 10:45:02.936240 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2fl2w\" (UniqueName: \"kubernetes.io/projected/52eea54a-9fe6-40ec-9354-8937bcc5a924-kube-api-access-2fl2w\") on node \"crc\" DevicePath \"\"" Nov 25 10:45:02 crc kubenswrapper[4821]: I1125 10:45:02.940784 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/f381f3655dcf4f93a6a93fe2921d660ff39f878b90de9b88a6ab8f46514g6tz"] Nov 25 10:45:02 crc kubenswrapper[4821]: W1125 10:45:02.948440 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbd4ce1de_05e0_49f4_ad3a_7baae72f3e09.slice/crio-aa33e02bc07289bec1dee2f5ab4b64e2840ac9c8886652c0ef5a244057d5b646 WatchSource:0}: Error finding container aa33e02bc07289bec1dee2f5ab4b64e2840ac9c8886652c0ef5a244057d5b646: Status 404 returned error can't find the container with id aa33e02bc07289bec1dee2f5ab4b64e2840ac9c8886652c0ef5a244057d5b646 Nov 25 10:45:03 crc kubenswrapper[4821]: I1125 10:45:03.335754 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401125-5p66d" event={"ID":"52eea54a-9fe6-40ec-9354-8937bcc5a924","Type":"ContainerDied","Data":"674b3a5d1a79a531a18a1d87f79aedd646047dae7ec3b5d0d4f873c43ee21601"} Nov 25 10:45:03 crc kubenswrapper[4821]: I1125 10:45:03.335800 4821 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="674b3a5d1a79a531a18a1d87f79aedd646047dae7ec3b5d0d4f873c43ee21601" Nov 25 10:45:03 crc kubenswrapper[4821]: I1125 10:45:03.335833 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401125-5p66d" Nov 25 10:45:03 crc kubenswrapper[4821]: I1125 10:45:03.337744 4821 generic.go:334] "Generic (PLEG): container finished" podID="bd4ce1de-05e0-49f4-ad3a-7baae72f3e09" containerID="4dd6f482e1bff2b508db89cc459e21dcdd9e38c7acbab46361a05a3d40cf7e09" exitCode=0 Nov 25 10:45:03 crc kubenswrapper[4821]: I1125 10:45:03.337785 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f381f3655dcf4f93a6a93fe2921d660ff39f878b90de9b88a6ab8f46514g6tz" event={"ID":"bd4ce1de-05e0-49f4-ad3a-7baae72f3e09","Type":"ContainerDied","Data":"4dd6f482e1bff2b508db89cc459e21dcdd9e38c7acbab46361a05a3d40cf7e09"} Nov 25 10:45:03 crc kubenswrapper[4821]: I1125 10:45:03.337810 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f381f3655dcf4f93a6a93fe2921d660ff39f878b90de9b88a6ab8f46514g6tz" event={"ID":"bd4ce1de-05e0-49f4-ad3a-7baae72f3e09","Type":"ContainerStarted","Data":"aa33e02bc07289bec1dee2f5ab4b64e2840ac9c8886652c0ef5a244057d5b646"} Nov 25 10:45:04 crc kubenswrapper[4821]: I1125 10:45:04.348258 4821 generic.go:334] "Generic (PLEG): container finished" podID="bd4ce1de-05e0-49f4-ad3a-7baae72f3e09" containerID="182297f2268a70460e7f4c9174a6aeaffe2afa945d84f43eea56c915a192a3bc" exitCode=0 Nov 25 10:45:04 crc kubenswrapper[4821]: I1125 10:45:04.348563 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f381f3655dcf4f93a6a93fe2921d660ff39f878b90de9b88a6ab8f46514g6tz" event={"ID":"bd4ce1de-05e0-49f4-ad3a-7baae72f3e09","Type":"ContainerDied","Data":"182297f2268a70460e7f4c9174a6aeaffe2afa945d84f43eea56c915a192a3bc"} Nov 25 10:45:04 crc kubenswrapper[4821]: I1125 10:45:04.715100 4821 patch_prober.go:28] interesting pod/machine-config-daemon-2krbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:45:04 crc kubenswrapper[4821]: I1125 10:45:04.715419 4821 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:45:05 crc kubenswrapper[4821]: I1125 10:45:05.356788 4821 generic.go:334] "Generic (PLEG): container finished" podID="bd4ce1de-05e0-49f4-ad3a-7baae72f3e09" containerID="2d83ddaf75003465649fa4b076e78b2d19d598701f9a4cac19f83c5e6c510359" exitCode=0 Nov 25 10:45:05 crc kubenswrapper[4821]: I1125 10:45:05.356838 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f381f3655dcf4f93a6a93fe2921d660ff39f878b90de9b88a6ab8f46514g6tz" event={"ID":"bd4ce1de-05e0-49f4-ad3a-7baae72f3e09","Type":"ContainerDied","Data":"2d83ddaf75003465649fa4b076e78b2d19d598701f9a4cac19f83c5e6c510359"} Nov 25 10:45:06 crc kubenswrapper[4821]: I1125 10:45:06.716801 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/f381f3655dcf4f93a6a93fe2921d660ff39f878b90de9b88a6ab8f46514g6tz" Nov 25 10:45:06 crc kubenswrapper[4821]: I1125 10:45:06.886035 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bd4ce1de-05e0-49f4-ad3a-7baae72f3e09-bundle\") pod \"bd4ce1de-05e0-49f4-ad3a-7baae72f3e09\" (UID: \"bd4ce1de-05e0-49f4-ad3a-7baae72f3e09\") " Nov 25 10:45:06 crc kubenswrapper[4821]: I1125 10:45:06.886136 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5jqw7\" (UniqueName: \"kubernetes.io/projected/bd4ce1de-05e0-49f4-ad3a-7baae72f3e09-kube-api-access-5jqw7\") pod \"bd4ce1de-05e0-49f4-ad3a-7baae72f3e09\" (UID: \"bd4ce1de-05e0-49f4-ad3a-7baae72f3e09\") " Nov 25 10:45:06 crc kubenswrapper[4821]: I1125 10:45:06.886185 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bd4ce1de-05e0-49f4-ad3a-7baae72f3e09-util\") pod \"bd4ce1de-05e0-49f4-ad3a-7baae72f3e09\" (UID: \"bd4ce1de-05e0-49f4-ad3a-7baae72f3e09\") " Nov 25 10:45:06 crc kubenswrapper[4821]: I1125 10:45:06.886862 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd4ce1de-05e0-49f4-ad3a-7baae72f3e09-bundle" (OuterVolumeSpecName: "bundle") pod "bd4ce1de-05e0-49f4-ad3a-7baae72f3e09" (UID: "bd4ce1de-05e0-49f4-ad3a-7baae72f3e09"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:45:06 crc kubenswrapper[4821]: I1125 10:45:06.893282 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd4ce1de-05e0-49f4-ad3a-7baae72f3e09-kube-api-access-5jqw7" (OuterVolumeSpecName: "kube-api-access-5jqw7") pod "bd4ce1de-05e0-49f4-ad3a-7baae72f3e09" (UID: "bd4ce1de-05e0-49f4-ad3a-7baae72f3e09"). InnerVolumeSpecName "kube-api-access-5jqw7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:45:06 crc kubenswrapper[4821]: I1125 10:45:06.900754 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd4ce1de-05e0-49f4-ad3a-7baae72f3e09-util" (OuterVolumeSpecName: "util") pod "bd4ce1de-05e0-49f4-ad3a-7baae72f3e09" (UID: "bd4ce1de-05e0-49f4-ad3a-7baae72f3e09"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:45:06 crc kubenswrapper[4821]: I1125 10:45:06.987408 4821 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bd4ce1de-05e0-49f4-ad3a-7baae72f3e09-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:45:06 crc kubenswrapper[4821]: I1125 10:45:06.987450 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5jqw7\" (UniqueName: \"kubernetes.io/projected/bd4ce1de-05e0-49f4-ad3a-7baae72f3e09-kube-api-access-5jqw7\") on node \"crc\" DevicePath \"\"" Nov 25 10:45:06 crc kubenswrapper[4821]: I1125 10:45:06.987461 4821 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bd4ce1de-05e0-49f4-ad3a-7baae72f3e09-util\") on node \"crc\" DevicePath \"\"" Nov 25 10:45:07 crc kubenswrapper[4821]: I1125 10:45:07.369584 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f381f3655dcf4f93a6a93fe2921d660ff39f878b90de9b88a6ab8f46514g6tz" event={"ID":"bd4ce1de-05e0-49f4-ad3a-7baae72f3e09","Type":"ContainerDied","Data":"aa33e02bc07289bec1dee2f5ab4b64e2840ac9c8886652c0ef5a244057d5b646"} Nov 25 10:45:07 crc kubenswrapper[4821]: I1125 10:45:07.369645 4821 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aa33e02bc07289bec1dee2f5ab4b64e2840ac9c8886652c0ef5a244057d5b646" Nov 25 10:45:07 crc kubenswrapper[4821]: I1125 10:45:07.369670 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/f381f3655dcf4f93a6a93fe2921d660ff39f878b90de9b88a6ab8f46514g6tz" Nov 25 10:45:14 crc kubenswrapper[4821]: I1125 10:45:14.226853 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-7f8f598fc5-8gsxh"] Nov 25 10:45:14 crc kubenswrapper[4821]: E1125 10:45:14.227792 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd4ce1de-05e0-49f4-ad3a-7baae72f3e09" containerName="util" Nov 25 10:45:14 crc kubenswrapper[4821]: I1125 10:45:14.227810 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd4ce1de-05e0-49f4-ad3a-7baae72f3e09" containerName="util" Nov 25 10:45:14 crc kubenswrapper[4821]: E1125 10:45:14.227822 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd4ce1de-05e0-49f4-ad3a-7baae72f3e09" containerName="pull" Nov 25 10:45:14 crc kubenswrapper[4821]: I1125 10:45:14.227830 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd4ce1de-05e0-49f4-ad3a-7baae72f3e09" containerName="pull" Nov 25 10:45:14 crc kubenswrapper[4821]: E1125 10:45:14.227843 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52eea54a-9fe6-40ec-9354-8937bcc5a924" containerName="collect-profiles" Nov 25 10:45:14 crc kubenswrapper[4821]: I1125 10:45:14.227850 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="52eea54a-9fe6-40ec-9354-8937bcc5a924" containerName="collect-profiles" Nov 25 10:45:14 crc kubenswrapper[4821]: E1125 10:45:14.227873 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd4ce1de-05e0-49f4-ad3a-7baae72f3e09" containerName="extract" Nov 25 10:45:14 crc kubenswrapper[4821]: I1125 10:45:14.227881 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd4ce1de-05e0-49f4-ad3a-7baae72f3e09" containerName="extract" Nov 25 10:45:14 crc kubenswrapper[4821]: I1125 10:45:14.228057 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd4ce1de-05e0-49f4-ad3a-7baae72f3e09" containerName="extract" Nov 25 10:45:14 crc kubenswrapper[4821]: I1125 10:45:14.228082 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="52eea54a-9fe6-40ec-9354-8937bcc5a924" containerName="collect-profiles" Nov 25 10:45:14 crc kubenswrapper[4821]: I1125 10:45:14.228973 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-7f8f598fc5-8gsxh" Nov 25 10:45:14 crc kubenswrapper[4821]: I1125 10:45:14.231506 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-hl6lk" Nov 25 10:45:14 crc kubenswrapper[4821]: I1125 10:45:14.241023 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-7f8f598fc5-8gsxh"] Nov 25 10:45:14 crc kubenswrapper[4821]: I1125 10:45:14.391410 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4h9bh\" (UniqueName: \"kubernetes.io/projected/1875cb8b-6a5c-48dc-bf01-07eef1d129fe-kube-api-access-4h9bh\") pod \"openstack-operator-controller-operator-7f8f598fc5-8gsxh\" (UID: \"1875cb8b-6a5c-48dc-bf01-07eef1d129fe\") " pod="openstack-operators/openstack-operator-controller-operator-7f8f598fc5-8gsxh" Nov 25 10:45:14 crc kubenswrapper[4821]: I1125 10:45:14.492809 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4h9bh\" (UniqueName: \"kubernetes.io/projected/1875cb8b-6a5c-48dc-bf01-07eef1d129fe-kube-api-access-4h9bh\") pod \"openstack-operator-controller-operator-7f8f598fc5-8gsxh\" (UID: \"1875cb8b-6a5c-48dc-bf01-07eef1d129fe\") " pod="openstack-operators/openstack-operator-controller-operator-7f8f598fc5-8gsxh" Nov 25 10:45:14 crc kubenswrapper[4821]: I1125 10:45:14.526384 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4h9bh\" (UniqueName: \"kubernetes.io/projected/1875cb8b-6a5c-48dc-bf01-07eef1d129fe-kube-api-access-4h9bh\") pod \"openstack-operator-controller-operator-7f8f598fc5-8gsxh\" (UID: \"1875cb8b-6a5c-48dc-bf01-07eef1d129fe\") " pod="openstack-operators/openstack-operator-controller-operator-7f8f598fc5-8gsxh" Nov 25 10:45:14 crc kubenswrapper[4821]: I1125 10:45:14.550693 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-7f8f598fc5-8gsxh" Nov 25 10:45:15 crc kubenswrapper[4821]: I1125 10:45:15.051030 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-7f8f598fc5-8gsxh"] Nov 25 10:45:15 crc kubenswrapper[4821]: I1125 10:45:15.413170 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-7f8f598fc5-8gsxh" event={"ID":"1875cb8b-6a5c-48dc-bf01-07eef1d129fe","Type":"ContainerStarted","Data":"e6a2423fec99b22bc0f9af04cf71b19278fda69eaf404f8b7f0fcd7d835945a7"} Nov 25 10:45:18 crc kubenswrapper[4821]: I1125 10:45:18.440364 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-7f8f598fc5-8gsxh" event={"ID":"1875cb8b-6a5c-48dc-bf01-07eef1d129fe","Type":"ContainerStarted","Data":"49ed0339fef860230fba81192b42d2e12bbf17bbfa08431912b5e66190be3f49"} Nov 25 10:45:18 crc kubenswrapper[4821]: I1125 10:45:18.441916 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-7f8f598fc5-8gsxh" Nov 25 10:45:18 crc kubenswrapper[4821]: I1125 10:45:18.475717 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-7f8f598fc5-8gsxh" podStartSLOduration=1.3061882040000001 podStartE2EDuration="4.475701285s" podCreationTimestamp="2025-11-25 10:45:14 +0000 UTC" firstStartedPulling="2025-11-25 10:45:15.072841398 +0000 UTC m=+785.609161245" lastFinishedPulling="2025-11-25 10:45:18.242354479 +0000 UTC m=+788.778674326" observedRunningTime="2025-11-25 10:45:18.475510899 +0000 UTC m=+789.011830746" watchObservedRunningTime="2025-11-25 10:45:18.475701285 +0000 UTC m=+789.012021132" Nov 25 10:45:24 crc kubenswrapper[4821]: I1125 10:45:24.520337 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-h746x"] Nov 25 10:45:24 crc kubenswrapper[4821]: I1125 10:45:24.522878 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h746x" Nov 25 10:45:24 crc kubenswrapper[4821]: I1125 10:45:24.527206 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-h746x"] Nov 25 10:45:24 crc kubenswrapper[4821]: I1125 10:45:24.553425 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-7f8f598fc5-8gsxh" Nov 25 10:45:24 crc kubenswrapper[4821]: I1125 10:45:24.630900 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6e0d698-a0d5-41db-927f-28141c472a2e-catalog-content\") pod \"redhat-operators-h746x\" (UID: \"b6e0d698-a0d5-41db-927f-28141c472a2e\") " pod="openshift-marketplace/redhat-operators-h746x" Nov 25 10:45:24 crc kubenswrapper[4821]: I1125 10:45:24.631136 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k9glt\" (UniqueName: \"kubernetes.io/projected/b6e0d698-a0d5-41db-927f-28141c472a2e-kube-api-access-k9glt\") pod \"redhat-operators-h746x\" (UID: \"b6e0d698-a0d5-41db-927f-28141c472a2e\") " pod="openshift-marketplace/redhat-operators-h746x" Nov 25 10:45:24 crc kubenswrapper[4821]: I1125 10:45:24.631248 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6e0d698-a0d5-41db-927f-28141c472a2e-utilities\") pod \"redhat-operators-h746x\" (UID: \"b6e0d698-a0d5-41db-927f-28141c472a2e\") " pod="openshift-marketplace/redhat-operators-h746x" Nov 25 10:45:24 crc kubenswrapper[4821]: I1125 10:45:24.732788 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k9glt\" (UniqueName: \"kubernetes.io/projected/b6e0d698-a0d5-41db-927f-28141c472a2e-kube-api-access-k9glt\") pod \"redhat-operators-h746x\" (UID: \"b6e0d698-a0d5-41db-927f-28141c472a2e\") " pod="openshift-marketplace/redhat-operators-h746x" Nov 25 10:45:24 crc kubenswrapper[4821]: I1125 10:45:24.732854 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6e0d698-a0d5-41db-927f-28141c472a2e-utilities\") pod \"redhat-operators-h746x\" (UID: \"b6e0d698-a0d5-41db-927f-28141c472a2e\") " pod="openshift-marketplace/redhat-operators-h746x" Nov 25 10:45:24 crc kubenswrapper[4821]: I1125 10:45:24.732927 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6e0d698-a0d5-41db-927f-28141c472a2e-catalog-content\") pod \"redhat-operators-h746x\" (UID: \"b6e0d698-a0d5-41db-927f-28141c472a2e\") " pod="openshift-marketplace/redhat-operators-h746x" Nov 25 10:45:24 crc kubenswrapper[4821]: I1125 10:45:24.733410 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6e0d698-a0d5-41db-927f-28141c472a2e-utilities\") pod \"redhat-operators-h746x\" (UID: \"b6e0d698-a0d5-41db-927f-28141c472a2e\") " pod="openshift-marketplace/redhat-operators-h746x" Nov 25 10:45:24 crc kubenswrapper[4821]: I1125 10:45:24.733447 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6e0d698-a0d5-41db-927f-28141c472a2e-catalog-content\") pod \"redhat-operators-h746x\" (UID: \"b6e0d698-a0d5-41db-927f-28141c472a2e\") " pod="openshift-marketplace/redhat-operators-h746x" Nov 25 10:45:24 crc kubenswrapper[4821]: I1125 10:45:24.750707 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k9glt\" (UniqueName: \"kubernetes.io/projected/b6e0d698-a0d5-41db-927f-28141c472a2e-kube-api-access-k9glt\") pod \"redhat-operators-h746x\" (UID: \"b6e0d698-a0d5-41db-927f-28141c472a2e\") " pod="openshift-marketplace/redhat-operators-h746x" Nov 25 10:45:24 crc kubenswrapper[4821]: I1125 10:45:24.867485 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h746x" Nov 25 10:45:25 crc kubenswrapper[4821]: I1125 10:45:25.343634 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-h746x"] Nov 25 10:45:25 crc kubenswrapper[4821]: I1125 10:45:25.479535 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h746x" event={"ID":"b6e0d698-a0d5-41db-927f-28141c472a2e","Type":"ContainerStarted","Data":"a8af14287bfddffa1e9ca67c140596145333f0f8f2fca0d220f76405b367c621"} Nov 25 10:45:26 crc kubenswrapper[4821]: I1125 10:45:26.488137 4821 generic.go:334] "Generic (PLEG): container finished" podID="b6e0d698-a0d5-41db-927f-28141c472a2e" containerID="d5b599eca6f1562bdbd3c19d24cf87d97866bbccf7fbee2ae33e78fa16d465fb" exitCode=0 Nov 25 10:45:26 crc kubenswrapper[4821]: I1125 10:45:26.489195 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h746x" event={"ID":"b6e0d698-a0d5-41db-927f-28141c472a2e","Type":"ContainerDied","Data":"d5b599eca6f1562bdbd3c19d24cf87d97866bbccf7fbee2ae33e78fa16d465fb"} Nov 25 10:45:27 crc kubenswrapper[4821]: I1125 10:45:27.496584 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h746x" event={"ID":"b6e0d698-a0d5-41db-927f-28141c472a2e","Type":"ContainerStarted","Data":"de345d260cbfddfddbb930b42612293383691d1ce59a284df8638de4c2660d47"} Nov 25 10:45:28 crc kubenswrapper[4821]: I1125 10:45:28.502763 4821 generic.go:334] "Generic (PLEG): container finished" podID="b6e0d698-a0d5-41db-927f-28141c472a2e" containerID="de345d260cbfddfddbb930b42612293383691d1ce59a284df8638de4c2660d47" exitCode=0 Nov 25 10:45:28 crc kubenswrapper[4821]: I1125 10:45:28.502827 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h746x" event={"ID":"b6e0d698-a0d5-41db-927f-28141c472a2e","Type":"ContainerDied","Data":"de345d260cbfddfddbb930b42612293383691d1ce59a284df8638de4c2660d47"} Nov 25 10:45:29 crc kubenswrapper[4821]: I1125 10:45:29.511454 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h746x" event={"ID":"b6e0d698-a0d5-41db-927f-28141c472a2e","Type":"ContainerStarted","Data":"a945b539fd6e06e8c0eb4c31970784d78c24eaaf1c31330e60910a3263bfcc76"} Nov 25 10:45:29 crc kubenswrapper[4821]: I1125 10:45:29.536546 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-h746x" podStartSLOduration=3.087573496 podStartE2EDuration="5.536530779s" podCreationTimestamp="2025-11-25 10:45:24 +0000 UTC" firstStartedPulling="2025-11-25 10:45:26.490389047 +0000 UTC m=+797.026708904" lastFinishedPulling="2025-11-25 10:45:28.93934632 +0000 UTC m=+799.475666187" observedRunningTime="2025-11-25 10:45:29.532533528 +0000 UTC m=+800.068853375" watchObservedRunningTime="2025-11-25 10:45:29.536530779 +0000 UTC m=+800.072850626" Nov 25 10:45:34 crc kubenswrapper[4821]: I1125 10:45:34.714569 4821 patch_prober.go:28] interesting pod/machine-config-daemon-2krbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:45:34 crc kubenswrapper[4821]: I1125 10:45:34.715740 4821 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:45:34 crc kubenswrapper[4821]: I1125 10:45:34.715832 4821 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" Nov 25 10:45:34 crc kubenswrapper[4821]: I1125 10:45:34.717090 4821 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d231cd7c0910d85e280090ca76750286ce837c9e610ff6ebf83f2f5b9e078c96"} pod="openshift-machine-config-operator/machine-config-daemon-2krbf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 10:45:34 crc kubenswrapper[4821]: I1125 10:45:34.717224 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerName="machine-config-daemon" containerID="cri-o://d231cd7c0910d85e280090ca76750286ce837c9e610ff6ebf83f2f5b9e078c96" gracePeriod=600 Nov 25 10:45:34 crc kubenswrapper[4821]: I1125 10:45:34.868678 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-h746x" Nov 25 10:45:34 crc kubenswrapper[4821]: I1125 10:45:34.869046 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-h746x" Nov 25 10:45:34 crc kubenswrapper[4821]: I1125 10:45:34.918490 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-h746x" Nov 25 10:45:35 crc kubenswrapper[4821]: I1125 10:45:35.546861 4821 generic.go:334] "Generic (PLEG): container finished" podID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerID="d231cd7c0910d85e280090ca76750286ce837c9e610ff6ebf83f2f5b9e078c96" exitCode=0 Nov 25 10:45:35 crc kubenswrapper[4821]: I1125 10:45:35.546952 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" event={"ID":"5f948b87-ac86-4de6-ad64-c2ef947f84d4","Type":"ContainerDied","Data":"d231cd7c0910d85e280090ca76750286ce837c9e610ff6ebf83f2f5b9e078c96"} Nov 25 10:45:35 crc kubenswrapper[4821]: I1125 10:45:35.547282 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" event={"ID":"5f948b87-ac86-4de6-ad64-c2ef947f84d4","Type":"ContainerStarted","Data":"c1742b8b51dd841b8f5beefd1352b72fe8669e2095b0356503a696291b279fa5"} Nov 25 10:45:35 crc kubenswrapper[4821]: I1125 10:45:35.547315 4821 scope.go:117] "RemoveContainer" containerID="808518ddff79187bc281fef8f1f64e070e5abe86921d58acacca289efda8d1d4" Nov 25 10:45:35 crc kubenswrapper[4821]: I1125 10:45:35.594953 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-h746x" Nov 25 10:45:35 crc kubenswrapper[4821]: I1125 10:45:35.648508 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-h746x"] Nov 25 10:45:37 crc kubenswrapper[4821]: I1125 10:45:37.561653 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-h746x" podUID="b6e0d698-a0d5-41db-927f-28141c472a2e" containerName="registry-server" containerID="cri-o://a945b539fd6e06e8c0eb4c31970784d78c24eaaf1c31330e60910a3263bfcc76" gracePeriod=2 Nov 25 10:45:39 crc kubenswrapper[4821]: I1125 10:45:39.574261 4821 generic.go:334] "Generic (PLEG): container finished" podID="b6e0d698-a0d5-41db-927f-28141c472a2e" containerID="a945b539fd6e06e8c0eb4c31970784d78c24eaaf1c31330e60910a3263bfcc76" exitCode=0 Nov 25 10:45:39 crc kubenswrapper[4821]: I1125 10:45:39.574330 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h746x" event={"ID":"b6e0d698-a0d5-41db-927f-28141c472a2e","Type":"ContainerDied","Data":"a945b539fd6e06e8c0eb4c31970784d78c24eaaf1c31330e60910a3263bfcc76"} Nov 25 10:45:40 crc kubenswrapper[4821]: I1125 10:45:40.165106 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h746x" Nov 25 10:45:40 crc kubenswrapper[4821]: I1125 10:45:40.235448 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k9glt\" (UniqueName: \"kubernetes.io/projected/b6e0d698-a0d5-41db-927f-28141c472a2e-kube-api-access-k9glt\") pod \"b6e0d698-a0d5-41db-927f-28141c472a2e\" (UID: \"b6e0d698-a0d5-41db-927f-28141c472a2e\") " Nov 25 10:45:40 crc kubenswrapper[4821]: I1125 10:45:40.235712 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6e0d698-a0d5-41db-927f-28141c472a2e-catalog-content\") pod \"b6e0d698-a0d5-41db-927f-28141c472a2e\" (UID: \"b6e0d698-a0d5-41db-927f-28141c472a2e\") " Nov 25 10:45:40 crc kubenswrapper[4821]: I1125 10:45:40.235769 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6e0d698-a0d5-41db-927f-28141c472a2e-utilities\") pod \"b6e0d698-a0d5-41db-927f-28141c472a2e\" (UID: \"b6e0d698-a0d5-41db-927f-28141c472a2e\") " Nov 25 10:45:40 crc kubenswrapper[4821]: I1125 10:45:40.237014 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b6e0d698-a0d5-41db-927f-28141c472a2e-utilities" (OuterVolumeSpecName: "utilities") pod "b6e0d698-a0d5-41db-927f-28141c472a2e" (UID: "b6e0d698-a0d5-41db-927f-28141c472a2e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:45:40 crc kubenswrapper[4821]: I1125 10:45:40.242507 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6e0d698-a0d5-41db-927f-28141c472a2e-kube-api-access-k9glt" (OuterVolumeSpecName: "kube-api-access-k9glt") pod "b6e0d698-a0d5-41db-927f-28141c472a2e" (UID: "b6e0d698-a0d5-41db-927f-28141c472a2e"). InnerVolumeSpecName "kube-api-access-k9glt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:45:40 crc kubenswrapper[4821]: I1125 10:45:40.332390 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b6e0d698-a0d5-41db-927f-28141c472a2e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b6e0d698-a0d5-41db-927f-28141c472a2e" (UID: "b6e0d698-a0d5-41db-927f-28141c472a2e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:45:40 crc kubenswrapper[4821]: I1125 10:45:40.337145 4821 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6e0d698-a0d5-41db-927f-28141c472a2e-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:45:40 crc kubenswrapper[4821]: I1125 10:45:40.337190 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k9glt\" (UniqueName: \"kubernetes.io/projected/b6e0d698-a0d5-41db-927f-28141c472a2e-kube-api-access-k9glt\") on node \"crc\" DevicePath \"\"" Nov 25 10:45:40 crc kubenswrapper[4821]: I1125 10:45:40.337205 4821 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6e0d698-a0d5-41db-927f-28141c472a2e-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:45:40 crc kubenswrapper[4821]: I1125 10:45:40.585381 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h746x" event={"ID":"b6e0d698-a0d5-41db-927f-28141c472a2e","Type":"ContainerDied","Data":"a8af14287bfddffa1e9ca67c140596145333f0f8f2fca0d220f76405b367c621"} Nov 25 10:45:40 crc kubenswrapper[4821]: I1125 10:45:40.585479 4821 scope.go:117] "RemoveContainer" containerID="a945b539fd6e06e8c0eb4c31970784d78c24eaaf1c31330e60910a3263bfcc76" Nov 25 10:45:40 crc kubenswrapper[4821]: I1125 10:45:40.585475 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h746x" Nov 25 10:45:40 crc kubenswrapper[4821]: I1125 10:45:40.628617 4821 scope.go:117] "RemoveContainer" containerID="de345d260cbfddfddbb930b42612293383691d1ce59a284df8638de4c2660d47" Nov 25 10:45:40 crc kubenswrapper[4821]: I1125 10:45:40.633144 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-h746x"] Nov 25 10:45:40 crc kubenswrapper[4821]: I1125 10:45:40.640254 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-h746x"] Nov 25 10:45:40 crc kubenswrapper[4821]: I1125 10:45:40.656002 4821 scope.go:117] "RemoveContainer" containerID="d5b599eca6f1562bdbd3c19d24cf87d97866bbccf7fbee2ae33e78fa16d465fb" Nov 25 10:45:42 crc kubenswrapper[4821]: I1125 10:45:42.128141 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6e0d698-a0d5-41db-927f-28141c472a2e" path="/var/lib/kubelet/pods/b6e0d698-a0d5-41db-927f-28141c472a2e/volumes" Nov 25 10:45:46 crc kubenswrapper[4821]: I1125 10:45:46.877865 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zwqgf"] Nov 25 10:45:46 crc kubenswrapper[4821]: E1125 10:45:46.878728 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6e0d698-a0d5-41db-927f-28141c472a2e" containerName="registry-server" Nov 25 10:45:46 crc kubenswrapper[4821]: I1125 10:45:46.878742 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6e0d698-a0d5-41db-927f-28141c472a2e" containerName="registry-server" Nov 25 10:45:46 crc kubenswrapper[4821]: E1125 10:45:46.878760 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6e0d698-a0d5-41db-927f-28141c472a2e" containerName="extract-utilities" Nov 25 10:45:46 crc kubenswrapper[4821]: I1125 10:45:46.878767 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6e0d698-a0d5-41db-927f-28141c472a2e" containerName="extract-utilities" Nov 25 10:45:46 crc kubenswrapper[4821]: E1125 10:45:46.878786 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6e0d698-a0d5-41db-927f-28141c472a2e" containerName="extract-content" Nov 25 10:45:46 crc kubenswrapper[4821]: I1125 10:45:46.878793 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6e0d698-a0d5-41db-927f-28141c472a2e" containerName="extract-content" Nov 25 10:45:46 crc kubenswrapper[4821]: I1125 10:45:46.878922 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6e0d698-a0d5-41db-927f-28141c472a2e" containerName="registry-server" Nov 25 10:45:46 crc kubenswrapper[4821]: I1125 10:45:46.879873 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zwqgf" Nov 25 10:45:46 crc kubenswrapper[4821]: I1125 10:45:46.894656 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zwqgf"] Nov 25 10:45:47 crc kubenswrapper[4821]: I1125 10:45:47.027260 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd8b5398-7942-4c13-8840-a53ff1aefcbf-catalog-content\") pod \"community-operators-zwqgf\" (UID: \"fd8b5398-7942-4c13-8840-a53ff1aefcbf\") " pod="openshift-marketplace/community-operators-zwqgf" Nov 25 10:45:47 crc kubenswrapper[4821]: I1125 10:45:47.027321 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4dnkx\" (UniqueName: \"kubernetes.io/projected/fd8b5398-7942-4c13-8840-a53ff1aefcbf-kube-api-access-4dnkx\") pod \"community-operators-zwqgf\" (UID: \"fd8b5398-7942-4c13-8840-a53ff1aefcbf\") " pod="openshift-marketplace/community-operators-zwqgf" Nov 25 10:45:47 crc kubenswrapper[4821]: I1125 10:45:47.027423 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd8b5398-7942-4c13-8840-a53ff1aefcbf-utilities\") pod \"community-operators-zwqgf\" (UID: \"fd8b5398-7942-4c13-8840-a53ff1aefcbf\") " pod="openshift-marketplace/community-operators-zwqgf" Nov 25 10:45:47 crc kubenswrapper[4821]: I1125 10:45:47.141491 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd8b5398-7942-4c13-8840-a53ff1aefcbf-utilities\") pod \"community-operators-zwqgf\" (UID: \"fd8b5398-7942-4c13-8840-a53ff1aefcbf\") " pod="openshift-marketplace/community-operators-zwqgf" Nov 25 10:45:47 crc kubenswrapper[4821]: I1125 10:45:47.141651 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd8b5398-7942-4c13-8840-a53ff1aefcbf-catalog-content\") pod \"community-operators-zwqgf\" (UID: \"fd8b5398-7942-4c13-8840-a53ff1aefcbf\") " pod="openshift-marketplace/community-operators-zwqgf" Nov 25 10:45:47 crc kubenswrapper[4821]: I1125 10:45:47.141691 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4dnkx\" (UniqueName: \"kubernetes.io/projected/fd8b5398-7942-4c13-8840-a53ff1aefcbf-kube-api-access-4dnkx\") pod \"community-operators-zwqgf\" (UID: \"fd8b5398-7942-4c13-8840-a53ff1aefcbf\") " pod="openshift-marketplace/community-operators-zwqgf" Nov 25 10:45:47 crc kubenswrapper[4821]: I1125 10:45:47.142501 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd8b5398-7942-4c13-8840-a53ff1aefcbf-utilities\") pod \"community-operators-zwqgf\" (UID: \"fd8b5398-7942-4c13-8840-a53ff1aefcbf\") " pod="openshift-marketplace/community-operators-zwqgf" Nov 25 10:45:47 crc kubenswrapper[4821]: I1125 10:45:47.142739 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd8b5398-7942-4c13-8840-a53ff1aefcbf-catalog-content\") pod \"community-operators-zwqgf\" (UID: \"fd8b5398-7942-4c13-8840-a53ff1aefcbf\") " pod="openshift-marketplace/community-operators-zwqgf" Nov 25 10:45:47 crc kubenswrapper[4821]: I1125 10:45:47.160747 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4dnkx\" (UniqueName: \"kubernetes.io/projected/fd8b5398-7942-4c13-8840-a53ff1aefcbf-kube-api-access-4dnkx\") pod \"community-operators-zwqgf\" (UID: \"fd8b5398-7942-4c13-8840-a53ff1aefcbf\") " pod="openshift-marketplace/community-operators-zwqgf" Nov 25 10:45:47 crc kubenswrapper[4821]: I1125 10:45:47.198438 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zwqgf" Nov 25 10:45:47 crc kubenswrapper[4821]: I1125 10:45:47.720013 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zwqgf"] Nov 25 10:45:47 crc kubenswrapper[4821]: W1125 10:45:47.729063 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfd8b5398_7942_4c13_8840_a53ff1aefcbf.slice/crio-e2067c3b5ea8206bc60ea0fa8a19321b79672783f08a29cb2f178e3cbb9976a0 WatchSource:0}: Error finding container e2067c3b5ea8206bc60ea0fa8a19321b79672783f08a29cb2f178e3cbb9976a0: Status 404 returned error can't find the container with id e2067c3b5ea8206bc60ea0fa8a19321b79672783f08a29cb2f178e3cbb9976a0 Nov 25 10:45:48 crc kubenswrapper[4821]: I1125 10:45:48.642370 4821 generic.go:334] "Generic (PLEG): container finished" podID="fd8b5398-7942-4c13-8840-a53ff1aefcbf" containerID="ca0979132b74469fa866210e56bd7eab29e532a6091a6be469b8fc79fbeb18c3" exitCode=0 Nov 25 10:45:48 crc kubenswrapper[4821]: I1125 10:45:48.642470 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zwqgf" event={"ID":"fd8b5398-7942-4c13-8840-a53ff1aefcbf","Type":"ContainerDied","Data":"ca0979132b74469fa866210e56bd7eab29e532a6091a6be469b8fc79fbeb18c3"} Nov 25 10:45:48 crc kubenswrapper[4821]: I1125 10:45:48.642693 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zwqgf" event={"ID":"fd8b5398-7942-4c13-8840-a53ff1aefcbf","Type":"ContainerStarted","Data":"e2067c3b5ea8206bc60ea0fa8a19321b79672783f08a29cb2f178e3cbb9976a0"} Nov 25 10:45:49 crc kubenswrapper[4821]: I1125 10:45:49.652820 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zwqgf" event={"ID":"fd8b5398-7942-4c13-8840-a53ff1aefcbf","Type":"ContainerStarted","Data":"2f9df2d4912b0774f58480ec42bf51a82a08a6e4ef317765ec4f4f582e28b25e"} Nov 25 10:45:50 crc kubenswrapper[4821]: I1125 10:45:50.659467 4821 generic.go:334] "Generic (PLEG): container finished" podID="fd8b5398-7942-4c13-8840-a53ff1aefcbf" containerID="2f9df2d4912b0774f58480ec42bf51a82a08a6e4ef317765ec4f4f582e28b25e" exitCode=0 Nov 25 10:45:50 crc kubenswrapper[4821]: I1125 10:45:50.659506 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zwqgf" event={"ID":"fd8b5398-7942-4c13-8840-a53ff1aefcbf","Type":"ContainerDied","Data":"2f9df2d4912b0774f58480ec42bf51a82a08a6e4ef317765ec4f4f582e28b25e"} Nov 25 10:45:51 crc kubenswrapper[4821]: I1125 10:45:51.666468 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zwqgf" event={"ID":"fd8b5398-7942-4c13-8840-a53ff1aefcbf","Type":"ContainerStarted","Data":"33cb6765f313a785652c5ee17c60ad16e2235717c721efbf3604550bf7e9d3bc"} Nov 25 10:45:51 crc kubenswrapper[4821]: I1125 10:45:51.695426 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zwqgf" podStartSLOduration=3.290378932 podStartE2EDuration="5.695402232s" podCreationTimestamp="2025-11-25 10:45:46 +0000 UTC" firstStartedPulling="2025-11-25 10:45:48.643813291 +0000 UTC m=+819.180133138" lastFinishedPulling="2025-11-25 10:45:51.048836591 +0000 UTC m=+821.585156438" observedRunningTime="2025-11-25 10:45:51.690362754 +0000 UTC m=+822.226682611" watchObservedRunningTime="2025-11-25 10:45:51.695402232 +0000 UTC m=+822.231722089" Nov 25 10:45:54 crc kubenswrapper[4821]: I1125 10:45:54.258608 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-b248b"] Nov 25 10:45:54 crc kubenswrapper[4821]: I1125 10:45:54.260577 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b248b" Nov 25 10:45:54 crc kubenswrapper[4821]: I1125 10:45:54.275997 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-b248b"] Nov 25 10:45:54 crc kubenswrapper[4821]: I1125 10:45:54.335885 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3efc5d6f-66aa-409b-adac-506291c1e810-utilities\") pod \"redhat-marketplace-b248b\" (UID: \"3efc5d6f-66aa-409b-adac-506291c1e810\") " pod="openshift-marketplace/redhat-marketplace-b248b" Nov 25 10:45:54 crc kubenswrapper[4821]: I1125 10:45:54.335939 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vv8ng\" (UniqueName: \"kubernetes.io/projected/3efc5d6f-66aa-409b-adac-506291c1e810-kube-api-access-vv8ng\") pod \"redhat-marketplace-b248b\" (UID: \"3efc5d6f-66aa-409b-adac-506291c1e810\") " pod="openshift-marketplace/redhat-marketplace-b248b" Nov 25 10:45:54 crc kubenswrapper[4821]: I1125 10:45:54.336093 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3efc5d6f-66aa-409b-adac-506291c1e810-catalog-content\") pod \"redhat-marketplace-b248b\" (UID: \"3efc5d6f-66aa-409b-adac-506291c1e810\") " pod="openshift-marketplace/redhat-marketplace-b248b" Nov 25 10:45:54 crc kubenswrapper[4821]: I1125 10:45:54.437874 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3efc5d6f-66aa-409b-adac-506291c1e810-utilities\") pod \"redhat-marketplace-b248b\" (UID: \"3efc5d6f-66aa-409b-adac-506291c1e810\") " pod="openshift-marketplace/redhat-marketplace-b248b" Nov 25 10:45:54 crc kubenswrapper[4821]: I1125 10:45:54.437946 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vv8ng\" (UniqueName: \"kubernetes.io/projected/3efc5d6f-66aa-409b-adac-506291c1e810-kube-api-access-vv8ng\") pod \"redhat-marketplace-b248b\" (UID: \"3efc5d6f-66aa-409b-adac-506291c1e810\") " pod="openshift-marketplace/redhat-marketplace-b248b" Nov 25 10:45:54 crc kubenswrapper[4821]: I1125 10:45:54.437992 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3efc5d6f-66aa-409b-adac-506291c1e810-catalog-content\") pod \"redhat-marketplace-b248b\" (UID: \"3efc5d6f-66aa-409b-adac-506291c1e810\") " pod="openshift-marketplace/redhat-marketplace-b248b" Nov 25 10:45:54 crc kubenswrapper[4821]: I1125 10:45:54.438625 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3efc5d6f-66aa-409b-adac-506291c1e810-catalog-content\") pod \"redhat-marketplace-b248b\" (UID: \"3efc5d6f-66aa-409b-adac-506291c1e810\") " pod="openshift-marketplace/redhat-marketplace-b248b" Nov 25 10:45:54 crc kubenswrapper[4821]: I1125 10:45:54.438779 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3efc5d6f-66aa-409b-adac-506291c1e810-utilities\") pod \"redhat-marketplace-b248b\" (UID: \"3efc5d6f-66aa-409b-adac-506291c1e810\") " pod="openshift-marketplace/redhat-marketplace-b248b" Nov 25 10:45:54 crc kubenswrapper[4821]: I1125 10:45:54.455306 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vv8ng\" (UniqueName: \"kubernetes.io/projected/3efc5d6f-66aa-409b-adac-506291c1e810-kube-api-access-vv8ng\") pod \"redhat-marketplace-b248b\" (UID: \"3efc5d6f-66aa-409b-adac-506291c1e810\") " pod="openshift-marketplace/redhat-marketplace-b248b" Nov 25 10:45:54 crc kubenswrapper[4821]: I1125 10:45:54.583242 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b248b" Nov 25 10:45:55 crc kubenswrapper[4821]: I1125 10:45:55.104437 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-b248b"] Nov 25 10:45:55 crc kubenswrapper[4821]: W1125 10:45:55.110782 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3efc5d6f_66aa_409b_adac_506291c1e810.slice/crio-e9b99fc6b314e18855f8e58f0c1b09bc440ef46e676efed63f5309df1e62c19d WatchSource:0}: Error finding container e9b99fc6b314e18855f8e58f0c1b09bc440ef46e676efed63f5309df1e62c19d: Status 404 returned error can't find the container with id e9b99fc6b314e18855f8e58f0c1b09bc440ef46e676efed63f5309df1e62c19d Nov 25 10:45:55 crc kubenswrapper[4821]: I1125 10:45:55.711524 4821 generic.go:334] "Generic (PLEG): container finished" podID="3efc5d6f-66aa-409b-adac-506291c1e810" containerID="f72d68759c12fed646c58d3fb778ab5e3e3dbc713cc2216b1873108a467d4080" exitCode=0 Nov 25 10:45:55 crc kubenswrapper[4821]: I1125 10:45:55.711924 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b248b" event={"ID":"3efc5d6f-66aa-409b-adac-506291c1e810","Type":"ContainerDied","Data":"f72d68759c12fed646c58d3fb778ab5e3e3dbc713cc2216b1873108a467d4080"} Nov 25 10:45:55 crc kubenswrapper[4821]: I1125 10:45:55.711984 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b248b" event={"ID":"3efc5d6f-66aa-409b-adac-506291c1e810","Type":"ContainerStarted","Data":"e9b99fc6b314e18855f8e58f0c1b09bc440ef46e676efed63f5309df1e62c19d"} Nov 25 10:45:56 crc kubenswrapper[4821]: I1125 10:45:56.723470 4821 generic.go:334] "Generic (PLEG): container finished" podID="3efc5d6f-66aa-409b-adac-506291c1e810" containerID="725d705414501cce458da9324ba24945e4c24c36c1b5574625b32bff21475e53" exitCode=0 Nov 25 10:45:56 crc kubenswrapper[4821]: I1125 10:45:56.723569 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b248b" event={"ID":"3efc5d6f-66aa-409b-adac-506291c1e810","Type":"ContainerDied","Data":"725d705414501cce458da9324ba24945e4c24c36c1b5574625b32bff21475e53"} Nov 25 10:45:57 crc kubenswrapper[4821]: I1125 10:45:57.198624 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zwqgf" Nov 25 10:45:57 crc kubenswrapper[4821]: I1125 10:45:57.198685 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zwqgf" Nov 25 10:45:57 crc kubenswrapper[4821]: I1125 10:45:57.241667 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zwqgf" Nov 25 10:45:57 crc kubenswrapper[4821]: I1125 10:45:57.732240 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b248b" event={"ID":"3efc5d6f-66aa-409b-adac-506291c1e810","Type":"ContainerStarted","Data":"355abcf932cff05e1a1aede668d3afcfd531d86f0aa99df4c1cf920c0e9642ff"} Nov 25 10:45:57 crc kubenswrapper[4821]: I1125 10:45:57.801844 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zwqgf" Nov 25 10:45:57 crc kubenswrapper[4821]: I1125 10:45:57.820588 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-b248b" podStartSLOduration=2.403113486 podStartE2EDuration="3.82057089s" podCreationTimestamp="2025-11-25 10:45:54 +0000 UTC" firstStartedPulling="2025-11-25 10:45:55.720146652 +0000 UTC m=+826.256466499" lastFinishedPulling="2025-11-25 10:45:57.137604056 +0000 UTC m=+827.673923903" observedRunningTime="2025-11-25 10:45:57.761283447 +0000 UTC m=+828.297603294" watchObservedRunningTime="2025-11-25 10:45:57.82057089 +0000 UTC m=+828.356890737" Nov 25 10:45:58 crc kubenswrapper[4821]: I1125 10:45:58.830032 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-86dc4d89c8-pnq4j"] Nov 25 10:45:58 crc kubenswrapper[4821]: I1125 10:45:58.831571 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-pnq4j" Nov 25 10:45:58 crc kubenswrapper[4821]: I1125 10:45:58.836620 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-tmdtj" Nov 25 10:45:58 crc kubenswrapper[4821]: I1125 10:45:58.839006 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-79856dc55c-p4bwc"] Nov 25 10:45:58 crc kubenswrapper[4821]: I1125 10:45:58.840586 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-p4bwc" Nov 25 10:45:58 crc kubenswrapper[4821]: I1125 10:45:58.842303 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-w42n8" Nov 25 10:45:58 crc kubenswrapper[4821]: I1125 10:45:58.856032 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-7d695c9b56-7sx9m"] Nov 25 10:45:58 crc kubenswrapper[4821]: I1125 10:45:58.857019 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-7sx9m" Nov 25 10:45:58 crc kubenswrapper[4821]: I1125 10:45:58.859502 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-lg752" Nov 25 10:45:58 crc kubenswrapper[4821]: I1125 10:45:58.871328 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-86dc4d89c8-pnq4j"] Nov 25 10:45:58 crc kubenswrapper[4821]: I1125 10:45:58.876259 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-7d695c9b56-7sx9m"] Nov 25 10:45:58 crc kubenswrapper[4821]: I1125 10:45:58.884441 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-79856dc55c-p4bwc"] Nov 25 10:45:58 crc kubenswrapper[4821]: I1125 10:45:58.895703 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6sbqh\" (UniqueName: \"kubernetes.io/projected/8f16d64f-ea14-4853-b30b-2df294ce6cc3-kube-api-access-6sbqh\") pod \"cinder-operator-controller-manager-79856dc55c-p4bwc\" (UID: \"8f16d64f-ea14-4853-b30b-2df294ce6cc3\") " pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-p4bwc" Nov 25 10:45:58 crc kubenswrapper[4821]: I1125 10:45:58.895818 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrhtz\" (UniqueName: \"kubernetes.io/projected/db28ec64-d8be-4b56-a427-e4cceb256fd0-kube-api-access-qrhtz\") pod \"barbican-operator-controller-manager-86dc4d89c8-pnq4j\" (UID: \"db28ec64-d8be-4b56-a427-e4cceb256fd0\") " pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-pnq4j" Nov 25 10:45:58 crc kubenswrapper[4821]: I1125 10:45:58.899053 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-68b95954c9-qgs92"] Nov 25 10:45:58 crc kubenswrapper[4821]: I1125 10:45:58.900347 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-68b95954c9-qgs92" Nov 25 10:45:58 crc kubenswrapper[4821]: I1125 10:45:58.903524 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-r4nbb" Nov 25 10:45:58 crc kubenswrapper[4821]: I1125 10:45:58.927344 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-774b86978c-s49f6"] Nov 25 10:45:58 crc kubenswrapper[4821]: I1125 10:45:58.928495 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-774b86978c-s49f6" Nov 25 10:45:58 crc kubenswrapper[4821]: I1125 10:45:58.930865 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-rbxx5" Nov 25 10:45:58 crc kubenswrapper[4821]: I1125 10:45:58.935139 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c9694994-2qzrg"] Nov 25 10:45:58 crc kubenswrapper[4821]: I1125 10:45:58.945834 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-2qzrg" Nov 25 10:45:58 crc kubenswrapper[4821]: I1125 10:45:58.948605 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-626rt" Nov 25 10:45:58 crc kubenswrapper[4821]: I1125 10:45:58.956931 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-68b95954c9-qgs92"] Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.001666 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6sbqh\" (UniqueName: \"kubernetes.io/projected/8f16d64f-ea14-4853-b30b-2df294ce6cc3-kube-api-access-6sbqh\") pod \"cinder-operator-controller-manager-79856dc55c-p4bwc\" (UID: \"8f16d64f-ea14-4853-b30b-2df294ce6cc3\") " pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-p4bwc" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.001738 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwxfn\" (UniqueName: \"kubernetes.io/projected/b7895321-27ec-4c75-aa5e-581fe46e0789-kube-api-access-gwxfn\") pod \"horizon-operator-controller-manager-68c9694994-2qzrg\" (UID: \"b7895321-27ec-4c75-aa5e-581fe46e0789\") " pod="openstack-operators/horizon-operator-controller-manager-68c9694994-2qzrg" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.001766 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrhtz\" (UniqueName: \"kubernetes.io/projected/db28ec64-d8be-4b56-a427-e4cceb256fd0-kube-api-access-qrhtz\") pod \"barbican-operator-controller-manager-86dc4d89c8-pnq4j\" (UID: \"db28ec64-d8be-4b56-a427-e4cceb256fd0\") " pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-pnq4j" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.001792 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-km7xn\" (UniqueName: \"kubernetes.io/projected/b35e42b0-2a6f-454f-b8e2-4f7cd9b9217f-kube-api-access-km7xn\") pod \"heat-operator-controller-manager-774b86978c-s49f6\" (UID: \"b35e42b0-2a6f-454f-b8e2-4f7cd9b9217f\") " pod="openstack-operators/heat-operator-controller-manager-774b86978c-s49f6" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.001850 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tvlxr\" (UniqueName: \"kubernetes.io/projected/2236607e-2a7b-4858-b5b9-c43e47c9375d-kube-api-access-tvlxr\") pod \"designate-operator-controller-manager-7d695c9b56-7sx9m\" (UID: \"2236607e-2a7b-4858-b5b9-c43e47c9375d\") " pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-7sx9m" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.001880 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lpdb2\" (UniqueName: \"kubernetes.io/projected/a3dc5167-4997-44bf-ace3-706829fd0f17-kube-api-access-lpdb2\") pod \"glance-operator-controller-manager-68b95954c9-qgs92\" (UID: \"a3dc5167-4997-44bf-ace3-706829fd0f17\") " pod="openstack-operators/glance-operator-controller-manager-68b95954c9-qgs92" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.013321 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-774b86978c-s49f6"] Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.039251 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-d5cc86f4b-8cqfp"] Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.040611 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-8cqfp" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.043511 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.043611 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-t64v2" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.045590 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6sbqh\" (UniqueName: \"kubernetes.io/projected/8f16d64f-ea14-4853-b30b-2df294ce6cc3-kube-api-access-6sbqh\") pod \"cinder-operator-controller-manager-79856dc55c-p4bwc\" (UID: \"8f16d64f-ea14-4853-b30b-2df294ce6cc3\") " pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-p4bwc" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.047663 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrhtz\" (UniqueName: \"kubernetes.io/projected/db28ec64-d8be-4b56-a427-e4cceb256fd0-kube-api-access-qrhtz\") pod \"barbican-operator-controller-manager-86dc4d89c8-pnq4j\" (UID: \"db28ec64-d8be-4b56-a427-e4cceb256fd0\") " pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-pnq4j" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.051449 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-d5cc86f4b-8cqfp"] Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.069257 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c9694994-2qzrg"] Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.071611 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5bfcdc958c-fqfwh"] Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.072658 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-fqfwh" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.079826 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-748dc6576f-f8tvv"] Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.080870 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-f8tvv" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.082432 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-zjwrg" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.088507 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-kx65r" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.090968 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-58bb8d67cc-vnf9x"] Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.092096 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5bfcdc958c-fqfwh"] Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.092194 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-vnf9x" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.095748 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-xxfmg" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.102858 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-748dc6576f-f8tvv"] Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.103191 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwxfn\" (UniqueName: \"kubernetes.io/projected/b7895321-27ec-4c75-aa5e-581fe46e0789-kube-api-access-gwxfn\") pod \"horizon-operator-controller-manager-68c9694994-2qzrg\" (UID: \"b7895321-27ec-4c75-aa5e-581fe46e0789\") " pod="openstack-operators/horizon-operator-controller-manager-68c9694994-2qzrg" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.103244 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxbmr\" (UniqueName: \"kubernetes.io/projected/7982734e-b887-4cea-9724-3a37f9369ac5-kube-api-access-qxbmr\") pod \"keystone-operator-controller-manager-748dc6576f-f8tvv\" (UID: \"7982734e-b887-4cea-9724-3a37f9369ac5\") " pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-f8tvv" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.103278 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-km7xn\" (UniqueName: \"kubernetes.io/projected/b35e42b0-2a6f-454f-b8e2-4f7cd9b9217f-kube-api-access-km7xn\") pod \"heat-operator-controller-manager-774b86978c-s49f6\" (UID: \"b35e42b0-2a6f-454f-b8e2-4f7cd9b9217f\") " pod="openstack-operators/heat-operator-controller-manager-774b86978c-s49f6" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.103323 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tvlxr\" (UniqueName: \"kubernetes.io/projected/2236607e-2a7b-4858-b5b9-c43e47c9375d-kube-api-access-tvlxr\") pod \"designate-operator-controller-manager-7d695c9b56-7sx9m\" (UID: \"2236607e-2a7b-4858-b5b9-c43e47c9375d\") " pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-7sx9m" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.103353 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lpdb2\" (UniqueName: \"kubernetes.io/projected/a3dc5167-4997-44bf-ace3-706829fd0f17-kube-api-access-lpdb2\") pod \"glance-operator-controller-manager-68b95954c9-qgs92\" (UID: \"a3dc5167-4997-44bf-ace3-706829fd0f17\") " pod="openstack-operators/glance-operator-controller-manager-68b95954c9-qgs92" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.103383 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k6lm5\" (UniqueName: \"kubernetes.io/projected/1976151f-2b7a-4cde-95d3-9020b17d938c-kube-api-access-k6lm5\") pod \"ironic-operator-controller-manager-5bfcdc958c-fqfwh\" (UID: \"1976151f-2b7a-4cde-95d3-9020b17d938c\") " pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-fqfwh" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.103413 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9db84cac-a8df-4359-8051-ebc064a97b4d-cert\") pod \"infra-operator-controller-manager-d5cc86f4b-8cqfp\" (UID: \"9db84cac-a8df-4359-8051-ebc064a97b4d\") " pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-8cqfp" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.103444 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fq5sc\" (UniqueName: \"kubernetes.io/projected/9db84cac-a8df-4359-8051-ebc064a97b4d-kube-api-access-fq5sc\") pod \"infra-operator-controller-manager-d5cc86f4b-8cqfp\" (UID: \"9db84cac-a8df-4359-8051-ebc064a97b4d\") " pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-8cqfp" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.103481 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cnjb7\" (UniqueName: \"kubernetes.io/projected/5fefd5f2-18cb-4c63-848f-2fc9d63bb828-kube-api-access-cnjb7\") pod \"manila-operator-controller-manager-58bb8d67cc-vnf9x\" (UID: \"5fefd5f2-18cb-4c63-848f-2fc9d63bb828\") " pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-vnf9x" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.123435 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-58bb8d67cc-vnf9x"] Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.138261 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-fw96v"] Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.140691 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-fw96v" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.149475 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-pnq4j" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.150038 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwxfn\" (UniqueName: \"kubernetes.io/projected/b7895321-27ec-4c75-aa5e-581fe46e0789-kube-api-access-gwxfn\") pod \"horizon-operator-controller-manager-68c9694994-2qzrg\" (UID: \"b7895321-27ec-4c75-aa5e-581fe46e0789\") " pod="openstack-operators/horizon-operator-controller-manager-68c9694994-2qzrg" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.151104 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tvlxr\" (UniqueName: \"kubernetes.io/projected/2236607e-2a7b-4858-b5b9-c43e47c9375d-kube-api-access-tvlxr\") pod \"designate-operator-controller-manager-7d695c9b56-7sx9m\" (UID: \"2236607e-2a7b-4858-b5b9-c43e47c9375d\") " pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-7sx9m" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.152082 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-9wszx" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.152774 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-km7xn\" (UniqueName: \"kubernetes.io/projected/b35e42b0-2a6f-454f-b8e2-4f7cd9b9217f-kube-api-access-km7xn\") pod \"heat-operator-controller-manager-774b86978c-s49f6\" (UID: \"b35e42b0-2a6f-454f-b8e2-4f7cd9b9217f\") " pod="openstack-operators/heat-operator-controller-manager-774b86978c-s49f6" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.160800 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-fw96v"] Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.161182 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-p4bwc" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.164660 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lpdb2\" (UniqueName: \"kubernetes.io/projected/a3dc5167-4997-44bf-ace3-706829fd0f17-kube-api-access-lpdb2\") pod \"glance-operator-controller-manager-68b95954c9-qgs92\" (UID: \"a3dc5167-4997-44bf-ace3-706829fd0f17\") " pod="openstack-operators/glance-operator-controller-manager-68b95954c9-qgs92" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.173214 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-7sx9m" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.181158 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-hwp4x"] Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.182219 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-hwp4x" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.184836 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-62vqh" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.193202 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-hwp4x"] Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.198203 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-fd75fd47d-5ncqq"] Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.199438 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-5ncqq" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.200970 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-6lj6j" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.206869 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j4f9s\" (UniqueName: \"kubernetes.io/projected/d42d0991-9e6f-46d0-9236-00e85c26a603-kube-api-access-j4f9s\") pod \"mariadb-operator-controller-manager-cb6c4fdb7-fw96v\" (UID: \"d42d0991-9e6f-46d0-9236-00e85c26a603\") " pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-fw96v" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.207741 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxbmr\" (UniqueName: \"kubernetes.io/projected/7982734e-b887-4cea-9724-3a37f9369ac5-kube-api-access-qxbmr\") pod \"keystone-operator-controller-manager-748dc6576f-f8tvv\" (UID: \"7982734e-b887-4cea-9724-3a37f9369ac5\") " pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-f8tvv" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.207981 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zj6np\" (UniqueName: \"kubernetes.io/projected/db1d55a6-1285-429d-b9dd-b2e100e602c6-kube-api-access-zj6np\") pod \"octavia-operator-controller-manager-fd75fd47d-5ncqq\" (UID: \"db1d55a6-1285-429d-b9dd-b2e100e602c6\") " pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-5ncqq" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.208056 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k6lm5\" (UniqueName: \"kubernetes.io/projected/1976151f-2b7a-4cde-95d3-9020b17d938c-kube-api-access-k6lm5\") pod \"ironic-operator-controller-manager-5bfcdc958c-fqfwh\" (UID: \"1976151f-2b7a-4cde-95d3-9020b17d938c\") " pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-fqfwh" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.208088 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9db84cac-a8df-4359-8051-ebc064a97b4d-cert\") pod \"infra-operator-controller-manager-d5cc86f4b-8cqfp\" (UID: \"9db84cac-a8df-4359-8051-ebc064a97b4d\") " pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-8cqfp" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.208120 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fq5sc\" (UniqueName: \"kubernetes.io/projected/9db84cac-a8df-4359-8051-ebc064a97b4d-kube-api-access-fq5sc\") pod \"infra-operator-controller-manager-d5cc86f4b-8cqfp\" (UID: \"9db84cac-a8df-4359-8051-ebc064a97b4d\") " pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-8cqfp" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.208191 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cnjb7\" (UniqueName: \"kubernetes.io/projected/5fefd5f2-18cb-4c63-848f-2fc9d63bb828-kube-api-access-cnjb7\") pod \"manila-operator-controller-manager-58bb8d67cc-vnf9x\" (UID: \"5fefd5f2-18cb-4c63-848f-2fc9d63bb828\") " pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-vnf9x" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.208226 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s7ldk\" (UniqueName: \"kubernetes.io/projected/c0376a5f-ed42-4399-8f41-2d2157a654fd-kube-api-access-s7ldk\") pod \"neutron-operator-controller-manager-7c57c8bbc4-hwp4x\" (UID: \"c0376a5f-ed42-4399-8f41-2d2157a654fd\") " pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-hwp4x" Nov 25 10:45:59 crc kubenswrapper[4821]: E1125 10:45:59.211928 4821 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 25 10:45:59 crc kubenswrapper[4821]: E1125 10:45:59.212008 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9db84cac-a8df-4359-8051-ebc064a97b4d-cert podName:9db84cac-a8df-4359-8051-ebc064a97b4d nodeName:}" failed. No retries permitted until 2025-11-25 10:45:59.711984934 +0000 UTC m=+830.248304781 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/9db84cac-a8df-4359-8051-ebc064a97b4d-cert") pod "infra-operator-controller-manager-d5cc86f4b-8cqfp" (UID: "9db84cac-a8df-4359-8051-ebc064a97b4d") : secret "infra-operator-webhook-server-cert" not found Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.216547 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-68b95954c9-qgs92" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.235860 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxbmr\" (UniqueName: \"kubernetes.io/projected/7982734e-b887-4cea-9724-3a37f9369ac5-kube-api-access-qxbmr\") pod \"keystone-operator-controller-manager-748dc6576f-f8tvv\" (UID: \"7982734e-b887-4cea-9724-3a37f9369ac5\") " pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-f8tvv" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.235935 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-fd75fd47d-5ncqq"] Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.248548 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k6lm5\" (UniqueName: \"kubernetes.io/projected/1976151f-2b7a-4cde-95d3-9020b17d938c-kube-api-access-k6lm5\") pod \"ironic-operator-controller-manager-5bfcdc958c-fqfwh\" (UID: \"1976151f-2b7a-4cde-95d3-9020b17d938c\") " pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-fqfwh" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.249943 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-79556f57fc-4ftbp"] Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.251707 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-4ftbp" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.256907 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-jdp2z" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.260731 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-lwp9z"] Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.261789 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-lwp9z" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.263528 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.263689 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-qff2t" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.277193 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-79556f57fc-4ftbp"] Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.281974 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fq5sc\" (UniqueName: \"kubernetes.io/projected/9db84cac-a8df-4359-8051-ebc064a97b4d-kube-api-access-fq5sc\") pod \"infra-operator-controller-manager-d5cc86f4b-8cqfp\" (UID: \"9db84cac-a8df-4359-8051-ebc064a97b4d\") " pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-8cqfp" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.282379 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-774b86978c-s49f6" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.283255 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cnjb7\" (UniqueName: \"kubernetes.io/projected/5fefd5f2-18cb-4c63-848f-2fc9d63bb828-kube-api-access-cnjb7\") pod \"manila-operator-controller-manager-58bb8d67cc-vnf9x\" (UID: \"5fefd5f2-18cb-4c63-848f-2fc9d63bb828\") " pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-vnf9x" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.290504 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-66cf5c67ff-qxgcp"] Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.295444 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-2qzrg" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.300636 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-qxgcp" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.303431 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-2tbf6" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.315262 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-5db546f9d9-66kf4"] Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.316855 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-66kf4" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.321252 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/14385dd9-9b06-473b-919d-9d2f9207ffe1-cert\") pod \"openstack-baremetal-operator-controller-manager-544b9bb9-lwp9z\" (UID: \"14385dd9-9b06-473b-919d-9d2f9207ffe1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-lwp9z" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.321302 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-flgg6\" (UniqueName: \"kubernetes.io/projected/1f69797f-4204-4989-a00b-d9722d44d3c6-kube-api-access-flgg6\") pod \"nova-operator-controller-manager-79556f57fc-4ftbp\" (UID: \"1f69797f-4204-4989-a00b-d9722d44d3c6\") " pod="openstack-operators/nova-operator-controller-manager-79556f57fc-4ftbp" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.321332 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ntb45\" (UniqueName: \"kubernetes.io/projected/2b44040a-e224-4468-a528-9f96a98e3891-kube-api-access-ntb45\") pod \"ovn-operator-controller-manager-66cf5c67ff-qxgcp\" (UID: \"2b44040a-e224-4468-a528-9f96a98e3891\") " pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-qxgcp" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.321358 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s7ldk\" (UniqueName: \"kubernetes.io/projected/c0376a5f-ed42-4399-8f41-2d2157a654fd-kube-api-access-s7ldk\") pod \"neutron-operator-controller-manager-7c57c8bbc4-hwp4x\" (UID: \"c0376a5f-ed42-4399-8f41-2d2157a654fd\") " pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-hwp4x" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.321386 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j4f9s\" (UniqueName: \"kubernetes.io/projected/d42d0991-9e6f-46d0-9236-00e85c26a603-kube-api-access-j4f9s\") pod \"mariadb-operator-controller-manager-cb6c4fdb7-fw96v\" (UID: \"d42d0991-9e6f-46d0-9236-00e85c26a603\") " pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-fw96v" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.321422 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zj6np\" (UniqueName: \"kubernetes.io/projected/db1d55a6-1285-429d-b9dd-b2e100e602c6-kube-api-access-zj6np\") pod \"octavia-operator-controller-manager-fd75fd47d-5ncqq\" (UID: \"db1d55a6-1285-429d-b9dd-b2e100e602c6\") " pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-5ncqq" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.321816 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-559p2\" (UniqueName: \"kubernetes.io/projected/14385dd9-9b06-473b-919d-9d2f9207ffe1-kube-api-access-559p2\") pod \"openstack-baremetal-operator-controller-manager-544b9bb9-lwp9z\" (UID: \"14385dd9-9b06-473b-919d-9d2f9207ffe1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-lwp9z" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.326039 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-dnpsf" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.333842 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-lwp9z"] Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.339374 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s7ldk\" (UniqueName: \"kubernetes.io/projected/c0376a5f-ed42-4399-8f41-2d2157a654fd-kube-api-access-s7ldk\") pod \"neutron-operator-controller-manager-7c57c8bbc4-hwp4x\" (UID: \"c0376a5f-ed42-4399-8f41-2d2157a654fd\") " pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-hwp4x" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.345710 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j4f9s\" (UniqueName: \"kubernetes.io/projected/d42d0991-9e6f-46d0-9236-00e85c26a603-kube-api-access-j4f9s\") pod \"mariadb-operator-controller-manager-cb6c4fdb7-fw96v\" (UID: \"d42d0991-9e6f-46d0-9236-00e85c26a603\") " pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-fw96v" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.347761 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zj6np\" (UniqueName: \"kubernetes.io/projected/db1d55a6-1285-429d-b9dd-b2e100e602c6-kube-api-access-zj6np\") pod \"octavia-operator-controller-manager-fd75fd47d-5ncqq\" (UID: \"db1d55a6-1285-429d-b9dd-b2e100e602c6\") " pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-5ncqq" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.347826 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-66cf5c67ff-qxgcp"] Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.363871 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5db546f9d9-66kf4"] Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.372704 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-6fdc4fcf86-75q4l"] Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.374215 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-75q4l" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.377050 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-hmwsq" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.397881 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-6fdc4fcf86-75q4l"] Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.415477 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-fqfwh" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.435984 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-559p2\" (UniqueName: \"kubernetes.io/projected/14385dd9-9b06-473b-919d-9d2f9207ffe1-kube-api-access-559p2\") pod \"openstack-baremetal-operator-controller-manager-544b9bb9-lwp9z\" (UID: \"14385dd9-9b06-473b-919d-9d2f9207ffe1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-lwp9z" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.436193 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/14385dd9-9b06-473b-919d-9d2f9207ffe1-cert\") pod \"openstack-baremetal-operator-controller-manager-544b9bb9-lwp9z\" (UID: \"14385dd9-9b06-473b-919d-9d2f9207ffe1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-lwp9z" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.436235 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-flgg6\" (UniqueName: \"kubernetes.io/projected/1f69797f-4204-4989-a00b-d9722d44d3c6-kube-api-access-flgg6\") pod \"nova-operator-controller-manager-79556f57fc-4ftbp\" (UID: \"1f69797f-4204-4989-a00b-d9722d44d3c6\") " pod="openstack-operators/nova-operator-controller-manager-79556f57fc-4ftbp" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.436294 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ntb45\" (UniqueName: \"kubernetes.io/projected/2b44040a-e224-4468-a528-9f96a98e3891-kube-api-access-ntb45\") pod \"ovn-operator-controller-manager-66cf5c67ff-qxgcp\" (UID: \"2b44040a-e224-4468-a528-9f96a98e3891\") " pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-qxgcp" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.438008 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-f8tvv" Nov 25 10:45:59 crc kubenswrapper[4821]: E1125 10:45:59.439493 4821 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 25 10:45:59 crc kubenswrapper[4821]: E1125 10:45:59.439559 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/14385dd9-9b06-473b-919d-9d2f9207ffe1-cert podName:14385dd9-9b06-473b-919d-9d2f9207ffe1 nodeName:}" failed. No retries permitted until 2025-11-25 10:45:59.939542452 +0000 UTC m=+830.475862299 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/14385dd9-9b06-473b-919d-9d2f9207ffe1-cert") pod "openstack-baremetal-operator-controller-manager-544b9bb9-lwp9z" (UID: "14385dd9-9b06-473b-919d-9d2f9207ffe1") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.495151 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-8dbc5685d-gcl85"] Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.496881 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-8dbc5685d-gcl85" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.497439 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ntb45\" (UniqueName: \"kubernetes.io/projected/2b44040a-e224-4468-a528-9f96a98e3891-kube-api-access-ntb45\") pod \"ovn-operator-controller-manager-66cf5c67ff-qxgcp\" (UID: \"2b44040a-e224-4468-a528-9f96a98e3891\") " pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-qxgcp" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.498285 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-vnf9x" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.498962 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-2fccf" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.507959 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-flgg6\" (UniqueName: \"kubernetes.io/projected/1f69797f-4204-4989-a00b-d9722d44d3c6-kube-api-access-flgg6\") pod \"nova-operator-controller-manager-79556f57fc-4ftbp\" (UID: \"1f69797f-4204-4989-a00b-d9722d44d3c6\") " pod="openstack-operators/nova-operator-controller-manager-79556f57fc-4ftbp" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.522775 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-559p2\" (UniqueName: \"kubernetes.io/projected/14385dd9-9b06-473b-919d-9d2f9207ffe1-kube-api-access-559p2\") pod \"openstack-baremetal-operator-controller-manager-544b9bb9-lwp9z\" (UID: \"14385dd9-9b06-473b-919d-9d2f9207ffe1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-lwp9z" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.548625 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-fw96v" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.549384 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dclkc\" (UniqueName: \"kubernetes.io/projected/c77cb0f4-1dff-466c-a8bf-6a987bc87ced-kube-api-access-dclkc\") pod \"swift-operator-controller-manager-6fdc4fcf86-75q4l\" (UID: \"c77cb0f4-1dff-466c-a8bf-6a987bc87ced\") " pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-75q4l" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.549423 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6p87t\" (UniqueName: \"kubernetes.io/projected/a64e88ce-fe1a-4cb1-984a-c6d8b5e5ffd9-kube-api-access-6p87t\") pod \"telemetry-operator-controller-manager-8dbc5685d-gcl85\" (UID: \"a64e88ce-fe1a-4cb1-984a-c6d8b5e5ffd9\") " pod="openstack-operators/telemetry-operator-controller-manager-8dbc5685d-gcl85" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.556269 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2czrm\" (UniqueName: \"kubernetes.io/projected/876c088c-7ee5-4b61-a798-eb36ac1e7e21-kube-api-access-2czrm\") pod \"placement-operator-controller-manager-5db546f9d9-66kf4\" (UID: \"876c088c-7ee5-4b61-a798-eb36ac1e7e21\") " pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-66kf4" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.556519 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-8dbc5685d-gcl85"] Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.563278 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-hwp4x" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.584716 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-5ncqq" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.600148 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5cb74df96-lppg4"] Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.601174 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5cb74df96-lppg4" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.602947 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5cb74df96-lppg4"] Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.605485 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-4ftbp" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.606137 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-52cml" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.639308 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-864885998-vnmtq"] Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.641783 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-864885998-vnmtq" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.643994 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-9fl88" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.645605 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-864885998-vnmtq"] Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.657601 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dclkc\" (UniqueName: \"kubernetes.io/projected/c77cb0f4-1dff-466c-a8bf-6a987bc87ced-kube-api-access-dclkc\") pod \"swift-operator-controller-manager-6fdc4fcf86-75q4l\" (UID: \"c77cb0f4-1dff-466c-a8bf-6a987bc87ced\") " pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-75q4l" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.659914 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6p87t\" (UniqueName: \"kubernetes.io/projected/a64e88ce-fe1a-4cb1-984a-c6d8b5e5ffd9-kube-api-access-6p87t\") pod \"telemetry-operator-controller-manager-8dbc5685d-gcl85\" (UID: \"a64e88ce-fe1a-4cb1-984a-c6d8b5e5ffd9\") " pod="openstack-operators/telemetry-operator-controller-manager-8dbc5685d-gcl85" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.660017 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2czrm\" (UniqueName: \"kubernetes.io/projected/876c088c-7ee5-4b61-a798-eb36ac1e7e21-kube-api-access-2czrm\") pod \"placement-operator-controller-manager-5db546f9d9-66kf4\" (UID: \"876c088c-7ee5-4b61-a798-eb36ac1e7e21\") " pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-66kf4" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.660079 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgrk7\" (UniqueName: \"kubernetes.io/projected/505faacd-65f5-4078-9867-9ae648652b50-kube-api-access-sgrk7\") pod \"watcher-operator-controller-manager-864885998-vnmtq\" (UID: \"505faacd-65f5-4078-9867-9ae648652b50\") " pod="openstack-operators/watcher-operator-controller-manager-864885998-vnmtq" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.660132 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4lkf4\" (UniqueName: \"kubernetes.io/projected/8fef5c2e-676a-49b6-a10c-94abe226600a-kube-api-access-4lkf4\") pod \"test-operator-controller-manager-5cb74df96-lppg4\" (UID: \"8fef5c2e-676a-49b6-a10c-94abe226600a\") " pod="openstack-operators/test-operator-controller-manager-5cb74df96-lppg4" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.678997 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-64b4f4bf8b-vzb6r"] Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.679326 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-qxgcp" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.679941 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-64b4f4bf8b-vzb6r" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.683307 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.684063 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-64b4f4bf8b-vzb6r"] Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.685208 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.685750 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-fvk7k" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.710443 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2czrm\" (UniqueName: \"kubernetes.io/projected/876c088c-7ee5-4b61-a798-eb36ac1e7e21-kube-api-access-2czrm\") pod \"placement-operator-controller-manager-5db546f9d9-66kf4\" (UID: \"876c088c-7ee5-4b61-a798-eb36ac1e7e21\") " pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-66kf4" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.714678 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-fprdh"] Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.715539 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-fprdh" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.717952 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-qtcnd" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.720303 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-fprdh"] Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.725488 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6p87t\" (UniqueName: \"kubernetes.io/projected/a64e88ce-fe1a-4cb1-984a-c6d8b5e5ffd9-kube-api-access-6p87t\") pod \"telemetry-operator-controller-manager-8dbc5685d-gcl85\" (UID: \"a64e88ce-fe1a-4cb1-984a-c6d8b5e5ffd9\") " pod="openstack-operators/telemetry-operator-controller-manager-8dbc5685d-gcl85" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.725755 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-66kf4" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.726898 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dclkc\" (UniqueName: \"kubernetes.io/projected/c77cb0f4-1dff-466c-a8bf-6a987bc87ced-kube-api-access-dclkc\") pod \"swift-operator-controller-manager-6fdc4fcf86-75q4l\" (UID: \"c77cb0f4-1dff-466c-a8bf-6a987bc87ced\") " pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-75q4l" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.747405 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-75q4l" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.760919 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9db84cac-a8df-4359-8051-ebc064a97b4d-cert\") pod \"infra-operator-controller-manager-d5cc86f4b-8cqfp\" (UID: \"9db84cac-a8df-4359-8051-ebc064a97b4d\") " pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-8cqfp" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.760995 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/40e2ab14-9d05-4569-8a3a-69a34586c504-webhook-certs\") pod \"openstack-operator-controller-manager-64b4f4bf8b-vzb6r\" (UID: \"40e2ab14-9d05-4569-8a3a-69a34586c504\") " pod="openstack-operators/openstack-operator-controller-manager-64b4f4bf8b-vzb6r" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.761048 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgrk7\" (UniqueName: \"kubernetes.io/projected/505faacd-65f5-4078-9867-9ae648652b50-kube-api-access-sgrk7\") pod \"watcher-operator-controller-manager-864885998-vnmtq\" (UID: \"505faacd-65f5-4078-9867-9ae648652b50\") " pod="openstack-operators/watcher-operator-controller-manager-864885998-vnmtq" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.761085 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/40e2ab14-9d05-4569-8a3a-69a34586c504-metrics-certs\") pod \"openstack-operator-controller-manager-64b4f4bf8b-vzb6r\" (UID: \"40e2ab14-9d05-4569-8a3a-69a34586c504\") " pod="openstack-operators/openstack-operator-controller-manager-64b4f4bf8b-vzb6r" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.761110 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gg7nk\" (UniqueName: \"kubernetes.io/projected/40e2ab14-9d05-4569-8a3a-69a34586c504-kube-api-access-gg7nk\") pod \"openstack-operator-controller-manager-64b4f4bf8b-vzb6r\" (UID: \"40e2ab14-9d05-4569-8a3a-69a34586c504\") " pod="openstack-operators/openstack-operator-controller-manager-64b4f4bf8b-vzb6r" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.761140 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4lkf4\" (UniqueName: \"kubernetes.io/projected/8fef5c2e-676a-49b6-a10c-94abe226600a-kube-api-access-4lkf4\") pod \"test-operator-controller-manager-5cb74df96-lppg4\" (UID: \"8fef5c2e-676a-49b6-a10c-94abe226600a\") " pod="openstack-operators/test-operator-controller-manager-5cb74df96-lppg4" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.761198 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l9qb6\" (UniqueName: \"kubernetes.io/projected/5cea2a2e-a19f-42fc-ac84-aadfa677ec4c-kube-api-access-l9qb6\") pod \"rabbitmq-cluster-operator-manager-668c99d594-fprdh\" (UID: \"5cea2a2e-a19f-42fc-ac84-aadfa677ec4c\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-fprdh" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.769181 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9db84cac-a8df-4359-8051-ebc064a97b4d-cert\") pod \"infra-operator-controller-manager-d5cc86f4b-8cqfp\" (UID: \"9db84cac-a8df-4359-8051-ebc064a97b4d\") " pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-8cqfp" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.781573 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4lkf4\" (UniqueName: \"kubernetes.io/projected/8fef5c2e-676a-49b6-a10c-94abe226600a-kube-api-access-4lkf4\") pod \"test-operator-controller-manager-5cb74df96-lppg4\" (UID: \"8fef5c2e-676a-49b6-a10c-94abe226600a\") " pod="openstack-operators/test-operator-controller-manager-5cb74df96-lppg4" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.784608 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgrk7\" (UniqueName: \"kubernetes.io/projected/505faacd-65f5-4078-9867-9ae648652b50-kube-api-access-sgrk7\") pod \"watcher-operator-controller-manager-864885998-vnmtq\" (UID: \"505faacd-65f5-4078-9867-9ae648652b50\") " pod="openstack-operators/watcher-operator-controller-manager-864885998-vnmtq" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.814872 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5cb74df96-lppg4" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.827358 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-864885998-vnmtq" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.862419 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/40e2ab14-9d05-4569-8a3a-69a34586c504-metrics-certs\") pod \"openstack-operator-controller-manager-64b4f4bf8b-vzb6r\" (UID: \"40e2ab14-9d05-4569-8a3a-69a34586c504\") " pod="openstack-operators/openstack-operator-controller-manager-64b4f4bf8b-vzb6r" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.862462 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gg7nk\" (UniqueName: \"kubernetes.io/projected/40e2ab14-9d05-4569-8a3a-69a34586c504-kube-api-access-gg7nk\") pod \"openstack-operator-controller-manager-64b4f4bf8b-vzb6r\" (UID: \"40e2ab14-9d05-4569-8a3a-69a34586c504\") " pod="openstack-operators/openstack-operator-controller-manager-64b4f4bf8b-vzb6r" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.862518 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l9qb6\" (UniqueName: \"kubernetes.io/projected/5cea2a2e-a19f-42fc-ac84-aadfa677ec4c-kube-api-access-l9qb6\") pod \"rabbitmq-cluster-operator-manager-668c99d594-fprdh\" (UID: \"5cea2a2e-a19f-42fc-ac84-aadfa677ec4c\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-fprdh" Nov 25 10:45:59 crc kubenswrapper[4821]: E1125 10:45:59.862561 4821 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.862597 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/40e2ab14-9d05-4569-8a3a-69a34586c504-webhook-certs\") pod \"openstack-operator-controller-manager-64b4f4bf8b-vzb6r\" (UID: \"40e2ab14-9d05-4569-8a3a-69a34586c504\") " pod="openstack-operators/openstack-operator-controller-manager-64b4f4bf8b-vzb6r" Nov 25 10:45:59 crc kubenswrapper[4821]: E1125 10:45:59.862614 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/40e2ab14-9d05-4569-8a3a-69a34586c504-metrics-certs podName:40e2ab14-9d05-4569-8a3a-69a34586c504 nodeName:}" failed. No retries permitted until 2025-11-25 10:46:00.362597988 +0000 UTC m=+830.898917835 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/40e2ab14-9d05-4569-8a3a-69a34586c504-metrics-certs") pod "openstack-operator-controller-manager-64b4f4bf8b-vzb6r" (UID: "40e2ab14-9d05-4569-8a3a-69a34586c504") : secret "metrics-server-cert" not found Nov 25 10:45:59 crc kubenswrapper[4821]: E1125 10:45:59.863526 4821 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 25 10:45:59 crc kubenswrapper[4821]: E1125 10:45:59.863550 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/40e2ab14-9d05-4569-8a3a-69a34586c504-webhook-certs podName:40e2ab14-9d05-4569-8a3a-69a34586c504 nodeName:}" failed. No retries permitted until 2025-11-25 10:46:00.363542311 +0000 UTC m=+830.899862158 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/40e2ab14-9d05-4569-8a3a-69a34586c504-webhook-certs") pod "openstack-operator-controller-manager-64b4f4bf8b-vzb6r" (UID: "40e2ab14-9d05-4569-8a3a-69a34586c504") : secret "webhook-server-cert" not found Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.880996 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-8dbc5685d-gcl85" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.884506 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gg7nk\" (UniqueName: \"kubernetes.io/projected/40e2ab14-9d05-4569-8a3a-69a34586c504-kube-api-access-gg7nk\") pod \"openstack-operator-controller-manager-64b4f4bf8b-vzb6r\" (UID: \"40e2ab14-9d05-4569-8a3a-69a34586c504\") " pod="openstack-operators/openstack-operator-controller-manager-64b4f4bf8b-vzb6r" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.890035 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l9qb6\" (UniqueName: \"kubernetes.io/projected/5cea2a2e-a19f-42fc-ac84-aadfa677ec4c-kube-api-access-l9qb6\") pod \"rabbitmq-cluster-operator-manager-668c99d594-fprdh\" (UID: \"5cea2a2e-a19f-42fc-ac84-aadfa677ec4c\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-fprdh" Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.964519 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/14385dd9-9b06-473b-919d-9d2f9207ffe1-cert\") pod \"openstack-baremetal-operator-controller-manager-544b9bb9-lwp9z\" (UID: \"14385dd9-9b06-473b-919d-9d2f9207ffe1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-lwp9z" Nov 25 10:45:59 crc kubenswrapper[4821]: E1125 10:45:59.964659 4821 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 25 10:45:59 crc kubenswrapper[4821]: E1125 10:45:59.964709 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/14385dd9-9b06-473b-919d-9d2f9207ffe1-cert podName:14385dd9-9b06-473b-919d-9d2f9207ffe1 nodeName:}" failed. No retries permitted until 2025-11-25 10:46:00.964693476 +0000 UTC m=+831.501013323 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/14385dd9-9b06-473b-919d-9d2f9207ffe1-cert") pod "openstack-baremetal-operator-controller-manager-544b9bb9-lwp9z" (UID: "14385dd9-9b06-473b-919d-9d2f9207ffe1") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 25 10:45:59 crc kubenswrapper[4821]: I1125 10:45:59.991363 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-8cqfp" Nov 25 10:46:00 crc kubenswrapper[4821]: I1125 10:46:00.056705 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zwqgf"] Nov 25 10:46:00 crc kubenswrapper[4821]: I1125 10:46:00.057174 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zwqgf" podUID="fd8b5398-7942-4c13-8840-a53ff1aefcbf" containerName="registry-server" containerID="cri-o://33cb6765f313a785652c5ee17c60ad16e2235717c721efbf3604550bf7e9d3bc" gracePeriod=2 Nov 25 10:46:00 crc kubenswrapper[4821]: I1125 10:46:00.139009 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-86dc4d89c8-pnq4j"] Nov 25 10:46:00 crc kubenswrapper[4821]: I1125 10:46:00.177103 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-fprdh" Nov 25 10:46:00 crc kubenswrapper[4821]: I1125 10:46:00.288592 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-68b95954c9-qgs92"] Nov 25 10:46:00 crc kubenswrapper[4821]: W1125 10:46:00.305349 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda3dc5167_4997_44bf_ace3_706829fd0f17.slice/crio-60a3be86aa020f12802c6610ef75450304c15b4e9fc68c71e27e25ed616599d9 WatchSource:0}: Error finding container 60a3be86aa020f12802c6610ef75450304c15b4e9fc68c71e27e25ed616599d9: Status 404 returned error can't find the container with id 60a3be86aa020f12802c6610ef75450304c15b4e9fc68c71e27e25ed616599d9 Nov 25 10:46:00 crc kubenswrapper[4821]: I1125 10:46:00.323631 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-7d695c9b56-7sx9m"] Nov 25 10:46:00 crc kubenswrapper[4821]: W1125 10:46:00.336549 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2236607e_2a7b_4858_b5b9_c43e47c9375d.slice/crio-2294128c065bf04b0d1cb74067d1d95e6a04fcc60009a17d5b366a044537906f WatchSource:0}: Error finding container 2294128c065bf04b0d1cb74067d1d95e6a04fcc60009a17d5b366a044537906f: Status 404 returned error can't find the container with id 2294128c065bf04b0d1cb74067d1d95e6a04fcc60009a17d5b366a044537906f Nov 25 10:46:00 crc kubenswrapper[4821]: I1125 10:46:00.370825 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/40e2ab14-9d05-4569-8a3a-69a34586c504-webhook-certs\") pod \"openstack-operator-controller-manager-64b4f4bf8b-vzb6r\" (UID: \"40e2ab14-9d05-4569-8a3a-69a34586c504\") " pod="openstack-operators/openstack-operator-controller-manager-64b4f4bf8b-vzb6r" Nov 25 10:46:00 crc kubenswrapper[4821]: I1125 10:46:00.370904 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/40e2ab14-9d05-4569-8a3a-69a34586c504-metrics-certs\") pod \"openstack-operator-controller-manager-64b4f4bf8b-vzb6r\" (UID: \"40e2ab14-9d05-4569-8a3a-69a34586c504\") " pod="openstack-operators/openstack-operator-controller-manager-64b4f4bf8b-vzb6r" Nov 25 10:46:00 crc kubenswrapper[4821]: E1125 10:46:00.371007 4821 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 25 10:46:00 crc kubenswrapper[4821]: E1125 10:46:00.371060 4821 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 25 10:46:00 crc kubenswrapper[4821]: E1125 10:46:00.371075 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/40e2ab14-9d05-4569-8a3a-69a34586c504-webhook-certs podName:40e2ab14-9d05-4569-8a3a-69a34586c504 nodeName:}" failed. No retries permitted until 2025-11-25 10:46:01.371059258 +0000 UTC m=+831.907379105 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/40e2ab14-9d05-4569-8a3a-69a34586c504-webhook-certs") pod "openstack-operator-controller-manager-64b4f4bf8b-vzb6r" (UID: "40e2ab14-9d05-4569-8a3a-69a34586c504") : secret "webhook-server-cert" not found Nov 25 10:46:00 crc kubenswrapper[4821]: E1125 10:46:00.371135 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/40e2ab14-9d05-4569-8a3a-69a34586c504-metrics-certs podName:40e2ab14-9d05-4569-8a3a-69a34586c504 nodeName:}" failed. No retries permitted until 2025-11-25 10:46:01.37111898 +0000 UTC m=+831.907438827 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/40e2ab14-9d05-4569-8a3a-69a34586c504-metrics-certs") pod "openstack-operator-controller-manager-64b4f4bf8b-vzb6r" (UID: "40e2ab14-9d05-4569-8a3a-69a34586c504") : secret "metrics-server-cert" not found Nov 25 10:46:00 crc kubenswrapper[4821]: I1125 10:46:00.445997 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5bfcdc958c-fqfwh"] Nov 25 10:46:00 crc kubenswrapper[4821]: W1125 10:46:00.468389 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7982734e_b887_4cea_9724_3a37f9369ac5.slice/crio-99f321d8ea082ec0054f8587e652d8ab325343f90cea6944f3c9ced7da110b97 WatchSource:0}: Error finding container 99f321d8ea082ec0054f8587e652d8ab325343f90cea6944f3c9ced7da110b97: Status 404 returned error can't find the container with id 99f321d8ea082ec0054f8587e652d8ab325343f90cea6944f3c9ced7da110b97 Nov 25 10:46:00 crc kubenswrapper[4821]: I1125 10:46:00.468727 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-748dc6576f-f8tvv"] Nov 25 10:46:00 crc kubenswrapper[4821]: I1125 10:46:00.476701 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-774b86978c-s49f6"] Nov 25 10:46:00 crc kubenswrapper[4821]: I1125 10:46:00.506841 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-79856dc55c-p4bwc"] Nov 25 10:46:00 crc kubenswrapper[4821]: I1125 10:46:00.512583 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c9694994-2qzrg"] Nov 25 10:46:00 crc kubenswrapper[4821]: W1125 10:46:00.556942 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb7895321_27ec_4c75_aa5e_581fe46e0789.slice/crio-026b42636230a5f25f47c16d535cbdccbc8a85f31c79e4abe694ced141ee80d6 WatchSource:0}: Error finding container 026b42636230a5f25f47c16d535cbdccbc8a85f31c79e4abe694ced141ee80d6: Status 404 returned error can't find the container with id 026b42636230a5f25f47c16d535cbdccbc8a85f31c79e4abe694ced141ee80d6 Nov 25 10:46:00 crc kubenswrapper[4821]: I1125 10:46:00.780284 4821 generic.go:334] "Generic (PLEG): container finished" podID="fd8b5398-7942-4c13-8840-a53ff1aefcbf" containerID="33cb6765f313a785652c5ee17c60ad16e2235717c721efbf3604550bf7e9d3bc" exitCode=0 Nov 25 10:46:00 crc kubenswrapper[4821]: I1125 10:46:00.780359 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zwqgf" event={"ID":"fd8b5398-7942-4c13-8840-a53ff1aefcbf","Type":"ContainerDied","Data":"33cb6765f313a785652c5ee17c60ad16e2235717c721efbf3604550bf7e9d3bc"} Nov 25 10:46:00 crc kubenswrapper[4821]: I1125 10:46:00.781423 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-2qzrg" event={"ID":"b7895321-27ec-4c75-aa5e-581fe46e0789","Type":"ContainerStarted","Data":"026b42636230a5f25f47c16d535cbdccbc8a85f31c79e4abe694ced141ee80d6"} Nov 25 10:46:00 crc kubenswrapper[4821]: I1125 10:46:00.782387 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-fqfwh" event={"ID":"1976151f-2b7a-4cde-95d3-9020b17d938c","Type":"ContainerStarted","Data":"ad702e8340d189629c69759c2ddd12e5f429fd79cd05cdb057bd776703ed8ab5"} Nov 25 10:46:00 crc kubenswrapper[4821]: I1125 10:46:00.783320 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-68b95954c9-qgs92" event={"ID":"a3dc5167-4997-44bf-ace3-706829fd0f17","Type":"ContainerStarted","Data":"60a3be86aa020f12802c6610ef75450304c15b4e9fc68c71e27e25ed616599d9"} Nov 25 10:46:00 crc kubenswrapper[4821]: I1125 10:46:00.784183 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-774b86978c-s49f6" event={"ID":"b35e42b0-2a6f-454f-b8e2-4f7cd9b9217f","Type":"ContainerStarted","Data":"34205e14deefdadd6aadc57792e7771a009d5a0190b4dd7b768c8bca77d5fe8c"} Nov 25 10:46:00 crc kubenswrapper[4821]: I1125 10:46:00.785660 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-p4bwc" event={"ID":"8f16d64f-ea14-4853-b30b-2df294ce6cc3","Type":"ContainerStarted","Data":"0ddc263eb184e5b6d13cb750ba835d2b680a7b08db47754d3ba9b3ecb0112cac"} Nov 25 10:46:00 crc kubenswrapper[4821]: I1125 10:46:00.787477 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-pnq4j" event={"ID":"db28ec64-d8be-4b56-a427-e4cceb256fd0","Type":"ContainerStarted","Data":"4ae65945ee7a1db293a57ae6e50b59e52404b1ed7961b2f3b7a2458bff202c66"} Nov 25 10:46:00 crc kubenswrapper[4821]: I1125 10:46:00.788321 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-f8tvv" event={"ID":"7982734e-b887-4cea-9724-3a37f9369ac5","Type":"ContainerStarted","Data":"99f321d8ea082ec0054f8587e652d8ab325343f90cea6944f3c9ced7da110b97"} Nov 25 10:46:00 crc kubenswrapper[4821]: I1125 10:46:00.790750 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-7sx9m" event={"ID":"2236607e-2a7b-4858-b5b9-c43e47c9375d","Type":"ContainerStarted","Data":"2294128c065bf04b0d1cb74067d1d95e6a04fcc60009a17d5b366a044537906f"} Nov 25 10:46:00 crc kubenswrapper[4821]: I1125 10:46:00.867020 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-6fdc4fcf86-75q4l"] Nov 25 10:46:00 crc kubenswrapper[4821]: I1125 10:46:00.876791 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5db546f9d9-66kf4"] Nov 25 10:46:00 crc kubenswrapper[4821]: I1125 10:46:00.890493 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-hwp4x"] Nov 25 10:46:00 crc kubenswrapper[4821]: W1125 10:46:00.890712 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod876c088c_7ee5_4b61_a798_eb36ac1e7e21.slice/crio-0b01c98a3ad8c633b6981940c1ec4c69c01c8f7268d015be7f2065a0f5227dec WatchSource:0}: Error finding container 0b01c98a3ad8c633b6981940c1ec4c69c01c8f7268d015be7f2065a0f5227dec: Status 404 returned error can't find the container with id 0b01c98a3ad8c633b6981940c1ec4c69c01c8f7268d015be7f2065a0f5227dec Nov 25 10:46:00 crc kubenswrapper[4821]: W1125 10:46:00.895883 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc77cb0f4_1dff_466c_a8bf_6a987bc87ced.slice/crio-0620f056ddc073f864c0ac55d52cc6b1397540a76e010942a63bb189b601c894 WatchSource:0}: Error finding container 0620f056ddc073f864c0ac55d52cc6b1397540a76e010942a63bb189b601c894: Status 404 returned error can't find the container with id 0620f056ddc073f864c0ac55d52cc6b1397540a76e010942a63bb189b601c894 Nov 25 10:46:00 crc kubenswrapper[4821]: I1125 10:46:00.908290 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5cb74df96-lppg4"] Nov 25 10:46:00 crc kubenswrapper[4821]: I1125 10:46:00.916901 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-fd75fd47d-5ncqq"] Nov 25 10:46:00 crc kubenswrapper[4821]: I1125 10:46:00.928082 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-58bb8d67cc-vnf9x"] Nov 25 10:46:00 crc kubenswrapper[4821]: I1125 10:46:00.948397 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-fw96v"] Nov 25 10:46:00 crc kubenswrapper[4821]: I1125 10:46:00.957458 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-66cf5c67ff-qxgcp"] Nov 25 10:46:00 crc kubenswrapper[4821]: I1125 10:46:00.962382 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-864885998-vnmtq"] Nov 25 10:46:00 crc kubenswrapper[4821]: E1125 10:46:00.962408 4821 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:442c269d79163f8da75505019c02e9f0815837aaadcaddacb8e6c12df297ca13,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zj6np,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-fd75fd47d-5ncqq_openstack-operators(db1d55a6-1285-429d-b9dd-b2e100e602c6): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 10:46:00 crc kubenswrapper[4821]: W1125 10:46:00.963353 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8fef5c2e_676a_49b6_a10c_94abe226600a.slice/crio-133d71c90730ff64388dbce1590894874ee5f792265c91e320199b80bda8ab49 WatchSource:0}: Error finding container 133d71c90730ff64388dbce1590894874ee5f792265c91e320199b80bda8ab49: Status 404 returned error can't find the container with id 133d71c90730ff64388dbce1590894874ee5f792265c91e320199b80bda8ab49 Nov 25 10:46:00 crc kubenswrapper[4821]: W1125 10:46:00.968781 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2b44040a_e224_4468_a528_9f96a98e3891.slice/crio-d150f77f8a5d0dee735f916ac42b8f5abd7d94cababd80ffc5c77616f666c43c WatchSource:0}: Error finding container d150f77f8a5d0dee735f916ac42b8f5abd7d94cababd80ffc5c77616f666c43c: Status 404 returned error can't find the container with id d150f77f8a5d0dee735f916ac42b8f5abd7d94cababd80ffc5c77616f666c43c Nov 25 10:46:00 crc kubenswrapper[4821]: W1125 10:46:00.972730 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda64e88ce_fe1a_4cb1_984a_c6d8b5e5ffd9.slice/crio-3d0bf63b16db578790974c0505b3cfe5461d808630554ee8d03f75a918fbc717 WatchSource:0}: Error finding container 3d0bf63b16db578790974c0505b3cfe5461d808630554ee8d03f75a918fbc717: Status 404 returned error can't find the container with id 3d0bf63b16db578790974c0505b3cfe5461d808630554ee8d03f75a918fbc717 Nov 25 10:46:00 crc kubenswrapper[4821]: W1125 10:46:00.974372 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9db84cac_a8df_4359_8051_ebc064a97b4d.slice/crio-cba4adc869b77f15d5bf350dc30ca7030634bef37c072976df022a7e4df5eb7d WatchSource:0}: Error finding container cba4adc869b77f15d5bf350dc30ca7030634bef37c072976df022a7e4df5eb7d: Status 404 returned error can't find the container with id cba4adc869b77f15d5bf350dc30ca7030634bef37c072976df022a7e4df5eb7d Nov 25 10:46:00 crc kubenswrapper[4821]: E1125 10:46:00.976515 4821 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zj6np,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-fd75fd47d-5ncqq_openstack-operators(db1d55a6-1285-429d-b9dd-b2e100e602c6): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 10:46:00 crc kubenswrapper[4821]: E1125 10:46:00.976812 4821 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:82207e753574d4be246f86c4b074500d66cf20214aa80f0a8525cf3287a35e6d,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4lkf4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5cb74df96-lppg4_openstack-operators(8fef5c2e-676a-49b6-a10c-94abe226600a): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 10:46:00 crc kubenswrapper[4821]: E1125 10:46:00.976812 4821 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:c053e34316044f14929e16e4f0d97f9f1b24cb68b5e22b925ca74c66aaaed0a7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-flgg6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-79556f57fc-4ftbp_openstack-operators(1f69797f-4204-4989-a00b-d9722d44d3c6): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 10:46:00 crc kubenswrapper[4821]: E1125 10:46:00.977074 4821 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/infra-operator@sha256:86df58f744c1d23233cc98f6ea17c8d6da637c50003d0fc8c100045594aa9894,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{600 -3} {} 600m DecimalSI},memory: {{2147483648 0} {} 2Gi BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{536870912 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fq5sc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infra-operator-controller-manager-d5cc86f4b-8cqfp_openstack-operators(9db84cac-a8df-4359-8051-ebc064a97b4d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 10:46:00 crc kubenswrapper[4821]: E1125 10:46:00.977195 4821 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:5d49d4594c66eda7b151746cc6e1d3c67c0129b4503eeb043a64ae8ec2da6a1b,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ntb45,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-66cf5c67ff-qxgcp_openstack-operators(2b44040a-e224-4468-a528-9f96a98e3891): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 10:46:00 crc kubenswrapper[4821]: E1125 10:46:00.977307 4821 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:38.102.83.162:5001/openstack-k8s-operators/telemetry-operator:2e68c244d50295e6912a58e349ce3258687ac72f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6p87t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-8dbc5685d-gcl85_openstack-operators(a64e88ce-fe1a-4cb1-984a-c6d8b5e5ffd9): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 10:46:00 crc kubenswrapper[4821]: E1125 10:46:00.977595 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-5ncqq" podUID="db1d55a6-1285-429d-b9dd-b2e100e602c6" Nov 25 10:46:00 crc kubenswrapper[4821]: I1125 10:46:00.978909 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-79556f57fc-4ftbp"] Nov 25 10:46:00 crc kubenswrapper[4821]: E1125 10:46:00.979705 4821 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-flgg6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-79556f57fc-4ftbp_openstack-operators(1f69797f-4204-4989-a00b-d9722d44d3c6): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 10:46:00 crc kubenswrapper[4821]: E1125 10:46:00.979775 4821 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4lkf4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5cb74df96-lppg4_openstack-operators(8fef5c2e-676a-49b6-a10c-94abe226600a): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 10:46:00 crc kubenswrapper[4821]: E1125 10:46:00.979835 4821 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6p87t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-8dbc5685d-gcl85_openstack-operators(a64e88ce-fe1a-4cb1-984a-c6d8b5e5ffd9): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 10:46:00 crc kubenswrapper[4821]: E1125 10:46:00.979963 4821 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fq5sc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infra-operator-controller-manager-d5cc86f4b-8cqfp_openstack-operators(9db84cac-a8df-4359-8051-ebc064a97b4d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 10:46:00 crc kubenswrapper[4821]: E1125 10:46:00.980023 4821 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ntb45,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-66cf5c67ff-qxgcp_openstack-operators(2b44040a-e224-4468-a528-9f96a98e3891): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 10:46:00 crc kubenswrapper[4821]: E1125 10:46:00.981985 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-qxgcp" podUID="2b44040a-e224-4468-a528-9f96a98e3891" Nov 25 10:46:00 crc kubenswrapper[4821]: E1125 10:46:00.982033 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-8cqfp" podUID="9db84cac-a8df-4359-8051-ebc064a97b4d" Nov 25 10:46:00 crc kubenswrapper[4821]: E1125 10:46:00.982058 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-4ftbp" podUID="1f69797f-4204-4989-a00b-d9722d44d3c6" Nov 25 10:46:00 crc kubenswrapper[4821]: E1125 10:46:00.982082 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/test-operator-controller-manager-5cb74df96-lppg4" podUID="8fef5c2e-676a-49b6-a10c-94abe226600a" Nov 25 10:46:00 crc kubenswrapper[4821]: E1125 10:46:00.982104 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/telemetry-operator-controller-manager-8dbc5685d-gcl85" podUID="a64e88ce-fe1a-4cb1-984a-c6d8b5e5ffd9" Nov 25 10:46:00 crc kubenswrapper[4821]: I1125 10:46:00.985521 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-d5cc86f4b-8cqfp"] Nov 25 10:46:00 crc kubenswrapper[4821]: I1125 10:46:00.992330 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-8dbc5685d-gcl85"] Nov 25 10:46:00 crc kubenswrapper[4821]: I1125 10:46:00.993085 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/14385dd9-9b06-473b-919d-9d2f9207ffe1-cert\") pod \"openstack-baremetal-operator-controller-manager-544b9bb9-lwp9z\" (UID: \"14385dd9-9b06-473b-919d-9d2f9207ffe1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-lwp9z" Nov 25 10:46:00 crc kubenswrapper[4821]: E1125 10:46:00.993309 4821 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 25 10:46:00 crc kubenswrapper[4821]: E1125 10:46:00.993376 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/14385dd9-9b06-473b-919d-9d2f9207ffe1-cert podName:14385dd9-9b06-473b-919d-9d2f9207ffe1 nodeName:}" failed. No retries permitted until 2025-11-25 10:46:02.993356844 +0000 UTC m=+833.529676691 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/14385dd9-9b06-473b-919d-9d2f9207ffe1-cert") pod "openstack-baremetal-operator-controller-manager-544b9bb9-lwp9z" (UID: "14385dd9-9b06-473b-919d-9d2f9207ffe1") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 25 10:46:00 crc kubenswrapper[4821]: E1125 10:46:00.999058 4821 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-l9qb6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-fprdh_openstack-operators(5cea2a2e-a19f-42fc-ac84-aadfa677ec4c): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 10:46:01 crc kubenswrapper[4821]: E1125 10:46:01.000957 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-fprdh" podUID="5cea2a2e-a19f-42fc-ac84-aadfa677ec4c" Nov 25 10:46:01 crc kubenswrapper[4821]: I1125 10:46:01.005744 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-fprdh"] Nov 25 10:46:01 crc kubenswrapper[4821]: I1125 10:46:01.148872 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zwqgf" Nov 25 10:46:01 crc kubenswrapper[4821]: I1125 10:46:01.198640 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd8b5398-7942-4c13-8840-a53ff1aefcbf-catalog-content\") pod \"fd8b5398-7942-4c13-8840-a53ff1aefcbf\" (UID: \"fd8b5398-7942-4c13-8840-a53ff1aefcbf\") " Nov 25 10:46:01 crc kubenswrapper[4821]: I1125 10:46:01.198692 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4dnkx\" (UniqueName: \"kubernetes.io/projected/fd8b5398-7942-4c13-8840-a53ff1aefcbf-kube-api-access-4dnkx\") pod \"fd8b5398-7942-4c13-8840-a53ff1aefcbf\" (UID: \"fd8b5398-7942-4c13-8840-a53ff1aefcbf\") " Nov 25 10:46:01 crc kubenswrapper[4821]: I1125 10:46:01.198723 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd8b5398-7942-4c13-8840-a53ff1aefcbf-utilities\") pod \"fd8b5398-7942-4c13-8840-a53ff1aefcbf\" (UID: \"fd8b5398-7942-4c13-8840-a53ff1aefcbf\") " Nov 25 10:46:01 crc kubenswrapper[4821]: I1125 10:46:01.201415 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fd8b5398-7942-4c13-8840-a53ff1aefcbf-utilities" (OuterVolumeSpecName: "utilities") pod "fd8b5398-7942-4c13-8840-a53ff1aefcbf" (UID: "fd8b5398-7942-4c13-8840-a53ff1aefcbf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:46:01 crc kubenswrapper[4821]: I1125 10:46:01.238066 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd8b5398-7942-4c13-8840-a53ff1aefcbf-kube-api-access-4dnkx" (OuterVolumeSpecName: "kube-api-access-4dnkx") pod "fd8b5398-7942-4c13-8840-a53ff1aefcbf" (UID: "fd8b5398-7942-4c13-8840-a53ff1aefcbf"). InnerVolumeSpecName "kube-api-access-4dnkx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:46:01 crc kubenswrapper[4821]: I1125 10:46:01.318496 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4dnkx\" (UniqueName: \"kubernetes.io/projected/fd8b5398-7942-4c13-8840-a53ff1aefcbf-kube-api-access-4dnkx\") on node \"crc\" DevicePath \"\"" Nov 25 10:46:01 crc kubenswrapper[4821]: I1125 10:46:01.318535 4821 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd8b5398-7942-4c13-8840-a53ff1aefcbf-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:46:01 crc kubenswrapper[4821]: I1125 10:46:01.342473 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fd8b5398-7942-4c13-8840-a53ff1aefcbf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fd8b5398-7942-4c13-8840-a53ff1aefcbf" (UID: "fd8b5398-7942-4c13-8840-a53ff1aefcbf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:46:01 crc kubenswrapper[4821]: I1125 10:46:01.421933 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/40e2ab14-9d05-4569-8a3a-69a34586c504-webhook-certs\") pod \"openstack-operator-controller-manager-64b4f4bf8b-vzb6r\" (UID: \"40e2ab14-9d05-4569-8a3a-69a34586c504\") " pod="openstack-operators/openstack-operator-controller-manager-64b4f4bf8b-vzb6r" Nov 25 10:46:01 crc kubenswrapper[4821]: I1125 10:46:01.422005 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/40e2ab14-9d05-4569-8a3a-69a34586c504-metrics-certs\") pod \"openstack-operator-controller-manager-64b4f4bf8b-vzb6r\" (UID: \"40e2ab14-9d05-4569-8a3a-69a34586c504\") " pod="openstack-operators/openstack-operator-controller-manager-64b4f4bf8b-vzb6r" Nov 25 10:46:01 crc kubenswrapper[4821]: I1125 10:46:01.422046 4821 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd8b5398-7942-4c13-8840-a53ff1aefcbf-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:46:01 crc kubenswrapper[4821]: E1125 10:46:01.422144 4821 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 25 10:46:01 crc kubenswrapper[4821]: E1125 10:46:01.422290 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/40e2ab14-9d05-4569-8a3a-69a34586c504-webhook-certs podName:40e2ab14-9d05-4569-8a3a-69a34586c504 nodeName:}" failed. No retries permitted until 2025-11-25 10:46:03.422269698 +0000 UTC m=+833.958589605 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/40e2ab14-9d05-4569-8a3a-69a34586c504-webhook-certs") pod "openstack-operator-controller-manager-64b4f4bf8b-vzb6r" (UID: "40e2ab14-9d05-4569-8a3a-69a34586c504") : secret "webhook-server-cert" not found Nov 25 10:46:01 crc kubenswrapper[4821]: E1125 10:46:01.422177 4821 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 25 10:46:01 crc kubenswrapper[4821]: E1125 10:46:01.423086 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/40e2ab14-9d05-4569-8a3a-69a34586c504-metrics-certs podName:40e2ab14-9d05-4569-8a3a-69a34586c504 nodeName:}" failed. No retries permitted until 2025-11-25 10:46:03.423052547 +0000 UTC m=+833.959372394 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/40e2ab14-9d05-4569-8a3a-69a34586c504-metrics-certs") pod "openstack-operator-controller-manager-64b4f4bf8b-vzb6r" (UID: "40e2ab14-9d05-4569-8a3a-69a34586c504") : secret "metrics-server-cert" not found Nov 25 10:46:01 crc kubenswrapper[4821]: I1125 10:46:01.800322 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-fprdh" event={"ID":"5cea2a2e-a19f-42fc-ac84-aadfa677ec4c","Type":"ContainerStarted","Data":"dc74da3c963fd1f03b13b2c08e6860c7a0b8205d35ae51b60faef95882b562dc"} Nov 25 10:46:01 crc kubenswrapper[4821]: I1125 10:46:01.802288 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-vnf9x" event={"ID":"5fefd5f2-18cb-4c63-848f-2fc9d63bb828","Type":"ContainerStarted","Data":"8013a939b797dee88eced9a2077e9039df8ecb4994d871a704c466ab5aa893d8"} Nov 25 10:46:01 crc kubenswrapper[4821]: E1125 10:46:01.802417 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-fprdh" podUID="5cea2a2e-a19f-42fc-ac84-aadfa677ec4c" Nov 25 10:46:01 crc kubenswrapper[4821]: I1125 10:46:01.804059 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-8cqfp" event={"ID":"9db84cac-a8df-4359-8051-ebc064a97b4d","Type":"ContainerStarted","Data":"cba4adc869b77f15d5bf350dc30ca7030634bef37c072976df022a7e4df5eb7d"} Nov 25 10:46:01 crc kubenswrapper[4821]: I1125 10:46:01.806030 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-8dbc5685d-gcl85" event={"ID":"a64e88ce-fe1a-4cb1-984a-c6d8b5e5ffd9","Type":"ContainerStarted","Data":"3d0bf63b16db578790974c0505b3cfe5461d808630554ee8d03f75a918fbc717"} Nov 25 10:46:01 crc kubenswrapper[4821]: E1125 10:46:01.808198 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.162:5001/openstack-k8s-operators/telemetry-operator:2e68c244d50295e6912a58e349ce3258687ac72f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-8dbc5685d-gcl85" podUID="a64e88ce-fe1a-4cb1-984a-c6d8b5e5ffd9" Nov 25 10:46:01 crc kubenswrapper[4821]: I1125 10:46:01.809370 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-864885998-vnmtq" event={"ID":"505faacd-65f5-4078-9867-9ae648652b50","Type":"ContainerStarted","Data":"9a666fa61b0e4e8c06f64aec0a10867c1f21b894720f7c694d33694612d7de5c"} Nov 25 10:46:01 crc kubenswrapper[4821]: E1125 10:46:01.810764 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:86df58f744c1d23233cc98f6ea17c8d6da637c50003d0fc8c100045594aa9894\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-8cqfp" podUID="9db84cac-a8df-4359-8051-ebc064a97b4d" Nov 25 10:46:01 crc kubenswrapper[4821]: I1125 10:46:01.812290 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-66kf4" event={"ID":"876c088c-7ee5-4b61-a798-eb36ac1e7e21","Type":"ContainerStarted","Data":"0b01c98a3ad8c633b6981940c1ec4c69c01c8f7268d015be7f2065a0f5227dec"} Nov 25 10:46:01 crc kubenswrapper[4821]: I1125 10:46:01.816847 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cb74df96-lppg4" event={"ID":"8fef5c2e-676a-49b6-a10c-94abe226600a","Type":"ContainerStarted","Data":"133d71c90730ff64388dbce1590894874ee5f792265c91e320199b80bda8ab49"} Nov 25 10:46:01 crc kubenswrapper[4821]: E1125 10:46:01.818575 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:82207e753574d4be246f86c4b074500d66cf20214aa80f0a8525cf3287a35e6d\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5cb74df96-lppg4" podUID="8fef5c2e-676a-49b6-a10c-94abe226600a" Nov 25 10:46:01 crc kubenswrapper[4821]: I1125 10:46:01.819734 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zwqgf" event={"ID":"fd8b5398-7942-4c13-8840-a53ff1aefcbf","Type":"ContainerDied","Data":"e2067c3b5ea8206bc60ea0fa8a19321b79672783f08a29cb2f178e3cbb9976a0"} Nov 25 10:46:01 crc kubenswrapper[4821]: I1125 10:46:01.819785 4821 scope.go:117] "RemoveContainer" containerID="33cb6765f313a785652c5ee17c60ad16e2235717c721efbf3604550bf7e9d3bc" Nov 25 10:46:01 crc kubenswrapper[4821]: I1125 10:46:01.819898 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zwqgf" Nov 25 10:46:01 crc kubenswrapper[4821]: I1125 10:46:01.822941 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-4ftbp" event={"ID":"1f69797f-4204-4989-a00b-d9722d44d3c6","Type":"ContainerStarted","Data":"6f03ee8ff48da2f797f0765248c40db7be6a53e6a12ac42708bf9309c3e8434f"} Nov 25 10:46:01 crc kubenswrapper[4821]: E1125 10:46:01.825086 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:c053e34316044f14929e16e4f0d97f9f1b24cb68b5e22b925ca74c66aaaed0a7\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-4ftbp" podUID="1f69797f-4204-4989-a00b-d9722d44d3c6" Nov 25 10:46:01 crc kubenswrapper[4821]: I1125 10:46:01.826495 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-fw96v" event={"ID":"d42d0991-9e6f-46d0-9236-00e85c26a603","Type":"ContainerStarted","Data":"aed7a25f1191879ff785e63f36233736bea1156ae75da922820aa281d5b81ae9"} Nov 25 10:46:01 crc kubenswrapper[4821]: I1125 10:46:01.837155 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-5ncqq" event={"ID":"db1d55a6-1285-429d-b9dd-b2e100e602c6","Type":"ContainerStarted","Data":"2bed996b1d5fd285edf5c3c17fbefcfeaddea0c5660477290d400b16a9d5048a"} Nov 25 10:46:01 crc kubenswrapper[4821]: E1125 10:46:01.845725 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:442c269d79163f8da75505019c02e9f0815837aaadcaddacb8e6c12df297ca13\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-5ncqq" podUID="db1d55a6-1285-429d-b9dd-b2e100e602c6" Nov 25 10:46:01 crc kubenswrapper[4821]: I1125 10:46:01.845942 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-hwp4x" event={"ID":"c0376a5f-ed42-4399-8f41-2d2157a654fd","Type":"ContainerStarted","Data":"4164cff244a6bf5f08c9d5ee6f8063e805c56ab40a26bde06aba3966231dda32"} Nov 25 10:46:01 crc kubenswrapper[4821]: I1125 10:46:01.852656 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-qxgcp" event={"ID":"2b44040a-e224-4468-a528-9f96a98e3891","Type":"ContainerStarted","Data":"d150f77f8a5d0dee735f916ac42b8f5abd7d94cababd80ffc5c77616f666c43c"} Nov 25 10:46:01 crc kubenswrapper[4821]: I1125 10:46:01.861861 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-75q4l" event={"ID":"c77cb0f4-1dff-466c-a8bf-6a987bc87ced","Type":"ContainerStarted","Data":"0620f056ddc073f864c0ac55d52cc6b1397540a76e010942a63bb189b601c894"} Nov 25 10:46:01 crc kubenswrapper[4821]: I1125 10:46:01.886978 4821 scope.go:117] "RemoveContainer" containerID="2f9df2d4912b0774f58480ec42bf51a82a08a6e4ef317765ec4f4f582e28b25e" Nov 25 10:46:01 crc kubenswrapper[4821]: E1125 10:46:01.888230 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:5d49d4594c66eda7b151746cc6e1d3c67c0129b4503eeb043a64ae8ec2da6a1b\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-qxgcp" podUID="2b44040a-e224-4468-a528-9f96a98e3891" Nov 25 10:46:01 crc kubenswrapper[4821]: I1125 10:46:01.921851 4821 scope.go:117] "RemoveContainer" containerID="ca0979132b74469fa866210e56bd7eab29e532a6091a6be469b8fc79fbeb18c3" Nov 25 10:46:01 crc kubenswrapper[4821]: I1125 10:46:01.984504 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zwqgf"] Nov 25 10:46:01 crc kubenswrapper[4821]: I1125 10:46:01.988620 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zwqgf"] Nov 25 10:46:02 crc kubenswrapper[4821]: I1125 10:46:02.142491 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd8b5398-7942-4c13-8840-a53ff1aefcbf" path="/var/lib/kubelet/pods/fd8b5398-7942-4c13-8840-a53ff1aefcbf/volumes" Nov 25 10:46:02 crc kubenswrapper[4821]: E1125 10:46:02.872842 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.162:5001/openstack-k8s-operators/telemetry-operator:2e68c244d50295e6912a58e349ce3258687ac72f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-8dbc5685d-gcl85" podUID="a64e88ce-fe1a-4cb1-984a-c6d8b5e5ffd9" Nov 25 10:46:02 crc kubenswrapper[4821]: E1125 10:46:02.872878 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:442c269d79163f8da75505019c02e9f0815837aaadcaddacb8e6c12df297ca13\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-5ncqq" podUID="db1d55a6-1285-429d-b9dd-b2e100e602c6" Nov 25 10:46:02 crc kubenswrapper[4821]: E1125 10:46:02.873581 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-fprdh" podUID="5cea2a2e-a19f-42fc-ac84-aadfa677ec4c" Nov 25 10:46:02 crc kubenswrapper[4821]: E1125 10:46:02.873555 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:82207e753574d4be246f86c4b074500d66cf20214aa80f0a8525cf3287a35e6d\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5cb74df96-lppg4" podUID="8fef5c2e-676a-49b6-a10c-94abe226600a" Nov 25 10:46:02 crc kubenswrapper[4821]: E1125 10:46:02.873773 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:86df58f744c1d23233cc98f6ea17c8d6da637c50003d0fc8c100045594aa9894\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-8cqfp" podUID="9db84cac-a8df-4359-8051-ebc064a97b4d" Nov 25 10:46:02 crc kubenswrapper[4821]: E1125 10:46:02.874423 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:5d49d4594c66eda7b151746cc6e1d3c67c0129b4503eeb043a64ae8ec2da6a1b\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-qxgcp" podUID="2b44040a-e224-4468-a528-9f96a98e3891" Nov 25 10:46:02 crc kubenswrapper[4821]: E1125 10:46:02.874991 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:c053e34316044f14929e16e4f0d97f9f1b24cb68b5e22b925ca74c66aaaed0a7\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-4ftbp" podUID="1f69797f-4204-4989-a00b-d9722d44d3c6" Nov 25 10:46:03 crc kubenswrapper[4821]: I1125 10:46:03.049293 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/14385dd9-9b06-473b-919d-9d2f9207ffe1-cert\") pod \"openstack-baremetal-operator-controller-manager-544b9bb9-lwp9z\" (UID: \"14385dd9-9b06-473b-919d-9d2f9207ffe1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-lwp9z" Nov 25 10:46:03 crc kubenswrapper[4821]: I1125 10:46:03.073508 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/14385dd9-9b06-473b-919d-9d2f9207ffe1-cert\") pod \"openstack-baremetal-operator-controller-manager-544b9bb9-lwp9z\" (UID: \"14385dd9-9b06-473b-919d-9d2f9207ffe1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-lwp9z" Nov 25 10:46:03 crc kubenswrapper[4821]: I1125 10:46:03.235350 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-lwp9z" Nov 25 10:46:03 crc kubenswrapper[4821]: I1125 10:46:03.458040 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/40e2ab14-9d05-4569-8a3a-69a34586c504-metrics-certs\") pod \"openstack-operator-controller-manager-64b4f4bf8b-vzb6r\" (UID: \"40e2ab14-9d05-4569-8a3a-69a34586c504\") " pod="openstack-operators/openstack-operator-controller-manager-64b4f4bf8b-vzb6r" Nov 25 10:46:03 crc kubenswrapper[4821]: I1125 10:46:03.458172 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/40e2ab14-9d05-4569-8a3a-69a34586c504-webhook-certs\") pod \"openstack-operator-controller-manager-64b4f4bf8b-vzb6r\" (UID: \"40e2ab14-9d05-4569-8a3a-69a34586c504\") " pod="openstack-operators/openstack-operator-controller-manager-64b4f4bf8b-vzb6r" Nov 25 10:46:03 crc kubenswrapper[4821]: I1125 10:46:03.461569 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/40e2ab14-9d05-4569-8a3a-69a34586c504-metrics-certs\") pod \"openstack-operator-controller-manager-64b4f4bf8b-vzb6r\" (UID: \"40e2ab14-9d05-4569-8a3a-69a34586c504\") " pod="openstack-operators/openstack-operator-controller-manager-64b4f4bf8b-vzb6r" Nov 25 10:46:03 crc kubenswrapper[4821]: I1125 10:46:03.461596 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/40e2ab14-9d05-4569-8a3a-69a34586c504-webhook-certs\") pod \"openstack-operator-controller-manager-64b4f4bf8b-vzb6r\" (UID: \"40e2ab14-9d05-4569-8a3a-69a34586c504\") " pod="openstack-operators/openstack-operator-controller-manager-64b4f4bf8b-vzb6r" Nov 25 10:46:03 crc kubenswrapper[4821]: I1125 10:46:03.761447 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-64b4f4bf8b-vzb6r" Nov 25 10:46:04 crc kubenswrapper[4821]: I1125 10:46:04.584156 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-b248b" Nov 25 10:46:04 crc kubenswrapper[4821]: I1125 10:46:04.584523 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-b248b" Nov 25 10:46:04 crc kubenswrapper[4821]: I1125 10:46:04.623017 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-b248b" Nov 25 10:46:04 crc kubenswrapper[4821]: I1125 10:46:04.935386 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-b248b" Nov 25 10:46:05 crc kubenswrapper[4821]: I1125 10:46:05.656353 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-b248b"] Nov 25 10:46:06 crc kubenswrapper[4821]: I1125 10:46:06.895445 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-b248b" podUID="3efc5d6f-66aa-409b-adac-506291c1e810" containerName="registry-server" containerID="cri-o://355abcf932cff05e1a1aede668d3afcfd531d86f0aa99df4c1cf920c0e9642ff" gracePeriod=2 Nov 25 10:46:07 crc kubenswrapper[4821]: I1125 10:46:07.914993 4821 generic.go:334] "Generic (PLEG): container finished" podID="3efc5d6f-66aa-409b-adac-506291c1e810" containerID="355abcf932cff05e1a1aede668d3afcfd531d86f0aa99df4c1cf920c0e9642ff" exitCode=0 Nov 25 10:46:07 crc kubenswrapper[4821]: I1125 10:46:07.915040 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b248b" event={"ID":"3efc5d6f-66aa-409b-adac-506291c1e810","Type":"ContainerDied","Data":"355abcf932cff05e1a1aede668d3afcfd531d86f0aa99df4c1cf920c0e9642ff"} Nov 25 10:46:14 crc kubenswrapper[4821]: E1125 10:46:14.963878 4821 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 355abcf932cff05e1a1aede668d3afcfd531d86f0aa99df4c1cf920c0e9642ff is running failed: container process not found" containerID="355abcf932cff05e1a1aede668d3afcfd531d86f0aa99df4c1cf920c0e9642ff" cmd=["grpc_health_probe","-addr=:50051"] Nov 25 10:46:14 crc kubenswrapper[4821]: E1125 10:46:14.967213 4821 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 355abcf932cff05e1a1aede668d3afcfd531d86f0aa99df4c1cf920c0e9642ff is running failed: container process not found" containerID="355abcf932cff05e1a1aede668d3afcfd531d86f0aa99df4c1cf920c0e9642ff" cmd=["grpc_health_probe","-addr=:50051"] Nov 25 10:46:14 crc kubenswrapper[4821]: E1125 10:46:14.968663 4821 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 355abcf932cff05e1a1aede668d3afcfd531d86f0aa99df4c1cf920c0e9642ff is running failed: container process not found" containerID="355abcf932cff05e1a1aede668d3afcfd531d86f0aa99df4c1cf920c0e9642ff" cmd=["grpc_health_probe","-addr=:50051"] Nov 25 10:46:14 crc kubenswrapper[4821]: E1125 10:46:14.968754 4821 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 355abcf932cff05e1a1aede668d3afcfd531d86f0aa99df4c1cf920c0e9642ff is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-b248b" podUID="3efc5d6f-66aa-409b-adac-506291c1e810" containerName="registry-server" Nov 25 10:46:14 crc kubenswrapper[4821]: E1125 10:46:14.984815 4821 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ironic-operator@sha256:b582189b55fddc180a6d468c9dba7078009a693db37b4093d4ba0c99ec675377" Nov 25 10:46:14 crc kubenswrapper[4821]: E1125 10:46:14.985022 4821 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ironic-operator@sha256:b582189b55fddc180a6d468c9dba7078009a693db37b4093d4ba0c99ec675377,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-k6lm5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-5bfcdc958c-fqfwh_openstack-operators(1976151f-2b7a-4cde-95d3-9020b17d938c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 10:46:15 crc kubenswrapper[4821]: E1125 10:46:15.339734 4821 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/cinder-operator@sha256:553b1288b330ad05771d59c6b73c1681c95f457e8475682f9ad0d2e6b85f37e9" Nov 25 10:46:15 crc kubenswrapper[4821]: E1125 10:46:15.340182 4821 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/cinder-operator@sha256:553b1288b330ad05771d59c6b73c1681c95f457e8475682f9ad0d2e6b85f37e9,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6sbqh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-operator-controller-manager-79856dc55c-p4bwc_openstack-operators(8f16d64f-ea14-4853-b30b-2df294ce6cc3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 10:46:16 crc kubenswrapper[4821]: E1125 10:46:16.005452 4821 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f" Nov 25 10:46:16 crc kubenswrapper[4821]: E1125 10:46:16.006620 4821 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-sgrk7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-864885998-vnmtq_openstack-operators(505faacd-65f5-4078-9867-9ae648652b50): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 10:46:16 crc kubenswrapper[4821]: I1125 10:46:16.050507 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b248b" Nov 25 10:46:16 crc kubenswrapper[4821]: I1125 10:46:16.178009 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vv8ng\" (UniqueName: \"kubernetes.io/projected/3efc5d6f-66aa-409b-adac-506291c1e810-kube-api-access-vv8ng\") pod \"3efc5d6f-66aa-409b-adac-506291c1e810\" (UID: \"3efc5d6f-66aa-409b-adac-506291c1e810\") " Nov 25 10:46:16 crc kubenswrapper[4821]: I1125 10:46:16.178761 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3efc5d6f-66aa-409b-adac-506291c1e810-utilities\") pod \"3efc5d6f-66aa-409b-adac-506291c1e810\" (UID: \"3efc5d6f-66aa-409b-adac-506291c1e810\") " Nov 25 10:46:16 crc kubenswrapper[4821]: I1125 10:46:16.178808 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3efc5d6f-66aa-409b-adac-506291c1e810-catalog-content\") pod \"3efc5d6f-66aa-409b-adac-506291c1e810\" (UID: \"3efc5d6f-66aa-409b-adac-506291c1e810\") " Nov 25 10:46:16 crc kubenswrapper[4821]: I1125 10:46:16.180420 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3efc5d6f-66aa-409b-adac-506291c1e810-utilities" (OuterVolumeSpecName: "utilities") pod "3efc5d6f-66aa-409b-adac-506291c1e810" (UID: "3efc5d6f-66aa-409b-adac-506291c1e810"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:46:16 crc kubenswrapper[4821]: I1125 10:46:16.182336 4821 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3efc5d6f-66aa-409b-adac-506291c1e810-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:46:16 crc kubenswrapper[4821]: I1125 10:46:16.193650 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3efc5d6f-66aa-409b-adac-506291c1e810-kube-api-access-vv8ng" (OuterVolumeSpecName: "kube-api-access-vv8ng") pod "3efc5d6f-66aa-409b-adac-506291c1e810" (UID: "3efc5d6f-66aa-409b-adac-506291c1e810"). InnerVolumeSpecName "kube-api-access-vv8ng". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:46:16 crc kubenswrapper[4821]: I1125 10:46:16.207225 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3efc5d6f-66aa-409b-adac-506291c1e810-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3efc5d6f-66aa-409b-adac-506291c1e810" (UID: "3efc5d6f-66aa-409b-adac-506291c1e810"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:46:16 crc kubenswrapper[4821]: I1125 10:46:16.298711 4821 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3efc5d6f-66aa-409b-adac-506291c1e810-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:46:16 crc kubenswrapper[4821]: I1125 10:46:16.298737 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vv8ng\" (UniqueName: \"kubernetes.io/projected/3efc5d6f-66aa-409b-adac-506291c1e810-kube-api-access-vv8ng\") on node \"crc\" DevicePath \"\"" Nov 25 10:46:16 crc kubenswrapper[4821]: I1125 10:46:16.447780 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-lwp9z"] Nov 25 10:46:16 crc kubenswrapper[4821]: I1125 10:46:16.481797 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-64b4f4bf8b-vzb6r"] Nov 25 10:46:16 crc kubenswrapper[4821]: W1125 10:46:16.605964 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod14385dd9_9b06_473b_919d_9d2f9207ffe1.slice/crio-f632672e05535ad3401ec56befcb8070136dcf6f726a80780ceabad6df360451 WatchSource:0}: Error finding container f632672e05535ad3401ec56befcb8070136dcf6f726a80780ceabad6df360451: Status 404 returned error can't find the container with id f632672e05535ad3401ec56befcb8070136dcf6f726a80780ceabad6df360451 Nov 25 10:46:16 crc kubenswrapper[4821]: I1125 10:46:16.990821 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-f8tvv" event={"ID":"7982734e-b887-4cea-9724-3a37f9369ac5","Type":"ContainerStarted","Data":"116d88a626fcfca604121c94c9187d604fb24e6c7f199897b9f5077ad01a8f96"} Nov 25 10:46:16 crc kubenswrapper[4821]: I1125 10:46:16.994192 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-2qzrg" event={"ID":"b7895321-27ec-4c75-aa5e-581fe46e0789","Type":"ContainerStarted","Data":"8015a4f5da8079577c3b1767269f09cda913f99b1d7f48adc1e4a8e4d8868914"} Nov 25 10:46:16 crc kubenswrapper[4821]: I1125 10:46:16.997724 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-vnf9x" event={"ID":"5fefd5f2-18cb-4c63-848f-2fc9d63bb828","Type":"ContainerStarted","Data":"2cf30b564bb5e631fa007e00c1a3d9ec9add5906ba1304b9dcc9a515138bdf51"} Nov 25 10:46:17 crc kubenswrapper[4821]: I1125 10:46:17.001188 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-hwp4x" event={"ID":"c0376a5f-ed42-4399-8f41-2d2157a654fd","Type":"ContainerStarted","Data":"76442c7c4aabefc1c368c2e08a0a60d2532d22383ebb106b2345b9288064e73c"} Nov 25 10:46:17 crc kubenswrapper[4821]: I1125 10:46:17.012864 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b248b" event={"ID":"3efc5d6f-66aa-409b-adac-506291c1e810","Type":"ContainerDied","Data":"e9b99fc6b314e18855f8e58f0c1b09bc440ef46e676efed63f5309df1e62c19d"} Nov 25 10:46:17 crc kubenswrapper[4821]: I1125 10:46:17.012930 4821 scope.go:117] "RemoveContainer" containerID="355abcf932cff05e1a1aede668d3afcfd531d86f0aa99df4c1cf920c0e9642ff" Nov 25 10:46:17 crc kubenswrapper[4821]: I1125 10:46:17.013062 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b248b" Nov 25 10:46:17 crc kubenswrapper[4821]: I1125 10:46:17.025951 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-7sx9m" event={"ID":"2236607e-2a7b-4858-b5b9-c43e47c9375d","Type":"ContainerStarted","Data":"5201a9fd0e0a9867841b1cdb6bea4429ee7b689ca188f7e7db7685041c33ca54"} Nov 25 10:46:17 crc kubenswrapper[4821]: I1125 10:46:17.027967 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-pnq4j" event={"ID":"db28ec64-d8be-4b56-a427-e4cceb256fd0","Type":"ContainerStarted","Data":"952ebc2db7b8dd148d1415f736d7b5a136b9518a12cbcedf7abf54bf7b536feb"} Nov 25 10:46:17 crc kubenswrapper[4821]: I1125 10:46:17.038945 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-lwp9z" event={"ID":"14385dd9-9b06-473b-919d-9d2f9207ffe1","Type":"ContainerStarted","Data":"f632672e05535ad3401ec56befcb8070136dcf6f726a80780ceabad6df360451"} Nov 25 10:46:17 crc kubenswrapper[4821]: I1125 10:46:17.044846 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-68b95954c9-qgs92" event={"ID":"a3dc5167-4997-44bf-ace3-706829fd0f17","Type":"ContainerStarted","Data":"f497b83d39cb586916285718e12071627c6c089a41b67e5999383ddfc524f570"} Nov 25 10:46:17 crc kubenswrapper[4821]: I1125 10:46:17.062076 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-fw96v" event={"ID":"d42d0991-9e6f-46d0-9236-00e85c26a603","Type":"ContainerStarted","Data":"fc0fa2dd435d97da1b6a1f2d93ef51a085bcbd7d481e2e67af73ea7fe86529ca"} Nov 25 10:46:17 crc kubenswrapper[4821]: I1125 10:46:17.069270 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-b248b"] Nov 25 10:46:17 crc kubenswrapper[4821]: I1125 10:46:17.073176 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-64b4f4bf8b-vzb6r" event={"ID":"40e2ab14-9d05-4569-8a3a-69a34586c504","Type":"ContainerStarted","Data":"5d273c1e227076e3e25dfe47f85b80f860ff97ea602f3ec723a4890100fe0e64"} Nov 25 10:46:17 crc kubenswrapper[4821]: I1125 10:46:17.073234 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-64b4f4bf8b-vzb6r" event={"ID":"40e2ab14-9d05-4569-8a3a-69a34586c504","Type":"ContainerStarted","Data":"9a88009cd3c2a774c25512c5d679195e02be1541dd15807b666eeb211588d23a"} Nov 25 10:46:17 crc kubenswrapper[4821]: I1125 10:46:17.074414 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-64b4f4bf8b-vzb6r" Nov 25 10:46:17 crc kubenswrapper[4821]: I1125 10:46:17.074451 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-b248b"] Nov 25 10:46:17 crc kubenswrapper[4821]: I1125 10:46:17.079436 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-66kf4" event={"ID":"876c088c-7ee5-4b61-a798-eb36ac1e7e21","Type":"ContainerStarted","Data":"a350f5eb2f2e00b43d3300d91f24ffe8731a0a154a5014689c2f49c7559851e9"} Nov 25 10:46:17 crc kubenswrapper[4821]: I1125 10:46:17.086092 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-774b86978c-s49f6" event={"ID":"b35e42b0-2a6f-454f-b8e2-4f7cd9b9217f","Type":"ContainerStarted","Data":"c71508a32973db3f69ea746f4c3d80578c3737039619c7d4d64c990282e2266f"} Nov 25 10:46:17 crc kubenswrapper[4821]: I1125 10:46:17.109040 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-64b4f4bf8b-vzb6r" podStartSLOduration=18.10901909 podStartE2EDuration="18.10901909s" podCreationTimestamp="2025-11-25 10:45:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:46:17.104145616 +0000 UTC m=+847.640465453" watchObservedRunningTime="2025-11-25 10:46:17.10901909 +0000 UTC m=+847.645338937" Nov 25 10:46:17 crc kubenswrapper[4821]: I1125 10:46:17.604465 4821 scope.go:117] "RemoveContainer" containerID="725d705414501cce458da9324ba24945e4c24c36c1b5574625b32bff21475e53" Nov 25 10:46:17 crc kubenswrapper[4821]: I1125 10:46:17.860812 4821 scope.go:117] "RemoveContainer" containerID="f72d68759c12fed646c58d3fb778ab5e3e3dbc713cc2216b1873108a467d4080" Nov 25 10:46:18 crc kubenswrapper[4821]: I1125 10:46:18.097010 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-75q4l" event={"ID":"c77cb0f4-1dff-466c-a8bf-6a987bc87ced","Type":"ContainerStarted","Data":"e2209694521fcc24c00cd6f5fde8370d3fdcdffe99cdd43d426da7772aa1469f"} Nov 25 10:46:18 crc kubenswrapper[4821]: I1125 10:46:18.146565 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3efc5d6f-66aa-409b-adac-506291c1e810" path="/var/lib/kubelet/pods/3efc5d6f-66aa-409b-adac-506291c1e810/volumes" Nov 25 10:46:23 crc kubenswrapper[4821]: I1125 10:46:23.768235 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-64b4f4bf8b-vzb6r" Nov 25 10:46:24 crc kubenswrapper[4821]: I1125 10:46:24.141673 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-8cqfp" event={"ID":"9db84cac-a8df-4359-8051-ebc064a97b4d","Type":"ContainerStarted","Data":"3745330f57b43c6a54e2ee22428eb219b95de4e23b98679e1999c7eaf46cb608"} Nov 25 10:46:24 crc kubenswrapper[4821]: I1125 10:46:24.143218 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-5ncqq" event={"ID":"db1d55a6-1285-429d-b9dd-b2e100e602c6","Type":"ContainerStarted","Data":"0c0ab3650a5e27612d6be6fac24c1d3c48dacb9bd06c337fa8eeab877e0b9395"} Nov 25 10:46:24 crc kubenswrapper[4821]: I1125 10:46:24.146314 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-fprdh" event={"ID":"5cea2a2e-a19f-42fc-ac84-aadfa677ec4c","Type":"ContainerStarted","Data":"85632fa8153d132f689d699cb2805fd368ce04f07f50c8c276c95064b18e3d07"} Nov 25 10:46:24 crc kubenswrapper[4821]: I1125 10:46:24.171868 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-fprdh" podStartSLOduration=3.307356158 podStartE2EDuration="25.171853097s" podCreationTimestamp="2025-11-25 10:45:59 +0000 UTC" firstStartedPulling="2025-11-25 10:46:00.998920215 +0000 UTC m=+831.535240072" lastFinishedPulling="2025-11-25 10:46:22.863417164 +0000 UTC m=+853.399737011" observedRunningTime="2025-11-25 10:46:24.17079893 +0000 UTC m=+854.707118777" watchObservedRunningTime="2025-11-25 10:46:24.171853097 +0000 UTC m=+854.708172944" Nov 25 10:46:24 crc kubenswrapper[4821]: E1125 10:46:24.189208 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-fqfwh" podUID="1976151f-2b7a-4cde-95d3-9020b17d938c" Nov 25 10:46:24 crc kubenswrapper[4821]: E1125 10:46:24.197376 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/watcher-operator-controller-manager-864885998-vnmtq" podUID="505faacd-65f5-4078-9867-9ae648652b50" Nov 25 10:46:24 crc kubenswrapper[4821]: E1125 10:46:24.420820 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-p4bwc" podUID="8f16d64f-ea14-4853-b30b-2df294ce6cc3" Nov 25 10:46:25 crc kubenswrapper[4821]: I1125 10:46:25.154702 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-75q4l" event={"ID":"c77cb0f4-1dff-466c-a8bf-6a987bc87ced","Type":"ContainerStarted","Data":"235b484f5f4406f356342dd27013cd07412a21901f86eee33334e502ccc9e413"} Nov 25 10:46:25 crc kubenswrapper[4821]: I1125 10:46:25.155047 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-75q4l" Nov 25 10:46:25 crc kubenswrapper[4821]: I1125 10:46:25.157286 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-68b95954c9-qgs92" event={"ID":"a3dc5167-4997-44bf-ace3-706829fd0f17","Type":"ContainerStarted","Data":"5408226c5ed95166f343edadce874ed63553f4d4008376741cdfe35a74bce573"} Nov 25 10:46:25 crc kubenswrapper[4821]: I1125 10:46:25.157442 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-68b95954c9-qgs92" Nov 25 10:46:25 crc kubenswrapper[4821]: I1125 10:46:25.158727 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-75q4l" Nov 25 10:46:25 crc kubenswrapper[4821]: I1125 10:46:25.159847 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-vnf9x" event={"ID":"5fefd5f2-18cb-4c63-848f-2fc9d63bb828","Type":"ContainerStarted","Data":"7f53421aa3daba030b2548e190b0edecaac9500fe1f9c9aaee0b532cc2a9b565"} Nov 25 10:46:25 crc kubenswrapper[4821]: I1125 10:46:25.160216 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-vnf9x" Nov 25 10:46:25 crc kubenswrapper[4821]: I1125 10:46:25.162181 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-vnf9x" Nov 25 10:46:25 crc kubenswrapper[4821]: I1125 10:46:25.162455 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-774b86978c-s49f6" event={"ID":"b35e42b0-2a6f-454f-b8e2-4f7cd9b9217f","Type":"ContainerStarted","Data":"558de1a0e6821a3903e544ec951e5a31fbeedc94ba538f872404c78923625675"} Nov 25 10:46:25 crc kubenswrapper[4821]: I1125 10:46:25.163134 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-774b86978c-s49f6" Nov 25 10:46:25 crc kubenswrapper[4821]: I1125 10:46:25.163812 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-fqfwh" event={"ID":"1976151f-2b7a-4cde-95d3-9020b17d938c","Type":"ContainerStarted","Data":"6c246c17cd9ac874e2ecc69b35c46584b5ba39d691f2c9496034ec57f25225fc"} Nov 25 10:46:25 crc kubenswrapper[4821]: I1125 10:46:25.165425 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-774b86978c-s49f6" Nov 25 10:46:25 crc kubenswrapper[4821]: I1125 10:46:25.165602 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-68b95954c9-qgs92" Nov 25 10:46:25 crc kubenswrapper[4821]: I1125 10:46:25.166808 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-4ftbp" event={"ID":"1f69797f-4204-4989-a00b-d9722d44d3c6","Type":"ContainerStarted","Data":"f7347fc33b5df67a9df49ff94711b64e141a5f7dcde2d64b3e70ea8cf600e96e"} Nov 25 10:46:25 crc kubenswrapper[4821]: I1125 10:46:25.166843 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-4ftbp" event={"ID":"1f69797f-4204-4989-a00b-d9722d44d3c6","Type":"ContainerStarted","Data":"1bad5f5b8c034897bfdf2622821006d3fecf50e688fe17be9141e0326a8b0078"} Nov 25 10:46:25 crc kubenswrapper[4821]: I1125 10:46:25.167495 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-4ftbp" Nov 25 10:46:25 crc kubenswrapper[4821]: I1125 10:46:25.172603 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-f8tvv" event={"ID":"7982734e-b887-4cea-9724-3a37f9369ac5","Type":"ContainerStarted","Data":"07eec6f222ab80360ed08568a8fb88c0e096323091c3c104891f712bbeb9730f"} Nov 25 10:46:25 crc kubenswrapper[4821]: I1125 10:46:25.173263 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-f8tvv" Nov 25 10:46:25 crc kubenswrapper[4821]: I1125 10:46:25.175069 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-7sx9m" event={"ID":"2236607e-2a7b-4858-b5b9-c43e47c9375d","Type":"ContainerStarted","Data":"c594d6215e0b91a1d395a6691f0cc1c2d1cbb833c1ef159f11d935bb61dd21e6"} Nov 25 10:46:25 crc kubenswrapper[4821]: I1125 10:46:25.175305 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-7sx9m" Nov 25 10:46:25 crc kubenswrapper[4821]: I1125 10:46:25.175736 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-f8tvv" Nov 25 10:46:25 crc kubenswrapper[4821]: I1125 10:46:25.176844 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-2qzrg" event={"ID":"b7895321-27ec-4c75-aa5e-581fe46e0789","Type":"ContainerStarted","Data":"61c520b1d54270b7300a770de93b11ef4e079f1ce10ece36ebe6a6e08ac97cba"} Nov 25 10:46:25 crc kubenswrapper[4821]: I1125 10:46:25.177633 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-2qzrg" Nov 25 10:46:25 crc kubenswrapper[4821]: I1125 10:46:25.179279 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-7sx9m" Nov 25 10:46:25 crc kubenswrapper[4821]: I1125 10:46:25.179488 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cb74df96-lppg4" event={"ID":"8fef5c2e-676a-49b6-a10c-94abe226600a","Type":"ContainerStarted","Data":"0acea6f2ce9c2a6794c105f80e3c0c505afa390b01f97b4c4621b708d65b5c86"} Nov 25 10:46:25 crc kubenswrapper[4821]: I1125 10:46:25.179523 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cb74df96-lppg4" event={"ID":"8fef5c2e-676a-49b6-a10c-94abe226600a","Type":"ContainerStarted","Data":"43712382dfc7ab002d4e691902b2406c8ee5da5258480f707b97c21727fa67e8"} Nov 25 10:46:25 crc kubenswrapper[4821]: I1125 10:46:25.179699 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5cb74df96-lppg4" Nov 25 10:46:25 crc kubenswrapper[4821]: I1125 10:46:25.182458 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-8cqfp" event={"ID":"9db84cac-a8df-4359-8051-ebc064a97b4d","Type":"ContainerStarted","Data":"09a5354288fdb396d8ac85933853d1e1f4184a1a6b66bab1c9a988871b4dc635"} Nov 25 10:46:25 crc kubenswrapper[4821]: I1125 10:46:25.182589 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-8cqfp" Nov 25 10:46:25 crc kubenswrapper[4821]: I1125 10:46:25.183922 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-75q4l" podStartSLOduration=3.465080299 podStartE2EDuration="26.183903799s" podCreationTimestamp="2025-11-25 10:45:59 +0000 UTC" firstStartedPulling="2025-11-25 10:46:00.907006425 +0000 UTC m=+831.443326272" lastFinishedPulling="2025-11-25 10:46:23.625829925 +0000 UTC m=+854.162149772" observedRunningTime="2025-11-25 10:46:25.178570363 +0000 UTC m=+855.714890210" watchObservedRunningTime="2025-11-25 10:46:25.183903799 +0000 UTC m=+855.720223646" Nov 25 10:46:25 crc kubenswrapper[4821]: I1125 10:46:25.184896 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-pnq4j" event={"ID":"db28ec64-d8be-4b56-a427-e4cceb256fd0","Type":"ContainerStarted","Data":"a2140ee7f11b734eaa6b5eca792b5cb78550b1cfc71e3922043a9d4aba2ec790"} Nov 25 10:46:25 crc kubenswrapper[4821]: I1125 10:46:25.185634 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-pnq4j" Nov 25 10:46:25 crc kubenswrapper[4821]: I1125 10:46:25.190417 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-pnq4j" Nov 25 10:46:25 crc kubenswrapper[4821]: I1125 10:46:25.190706 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-8dbc5685d-gcl85" event={"ID":"a64e88ce-fe1a-4cb1-984a-c6d8b5e5ffd9","Type":"ContainerStarted","Data":"98f0a588b5ebb44503defc350a62cda135edd4f42f35070044c6f4c0b1dfc77a"} Nov 25 10:46:25 crc kubenswrapper[4821]: I1125 10:46:25.190756 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-8dbc5685d-gcl85" event={"ID":"a64e88ce-fe1a-4cb1-984a-c6d8b5e5ffd9","Type":"ContainerStarted","Data":"31074175cdda4b8084d9feb7f3d761569f4fdc5e90caa9ba92f5b2529678a949"} Nov 25 10:46:25 crc kubenswrapper[4821]: I1125 10:46:25.190937 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-8dbc5685d-gcl85" Nov 25 10:46:25 crc kubenswrapper[4821]: I1125 10:46:25.197428 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-2qzrg" Nov 25 10:46:25 crc kubenswrapper[4821]: I1125 10:46:25.197734 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-fw96v" event={"ID":"d42d0991-9e6f-46d0-9236-00e85c26a603","Type":"ContainerStarted","Data":"9a047ec2f1b01f0a3c9422304b4335cc041b3808b663ef3afae9980a2aac4902"} Nov 25 10:46:25 crc kubenswrapper[4821]: I1125 10:46:25.197941 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-fw96v" Nov 25 10:46:25 crc kubenswrapper[4821]: I1125 10:46:25.202721 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-fw96v" Nov 25 10:46:25 crc kubenswrapper[4821]: I1125 10:46:25.203936 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-hwp4x" event={"ID":"c0376a5f-ed42-4399-8f41-2d2157a654fd","Type":"ContainerStarted","Data":"52afafdb42579b48e666f0fe67a58e365cfdf447229d452a2fedab1dc96e1ca8"} Nov 25 10:46:25 crc kubenswrapper[4821]: I1125 10:46:25.204188 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-hwp4x" Nov 25 10:46:25 crc kubenswrapper[4821]: I1125 10:46:25.206055 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-864885998-vnmtq" event={"ID":"505faacd-65f5-4078-9867-9ae648652b50","Type":"ContainerStarted","Data":"2cc7ea38366c72c8981cbce3716cd30fe573efd356a777728956691aff2bb07e"} Nov 25 10:46:25 crc kubenswrapper[4821]: E1125 10:46:25.206974 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-864885998-vnmtq" podUID="505faacd-65f5-4078-9867-9ae648652b50" Nov 25 10:46:25 crc kubenswrapper[4821]: I1125 10:46:25.207434 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-hwp4x" Nov 25 10:46:25 crc kubenswrapper[4821]: I1125 10:46:25.212135 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-p4bwc" event={"ID":"8f16d64f-ea14-4853-b30b-2df294ce6cc3","Type":"ContainerStarted","Data":"986d0ecfb4f3ca1fd62a46afff3ef46c41ab4c5de149189920e985ea86f35f30"} Nov 25 10:46:25 crc kubenswrapper[4821]: E1125 10:46:25.216405 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/cinder-operator@sha256:553b1288b330ad05771d59c6b73c1681c95f457e8475682f9ad0d2e6b85f37e9\\\"\"" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-p4bwc" podUID="8f16d64f-ea14-4853-b30b-2df294ce6cc3" Nov 25 10:46:25 crc kubenswrapper[4821]: I1125 10:46:25.228698 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-qxgcp" event={"ID":"2b44040a-e224-4468-a528-9f96a98e3891","Type":"ContainerStarted","Data":"48dae76d7bd5e5e1e6d3906512788d6a440ae019dcd290253b199b432da4aa9b"} Nov 25 10:46:25 crc kubenswrapper[4821]: I1125 10:46:25.228741 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-qxgcp" event={"ID":"2b44040a-e224-4468-a528-9f96a98e3891","Type":"ContainerStarted","Data":"ee2ddd78e91c004475af95404dde8581207d4f9c23e02a1aa8b7d072f692e2de"} Nov 25 10:46:25 crc kubenswrapper[4821]: I1125 10:46:25.228967 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-qxgcp" Nov 25 10:46:25 crc kubenswrapper[4821]: I1125 10:46:25.235352 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-66kf4" event={"ID":"876c088c-7ee5-4b61-a798-eb36ac1e7e21","Type":"ContainerStarted","Data":"62185a9c08fd81c72d8288cebdd1f030e6fe55e1ceff7967b429971d3d4b452a"} Nov 25 10:46:25 crc kubenswrapper[4821]: I1125 10:46:25.235456 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-66kf4" Nov 25 10:46:25 crc kubenswrapper[4821]: I1125 10:46:25.238377 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-66kf4" Nov 25 10:46:25 crc kubenswrapper[4821]: I1125 10:46:25.247892 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-lwp9z" event={"ID":"14385dd9-9b06-473b-919d-9d2f9207ffe1","Type":"ContainerStarted","Data":"0fedfbbfb1fcbd6170c8c76e41b7baba15437c8c4003fc8e4bbd99be3795deaf"} Nov 25 10:46:25 crc kubenswrapper[4821]: I1125 10:46:25.247939 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-lwp9z" event={"ID":"14385dd9-9b06-473b-919d-9d2f9207ffe1","Type":"ContainerStarted","Data":"c710739953388663b84d8d487280f3bdffc9bb7306955871cf115e123301339e"} Nov 25 10:46:25 crc kubenswrapper[4821]: I1125 10:46:25.248654 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-lwp9z" Nov 25 10:46:25 crc kubenswrapper[4821]: I1125 10:46:25.253264 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-f8tvv" podStartSLOduration=4.210887385 podStartE2EDuration="27.253250558s" podCreationTimestamp="2025-11-25 10:45:58 +0000 UTC" firstStartedPulling="2025-11-25 10:46:00.483793415 +0000 UTC m=+831.020113262" lastFinishedPulling="2025-11-25 10:46:23.526156578 +0000 UTC m=+854.062476435" observedRunningTime="2025-11-25 10:46:25.224212833 +0000 UTC m=+855.760532680" watchObservedRunningTime="2025-11-25 10:46:25.253250558 +0000 UTC m=+855.789570405" Nov 25 10:46:25 crc kubenswrapper[4821]: I1125 10:46:25.259551 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-774b86978c-s49f6" podStartSLOduration=4.149421047 podStartE2EDuration="27.259530908s" podCreationTimestamp="2025-11-25 10:45:58 +0000 UTC" firstStartedPulling="2025-11-25 10:46:00.483934549 +0000 UTC m=+831.020254406" lastFinishedPulling="2025-11-25 10:46:23.59404442 +0000 UTC m=+854.130364267" observedRunningTime="2025-11-25 10:46:25.251827331 +0000 UTC m=+855.788147188" watchObservedRunningTime="2025-11-25 10:46:25.259530908 +0000 UTC m=+855.795850765" Nov 25 10:46:25 crc kubenswrapper[4821]: I1125 10:46:25.267886 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-5ncqq" event={"ID":"db1d55a6-1285-429d-b9dd-b2e100e602c6","Type":"ContainerStarted","Data":"7e776a4e43534577a4b5d3919a76ed83ccf80ec133de70bdd24b554d63d91020"} Nov 25 10:46:25 crc kubenswrapper[4821]: I1125 10:46:25.268779 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-5ncqq" Nov 25 10:46:25 crc kubenswrapper[4821]: I1125 10:46:25.330477 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-2qzrg" podStartSLOduration=4.273067286 podStartE2EDuration="27.330461548s" podCreationTimestamp="2025-11-25 10:45:58 +0000 UTC" firstStartedPulling="2025-11-25 10:46:00.562231184 +0000 UTC m=+831.098551031" lastFinishedPulling="2025-11-25 10:46:23.619625446 +0000 UTC m=+854.155945293" observedRunningTime="2025-11-25 10:46:25.302616003 +0000 UTC m=+855.838935850" watchObservedRunningTime="2025-11-25 10:46:25.330461548 +0000 UTC m=+855.866781415" Nov 25 10:46:25 crc kubenswrapper[4821]: I1125 10:46:25.331052 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-4ftbp" podStartSLOduration=3.781434571 podStartE2EDuration="26.331044073s" podCreationTimestamp="2025-11-25 10:45:59 +0000 UTC" firstStartedPulling="2025-11-25 10:46:00.97666474 +0000 UTC m=+831.512984587" lastFinishedPulling="2025-11-25 10:46:23.526274232 +0000 UTC m=+854.062594089" observedRunningTime="2025-11-25 10:46:25.324843984 +0000 UTC m=+855.861163831" watchObservedRunningTime="2025-11-25 10:46:25.331044073 +0000 UTC m=+855.867363940" Nov 25 10:46:25 crc kubenswrapper[4821]: I1125 10:46:25.376089 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-7sx9m" podStartSLOduration=4.392603327 podStartE2EDuration="27.376069178s" podCreationTimestamp="2025-11-25 10:45:58 +0000 UTC" firstStartedPulling="2025-11-25 10:46:00.341369896 +0000 UTC m=+830.877689743" lastFinishedPulling="2025-11-25 10:46:23.324835727 +0000 UTC m=+853.861155594" observedRunningTime="2025-11-25 10:46:25.348152052 +0000 UTC m=+855.884471889" watchObservedRunningTime="2025-11-25 10:46:25.376069178 +0000 UTC m=+855.912389025" Nov 25 10:46:25 crc kubenswrapper[4821]: I1125 10:46:25.404047 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-vnf9x" podStartSLOduration=4.784050053 podStartE2EDuration="27.404029425s" podCreationTimestamp="2025-11-25 10:45:58 +0000 UTC" firstStartedPulling="2025-11-25 10:46:00.929330881 +0000 UTC m=+831.465650728" lastFinishedPulling="2025-11-25 10:46:23.549310243 +0000 UTC m=+854.085630100" observedRunningTime="2025-11-25 10:46:25.378250643 +0000 UTC m=+855.914570490" watchObservedRunningTime="2025-11-25 10:46:25.404029425 +0000 UTC m=+855.940349272" Nov 25 10:46:25 crc kubenswrapper[4821]: I1125 10:46:25.408019 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-68b95954c9-qgs92" podStartSLOduration=4.077506157 podStartE2EDuration="27.408003137s" podCreationTimestamp="2025-11-25 10:45:58 +0000 UTC" firstStartedPulling="2025-11-25 10:46:00.308297377 +0000 UTC m=+830.844617224" lastFinishedPulling="2025-11-25 10:46:23.638794357 +0000 UTC m=+854.175114204" observedRunningTime="2025-11-25 10:46:25.403524321 +0000 UTC m=+855.939844168" watchObservedRunningTime="2025-11-25 10:46:25.408003137 +0000 UTC m=+855.944322984" Nov 25 10:46:25 crc kubenswrapper[4821]: I1125 10:46:25.476912 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-lwp9z" podStartSLOduration=19.748458551 podStartE2EDuration="26.476895793s" podCreationTimestamp="2025-11-25 10:45:59 +0000 UTC" firstStartedPulling="2025-11-25 10:46:16.657892212 +0000 UTC m=+847.194212059" lastFinishedPulling="2025-11-25 10:46:23.386329424 +0000 UTC m=+853.922649301" observedRunningTime="2025-11-25 10:46:25.476292797 +0000 UTC m=+856.012612644" watchObservedRunningTime="2025-11-25 10:46:25.476895793 +0000 UTC m=+856.013215640" Nov 25 10:46:25 crc kubenswrapper[4821]: I1125 10:46:25.511172 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-8cqfp" podStartSLOduration=5.18181112 podStartE2EDuration="27.511140171s" podCreationTimestamp="2025-11-25 10:45:58 +0000 UTC" firstStartedPulling="2025-11-25 10:46:00.97700257 +0000 UTC m=+831.513322407" lastFinishedPulling="2025-11-25 10:46:23.306331611 +0000 UTC m=+853.842651458" observedRunningTime="2025-11-25 10:46:25.501401501 +0000 UTC m=+856.037721358" watchObservedRunningTime="2025-11-25 10:46:25.511140171 +0000 UTC m=+856.047460018" Nov 25 10:46:25 crc kubenswrapper[4821]: I1125 10:46:25.521815 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-pnq4j" podStartSLOduration=4.071352729 podStartE2EDuration="27.521800795s" podCreationTimestamp="2025-11-25 10:45:58 +0000 UTC" firstStartedPulling="2025-11-25 10:46:00.139723864 +0000 UTC m=+830.676043711" lastFinishedPulling="2025-11-25 10:46:23.5901719 +0000 UTC m=+854.126491777" observedRunningTime="2025-11-25 10:46:25.51967758 +0000 UTC m=+856.055997427" watchObservedRunningTime="2025-11-25 10:46:25.521800795 +0000 UTC m=+856.058120642" Nov 25 10:46:25 crc kubenswrapper[4821]: I1125 10:46:25.577741 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-66kf4" podStartSLOduration=3.852905526 podStartE2EDuration="26.577727669s" podCreationTimestamp="2025-11-25 10:45:59 +0000 UTC" firstStartedPulling="2025-11-25 10:46:00.906372609 +0000 UTC m=+831.442692456" lastFinishedPulling="2025-11-25 10:46:23.631194752 +0000 UTC m=+854.167514599" observedRunningTime="2025-11-25 10:46:25.573549802 +0000 UTC m=+856.109869649" watchObservedRunningTime="2025-11-25 10:46:25.577727669 +0000 UTC m=+856.114047516" Nov 25 10:46:25 crc kubenswrapper[4821]: I1125 10:46:25.580099 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5cb74df96-lppg4" podStartSLOduration=4.030040966 podStartE2EDuration="26.580093429s" podCreationTimestamp="2025-11-25 10:45:59 +0000 UTC" firstStartedPulling="2025-11-25 10:46:00.976699681 +0000 UTC m=+831.513019528" lastFinishedPulling="2025-11-25 10:46:23.526752134 +0000 UTC m=+854.063071991" observedRunningTime="2025-11-25 10:46:25.545068471 +0000 UTC m=+856.081388318" watchObservedRunningTime="2025-11-25 10:46:25.580093429 +0000 UTC m=+856.116413276" Nov 25 10:46:25 crc kubenswrapper[4821]: I1125 10:46:25.631647 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-5ncqq" podStartSLOduration=4.269093761 podStartE2EDuration="26.631629881s" podCreationTimestamp="2025-11-25 10:45:59 +0000 UTC" firstStartedPulling="2025-11-25 10:46:00.962270106 +0000 UTC m=+831.498589953" lastFinishedPulling="2025-11-25 10:46:23.324806236 +0000 UTC m=+853.861126073" observedRunningTime="2025-11-25 10:46:25.631129958 +0000 UTC m=+856.167449805" watchObservedRunningTime="2025-11-25 10:46:25.631629881 +0000 UTC m=+856.167949728" Nov 25 10:46:25 crc kubenswrapper[4821]: I1125 10:46:25.679537 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-fw96v" podStartSLOduration=4.035255298 podStartE2EDuration="26.679522899s" podCreationTimestamp="2025-11-25 10:45:59 +0000 UTC" firstStartedPulling="2025-11-25 10:46:00.948201279 +0000 UTC m=+831.484521126" lastFinishedPulling="2025-11-25 10:46:23.59246888 +0000 UTC m=+854.128788727" observedRunningTime="2025-11-25 10:46:25.678272317 +0000 UTC m=+856.214592164" watchObservedRunningTime="2025-11-25 10:46:25.679522899 +0000 UTC m=+856.215842746" Nov 25 10:46:25 crc kubenswrapper[4821]: I1125 10:46:25.702083 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-qxgcp" podStartSLOduration=4.095499366 podStartE2EDuration="26.702067758s" podCreationTimestamp="2025-11-25 10:45:59 +0000 UTC" firstStartedPulling="2025-11-25 10:46:00.977128583 +0000 UTC m=+831.513448430" lastFinishedPulling="2025-11-25 10:46:23.583696975 +0000 UTC m=+854.120016822" observedRunningTime="2025-11-25 10:46:25.700872637 +0000 UTC m=+856.237192484" watchObservedRunningTime="2025-11-25 10:46:25.702067758 +0000 UTC m=+856.238387605" Nov 25 10:46:25 crc kubenswrapper[4821]: I1125 10:46:25.738806 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-8dbc5685d-gcl85" podStartSLOduration=4.09569777 podStartE2EDuration="26.738783619s" podCreationTimestamp="2025-11-25 10:45:59 +0000 UTC" firstStartedPulling="2025-11-25 10:46:00.977238495 +0000 UTC m=+831.513558342" lastFinishedPulling="2025-11-25 10:46:23.620324344 +0000 UTC m=+854.156644191" observedRunningTime="2025-11-25 10:46:25.730220029 +0000 UTC m=+856.266539876" watchObservedRunningTime="2025-11-25 10:46:25.738783619 +0000 UTC m=+856.275103466" Nov 25 10:46:26 crc kubenswrapper[4821]: I1125 10:46:26.276004 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-fqfwh" event={"ID":"1976151f-2b7a-4cde-95d3-9020b17d938c","Type":"ContainerStarted","Data":"564a71a49342b04efb57b9da24f46b8f7e1840df18bf0cd5299703371cc48e65"} Nov 25 10:46:26 crc kubenswrapper[4821]: I1125 10:46:26.298890 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-hwp4x" podStartSLOduration=4.656496303 podStartE2EDuration="27.298873732s" podCreationTimestamp="2025-11-25 10:45:59 +0000 UTC" firstStartedPulling="2025-11-25 10:46:00.941182451 +0000 UTC m=+831.477502298" lastFinishedPulling="2025-11-25 10:46:23.58355988 +0000 UTC m=+854.119879727" observedRunningTime="2025-11-25 10:46:25.781388471 +0000 UTC m=+856.317708318" watchObservedRunningTime="2025-11-25 10:46:26.298873732 +0000 UTC m=+856.835193579" Nov 25 10:46:26 crc kubenswrapper[4821]: I1125 10:46:26.331475 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-fqfwh" podStartSLOduration=3.22267887 podStartE2EDuration="28.331457497s" podCreationTimestamp="2025-11-25 10:45:58 +0000 UTC" firstStartedPulling="2025-11-25 10:46:00.459234123 +0000 UTC m=+830.995553960" lastFinishedPulling="2025-11-25 10:46:25.56801274 +0000 UTC m=+856.104332587" observedRunningTime="2025-11-25 10:46:26.326176011 +0000 UTC m=+856.862495858" watchObservedRunningTime="2025-11-25 10:46:26.331457497 +0000 UTC m=+856.867777344" Nov 25 10:46:27 crc kubenswrapper[4821]: I1125 10:46:27.284823 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-864885998-vnmtq" event={"ID":"505faacd-65f5-4078-9867-9ae648652b50","Type":"ContainerStarted","Data":"828f863abf7563127314b05f39f02119f6e8b159986d5a99c11438c0f8e8d3fa"} Nov 25 10:46:27 crc kubenswrapper[4821]: I1125 10:46:27.285094 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-864885998-vnmtq" Nov 25 10:46:27 crc kubenswrapper[4821]: I1125 10:46:27.287325 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-p4bwc" event={"ID":"8f16d64f-ea14-4853-b30b-2df294ce6cc3","Type":"ContainerStarted","Data":"17ff6114f7de7b7be1ec990c89120a49cf0d46b76c9f5ebc2d5b15102b2c532b"} Nov 25 10:46:27 crc kubenswrapper[4821]: I1125 10:46:27.288472 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-fqfwh" Nov 25 10:46:27 crc kubenswrapper[4821]: I1125 10:46:27.319414 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-p4bwc" podStartSLOduration=3.159354839 podStartE2EDuration="29.319395852s" podCreationTimestamp="2025-11-25 10:45:58 +0000 UTC" firstStartedPulling="2025-11-25 10:46:00.554615241 +0000 UTC m=+831.090935088" lastFinishedPulling="2025-11-25 10:46:26.714656254 +0000 UTC m=+857.250976101" observedRunningTime="2025-11-25 10:46:27.31858168 +0000 UTC m=+857.854901527" watchObservedRunningTime="2025-11-25 10:46:27.319395852 +0000 UTC m=+857.855715699" Nov 25 10:46:27 crc kubenswrapper[4821]: I1125 10:46:27.319498 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-864885998-vnmtq" podStartSLOduration=2.5616949890000003 podStartE2EDuration="28.319495364s" podCreationTimestamp="2025-11-25 10:45:59 +0000 UTC" firstStartedPulling="2025-11-25 10:46:00.955838512 +0000 UTC m=+831.492158359" lastFinishedPulling="2025-11-25 10:46:26.713638887 +0000 UTC m=+857.249958734" observedRunningTime="2025-11-25 10:46:27.302946869 +0000 UTC m=+857.839266716" watchObservedRunningTime="2025-11-25 10:46:27.319495364 +0000 UTC m=+857.855815211" Nov 25 10:46:29 crc kubenswrapper[4821]: I1125 10:46:29.161495 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-p4bwc" Nov 25 10:46:29 crc kubenswrapper[4821]: I1125 10:46:29.587833 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-5ncqq" Nov 25 10:46:29 crc kubenswrapper[4821]: I1125 10:46:29.608127 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-4ftbp" Nov 25 10:46:29 crc kubenswrapper[4821]: I1125 10:46:29.682559 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-qxgcp" Nov 25 10:46:29 crc kubenswrapper[4821]: I1125 10:46:29.819460 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5cb74df96-lppg4" Nov 25 10:46:29 crc kubenswrapper[4821]: I1125 10:46:29.890532 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-8dbc5685d-gcl85" Nov 25 10:46:29 crc kubenswrapper[4821]: I1125 10:46:29.997549 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-8cqfp" Nov 25 10:46:33 crc kubenswrapper[4821]: I1125 10:46:33.244415 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-lwp9z" Nov 25 10:46:39 crc kubenswrapper[4821]: I1125 10:46:39.163869 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-p4bwc" Nov 25 10:46:39 crc kubenswrapper[4821]: I1125 10:46:39.419476 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-fqfwh" Nov 25 10:46:39 crc kubenswrapper[4821]: I1125 10:46:39.831762 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-864885998-vnmtq" Nov 25 10:46:40 crc kubenswrapper[4821]: I1125 10:46:40.159673 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8phxk"] Nov 25 10:46:40 crc kubenswrapper[4821]: E1125 10:46:40.160046 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3efc5d6f-66aa-409b-adac-506291c1e810" containerName="extract-content" Nov 25 10:46:40 crc kubenswrapper[4821]: I1125 10:46:40.160061 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="3efc5d6f-66aa-409b-adac-506291c1e810" containerName="extract-content" Nov 25 10:46:40 crc kubenswrapper[4821]: E1125 10:46:40.160094 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3efc5d6f-66aa-409b-adac-506291c1e810" containerName="registry-server" Nov 25 10:46:40 crc kubenswrapper[4821]: I1125 10:46:40.160103 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="3efc5d6f-66aa-409b-adac-506291c1e810" containerName="registry-server" Nov 25 10:46:40 crc kubenswrapper[4821]: E1125 10:46:40.160123 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd8b5398-7942-4c13-8840-a53ff1aefcbf" containerName="extract-content" Nov 25 10:46:40 crc kubenswrapper[4821]: I1125 10:46:40.160131 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd8b5398-7942-4c13-8840-a53ff1aefcbf" containerName="extract-content" Nov 25 10:46:40 crc kubenswrapper[4821]: E1125 10:46:40.160148 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3efc5d6f-66aa-409b-adac-506291c1e810" containerName="extract-utilities" Nov 25 10:46:40 crc kubenswrapper[4821]: I1125 10:46:40.160156 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="3efc5d6f-66aa-409b-adac-506291c1e810" containerName="extract-utilities" Nov 25 10:46:40 crc kubenswrapper[4821]: E1125 10:46:40.160196 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd8b5398-7942-4c13-8840-a53ff1aefcbf" containerName="registry-server" Nov 25 10:46:40 crc kubenswrapper[4821]: I1125 10:46:40.160208 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd8b5398-7942-4c13-8840-a53ff1aefcbf" containerName="registry-server" Nov 25 10:46:40 crc kubenswrapper[4821]: E1125 10:46:40.160222 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd8b5398-7942-4c13-8840-a53ff1aefcbf" containerName="extract-utilities" Nov 25 10:46:40 crc kubenswrapper[4821]: I1125 10:46:40.160228 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd8b5398-7942-4c13-8840-a53ff1aefcbf" containerName="extract-utilities" Nov 25 10:46:40 crc kubenswrapper[4821]: I1125 10:46:40.160403 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="3efc5d6f-66aa-409b-adac-506291c1e810" containerName="registry-server" Nov 25 10:46:40 crc kubenswrapper[4821]: I1125 10:46:40.160430 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd8b5398-7942-4c13-8840-a53ff1aefcbf" containerName="registry-server" Nov 25 10:46:40 crc kubenswrapper[4821]: I1125 10:46:40.161818 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8phxk" Nov 25 10:46:40 crc kubenswrapper[4821]: I1125 10:46:40.188002 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8phxk"] Nov 25 10:46:40 crc kubenswrapper[4821]: I1125 10:46:40.268575 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e93e112b-d64e-4bab-b62d-cfee3e1a74eb-catalog-content\") pod \"certified-operators-8phxk\" (UID: \"e93e112b-d64e-4bab-b62d-cfee3e1a74eb\") " pod="openshift-marketplace/certified-operators-8phxk" Nov 25 10:46:40 crc kubenswrapper[4821]: I1125 10:46:40.268686 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4l272\" (UniqueName: \"kubernetes.io/projected/e93e112b-d64e-4bab-b62d-cfee3e1a74eb-kube-api-access-4l272\") pod \"certified-operators-8phxk\" (UID: \"e93e112b-d64e-4bab-b62d-cfee3e1a74eb\") " pod="openshift-marketplace/certified-operators-8phxk" Nov 25 10:46:40 crc kubenswrapper[4821]: I1125 10:46:40.268735 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e93e112b-d64e-4bab-b62d-cfee3e1a74eb-utilities\") pod \"certified-operators-8phxk\" (UID: \"e93e112b-d64e-4bab-b62d-cfee3e1a74eb\") " pod="openshift-marketplace/certified-operators-8phxk" Nov 25 10:46:40 crc kubenswrapper[4821]: I1125 10:46:40.369649 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e93e112b-d64e-4bab-b62d-cfee3e1a74eb-catalog-content\") pod \"certified-operators-8phxk\" (UID: \"e93e112b-d64e-4bab-b62d-cfee3e1a74eb\") " pod="openshift-marketplace/certified-operators-8phxk" Nov 25 10:46:40 crc kubenswrapper[4821]: I1125 10:46:40.369691 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4l272\" (UniqueName: \"kubernetes.io/projected/e93e112b-d64e-4bab-b62d-cfee3e1a74eb-kube-api-access-4l272\") pod \"certified-operators-8phxk\" (UID: \"e93e112b-d64e-4bab-b62d-cfee3e1a74eb\") " pod="openshift-marketplace/certified-operators-8phxk" Nov 25 10:46:40 crc kubenswrapper[4821]: I1125 10:46:40.369713 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e93e112b-d64e-4bab-b62d-cfee3e1a74eb-utilities\") pod \"certified-operators-8phxk\" (UID: \"e93e112b-d64e-4bab-b62d-cfee3e1a74eb\") " pod="openshift-marketplace/certified-operators-8phxk" Nov 25 10:46:40 crc kubenswrapper[4821]: I1125 10:46:40.370207 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e93e112b-d64e-4bab-b62d-cfee3e1a74eb-catalog-content\") pod \"certified-operators-8phxk\" (UID: \"e93e112b-d64e-4bab-b62d-cfee3e1a74eb\") " pod="openshift-marketplace/certified-operators-8phxk" Nov 25 10:46:40 crc kubenswrapper[4821]: I1125 10:46:40.370265 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e93e112b-d64e-4bab-b62d-cfee3e1a74eb-utilities\") pod \"certified-operators-8phxk\" (UID: \"e93e112b-d64e-4bab-b62d-cfee3e1a74eb\") " pod="openshift-marketplace/certified-operators-8phxk" Nov 25 10:46:40 crc kubenswrapper[4821]: I1125 10:46:40.390844 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4l272\" (UniqueName: \"kubernetes.io/projected/e93e112b-d64e-4bab-b62d-cfee3e1a74eb-kube-api-access-4l272\") pod \"certified-operators-8phxk\" (UID: \"e93e112b-d64e-4bab-b62d-cfee3e1a74eb\") " pod="openshift-marketplace/certified-operators-8phxk" Nov 25 10:46:40 crc kubenswrapper[4821]: I1125 10:46:40.482114 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8phxk" Nov 25 10:46:40 crc kubenswrapper[4821]: I1125 10:46:40.925081 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8phxk"] Nov 25 10:46:40 crc kubenswrapper[4821]: W1125 10:46:40.931348 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode93e112b_d64e_4bab_b62d_cfee3e1a74eb.slice/crio-97514406c4d204378d6867d011b122723a510d456aeb008ebbd1442cc6656fb1 WatchSource:0}: Error finding container 97514406c4d204378d6867d011b122723a510d456aeb008ebbd1442cc6656fb1: Status 404 returned error can't find the container with id 97514406c4d204378d6867d011b122723a510d456aeb008ebbd1442cc6656fb1 Nov 25 10:46:41 crc kubenswrapper[4821]: I1125 10:46:41.405531 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8phxk" event={"ID":"e93e112b-d64e-4bab-b62d-cfee3e1a74eb","Type":"ContainerStarted","Data":"97514406c4d204378d6867d011b122723a510d456aeb008ebbd1442cc6656fb1"} Nov 25 10:46:47 crc kubenswrapper[4821]: I1125 10:46:47.452140 4821 generic.go:334] "Generic (PLEG): container finished" podID="e93e112b-d64e-4bab-b62d-cfee3e1a74eb" containerID="c8df2db54459c41b5438811f629d46a12cedcf62796b6e4dcdb43e6eefecedc6" exitCode=0 Nov 25 10:46:47 crc kubenswrapper[4821]: I1125 10:46:47.452220 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8phxk" event={"ID":"e93e112b-d64e-4bab-b62d-cfee3e1a74eb","Type":"ContainerDied","Data":"c8df2db54459c41b5438811f629d46a12cedcf62796b6e4dcdb43e6eefecedc6"} Nov 25 10:46:48 crc kubenswrapper[4821]: I1125 10:46:48.474933 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8phxk" event={"ID":"e93e112b-d64e-4bab-b62d-cfee3e1a74eb","Type":"ContainerStarted","Data":"37c74a014dadf4083716ab00ae6e0e26e84722e23c5ddd3ddc9d04b54a9a247d"} Nov 25 10:46:49 crc kubenswrapper[4821]: I1125 10:46:49.483579 4821 generic.go:334] "Generic (PLEG): container finished" podID="e93e112b-d64e-4bab-b62d-cfee3e1a74eb" containerID="37c74a014dadf4083716ab00ae6e0e26e84722e23c5ddd3ddc9d04b54a9a247d" exitCode=0 Nov 25 10:46:49 crc kubenswrapper[4821]: I1125 10:46:49.483681 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8phxk" event={"ID":"e93e112b-d64e-4bab-b62d-cfee3e1a74eb","Type":"ContainerDied","Data":"37c74a014dadf4083716ab00ae6e0e26e84722e23c5ddd3ddc9d04b54a9a247d"} Nov 25 10:46:50 crc kubenswrapper[4821]: I1125 10:46:50.496098 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8phxk" event={"ID":"e93e112b-d64e-4bab-b62d-cfee3e1a74eb","Type":"ContainerStarted","Data":"afb0d51c47960d814a1cea6065cbd96a29f2014cf8af36ff98b91e9fa73c658e"} Nov 25 10:46:50 crc kubenswrapper[4821]: I1125 10:46:50.527729 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8phxk" podStartSLOduration=8.084514625 podStartE2EDuration="10.527705747s" podCreationTimestamp="2025-11-25 10:46:40 +0000 UTC" firstStartedPulling="2025-11-25 10:46:47.454109129 +0000 UTC m=+877.990428976" lastFinishedPulling="2025-11-25 10:46:49.897300251 +0000 UTC m=+880.433620098" observedRunningTime="2025-11-25 10:46:50.52355113 +0000 UTC m=+881.059870987" watchObservedRunningTime="2025-11-25 10:46:50.527705747 +0000 UTC m=+881.064025594" Nov 25 10:46:56 crc kubenswrapper[4821]: I1125 10:46:56.480574 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-shhlw"] Nov 25 10:46:56 crc kubenswrapper[4821]: I1125 10:46:56.499764 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-shhlw" Nov 25 10:46:56 crc kubenswrapper[4821]: I1125 10:46:56.518230 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-shhlw"] Nov 25 10:46:56 crc kubenswrapper[4821]: I1125 10:46:56.518601 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Nov 25 10:46:56 crc kubenswrapper[4821]: I1125 10:46:56.518829 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Nov 25 10:46:56 crc kubenswrapper[4821]: I1125 10:46:56.518950 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Nov 25 10:46:56 crc kubenswrapper[4821]: I1125 10:46:56.519097 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-zt27k" Nov 25 10:46:56 crc kubenswrapper[4821]: I1125 10:46:56.582059 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-srdjw"] Nov 25 10:46:56 crc kubenswrapper[4821]: I1125 10:46:56.583761 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-srdjw" Nov 25 10:46:56 crc kubenswrapper[4821]: I1125 10:46:56.585897 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Nov 25 10:46:56 crc kubenswrapper[4821]: I1125 10:46:56.590626 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-srdjw"] Nov 25 10:46:56 crc kubenswrapper[4821]: I1125 10:46:56.595782 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99bb9\" (UniqueName: \"kubernetes.io/projected/b7a3ac30-3925-493e-9e13-0f1a9a7e49c2-kube-api-access-99bb9\") pod \"dnsmasq-dns-675f4bcbfc-shhlw\" (UID: \"b7a3ac30-3925-493e-9e13-0f1a9a7e49c2\") " pod="openstack/dnsmasq-dns-675f4bcbfc-shhlw" Nov 25 10:46:56 crc kubenswrapper[4821]: I1125 10:46:56.595845 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7a3ac30-3925-493e-9e13-0f1a9a7e49c2-config\") pod \"dnsmasq-dns-675f4bcbfc-shhlw\" (UID: \"b7a3ac30-3925-493e-9e13-0f1a9a7e49c2\") " pod="openstack/dnsmasq-dns-675f4bcbfc-shhlw" Nov 25 10:46:56 crc kubenswrapper[4821]: I1125 10:46:56.696962 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xqr4f\" (UniqueName: \"kubernetes.io/projected/68a5afc7-5e47-4bed-a6c2-8ff5b6b0236c-kube-api-access-xqr4f\") pod \"dnsmasq-dns-78dd6ddcc-srdjw\" (UID: \"68a5afc7-5e47-4bed-a6c2-8ff5b6b0236c\") " pod="openstack/dnsmasq-dns-78dd6ddcc-srdjw" Nov 25 10:46:56 crc kubenswrapper[4821]: I1125 10:46:56.697044 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/68a5afc7-5e47-4bed-a6c2-8ff5b6b0236c-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-srdjw\" (UID: \"68a5afc7-5e47-4bed-a6c2-8ff5b6b0236c\") " pod="openstack/dnsmasq-dns-78dd6ddcc-srdjw" Nov 25 10:46:56 crc kubenswrapper[4821]: I1125 10:46:56.697092 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99bb9\" (UniqueName: \"kubernetes.io/projected/b7a3ac30-3925-493e-9e13-0f1a9a7e49c2-kube-api-access-99bb9\") pod \"dnsmasq-dns-675f4bcbfc-shhlw\" (UID: \"b7a3ac30-3925-493e-9e13-0f1a9a7e49c2\") " pod="openstack/dnsmasq-dns-675f4bcbfc-shhlw" Nov 25 10:46:56 crc kubenswrapper[4821]: I1125 10:46:56.697130 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7a3ac30-3925-493e-9e13-0f1a9a7e49c2-config\") pod \"dnsmasq-dns-675f4bcbfc-shhlw\" (UID: \"b7a3ac30-3925-493e-9e13-0f1a9a7e49c2\") " pod="openstack/dnsmasq-dns-675f4bcbfc-shhlw" Nov 25 10:46:56 crc kubenswrapper[4821]: I1125 10:46:56.697209 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68a5afc7-5e47-4bed-a6c2-8ff5b6b0236c-config\") pod \"dnsmasq-dns-78dd6ddcc-srdjw\" (UID: \"68a5afc7-5e47-4bed-a6c2-8ff5b6b0236c\") " pod="openstack/dnsmasq-dns-78dd6ddcc-srdjw" Nov 25 10:46:56 crc kubenswrapper[4821]: I1125 10:46:56.697965 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7a3ac30-3925-493e-9e13-0f1a9a7e49c2-config\") pod \"dnsmasq-dns-675f4bcbfc-shhlw\" (UID: \"b7a3ac30-3925-493e-9e13-0f1a9a7e49c2\") " pod="openstack/dnsmasq-dns-675f4bcbfc-shhlw" Nov 25 10:46:56 crc kubenswrapper[4821]: I1125 10:46:56.714565 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99bb9\" (UniqueName: \"kubernetes.io/projected/b7a3ac30-3925-493e-9e13-0f1a9a7e49c2-kube-api-access-99bb9\") pod \"dnsmasq-dns-675f4bcbfc-shhlw\" (UID: \"b7a3ac30-3925-493e-9e13-0f1a9a7e49c2\") " pod="openstack/dnsmasq-dns-675f4bcbfc-shhlw" Nov 25 10:46:56 crc kubenswrapper[4821]: I1125 10:46:56.798219 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68a5afc7-5e47-4bed-a6c2-8ff5b6b0236c-config\") pod \"dnsmasq-dns-78dd6ddcc-srdjw\" (UID: \"68a5afc7-5e47-4bed-a6c2-8ff5b6b0236c\") " pod="openstack/dnsmasq-dns-78dd6ddcc-srdjw" Nov 25 10:46:56 crc kubenswrapper[4821]: I1125 10:46:56.798815 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xqr4f\" (UniqueName: \"kubernetes.io/projected/68a5afc7-5e47-4bed-a6c2-8ff5b6b0236c-kube-api-access-xqr4f\") pod \"dnsmasq-dns-78dd6ddcc-srdjw\" (UID: \"68a5afc7-5e47-4bed-a6c2-8ff5b6b0236c\") " pod="openstack/dnsmasq-dns-78dd6ddcc-srdjw" Nov 25 10:46:56 crc kubenswrapper[4821]: I1125 10:46:56.798946 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/68a5afc7-5e47-4bed-a6c2-8ff5b6b0236c-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-srdjw\" (UID: \"68a5afc7-5e47-4bed-a6c2-8ff5b6b0236c\") " pod="openstack/dnsmasq-dns-78dd6ddcc-srdjw" Nov 25 10:46:56 crc kubenswrapper[4821]: I1125 10:46:56.799371 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68a5afc7-5e47-4bed-a6c2-8ff5b6b0236c-config\") pod \"dnsmasq-dns-78dd6ddcc-srdjw\" (UID: \"68a5afc7-5e47-4bed-a6c2-8ff5b6b0236c\") " pod="openstack/dnsmasq-dns-78dd6ddcc-srdjw" Nov 25 10:46:56 crc kubenswrapper[4821]: I1125 10:46:56.799815 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/68a5afc7-5e47-4bed-a6c2-8ff5b6b0236c-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-srdjw\" (UID: \"68a5afc7-5e47-4bed-a6c2-8ff5b6b0236c\") " pod="openstack/dnsmasq-dns-78dd6ddcc-srdjw" Nov 25 10:46:56 crc kubenswrapper[4821]: I1125 10:46:56.818923 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xqr4f\" (UniqueName: \"kubernetes.io/projected/68a5afc7-5e47-4bed-a6c2-8ff5b6b0236c-kube-api-access-xqr4f\") pod \"dnsmasq-dns-78dd6ddcc-srdjw\" (UID: \"68a5afc7-5e47-4bed-a6c2-8ff5b6b0236c\") " pod="openstack/dnsmasq-dns-78dd6ddcc-srdjw" Nov 25 10:46:56 crc kubenswrapper[4821]: I1125 10:46:56.836986 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-shhlw" Nov 25 10:46:56 crc kubenswrapper[4821]: I1125 10:46:56.899767 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-srdjw" Nov 25 10:46:57 crc kubenswrapper[4821]: I1125 10:46:57.286837 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-shhlw"] Nov 25 10:46:57 crc kubenswrapper[4821]: I1125 10:46:57.295012 4821 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 10:46:57 crc kubenswrapper[4821]: I1125 10:46:57.366244 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-srdjw"] Nov 25 10:46:57 crc kubenswrapper[4821]: W1125 10:46:57.369898 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod68a5afc7_5e47_4bed_a6c2_8ff5b6b0236c.slice/crio-2c793cc86cb38df469836329e3d4d5652b95d368e90ae7040096dcd72bc0fe99 WatchSource:0}: Error finding container 2c793cc86cb38df469836329e3d4d5652b95d368e90ae7040096dcd72bc0fe99: Status 404 returned error can't find the container with id 2c793cc86cb38df469836329e3d4d5652b95d368e90ae7040096dcd72bc0fe99 Nov 25 10:46:57 crc kubenswrapper[4821]: I1125 10:46:57.554368 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-shhlw" event={"ID":"b7a3ac30-3925-493e-9e13-0f1a9a7e49c2","Type":"ContainerStarted","Data":"23de78eedabd01795c0b200e44b5388d4d4d3a58f73bd2db657ab9e800ea9af6"} Nov 25 10:46:57 crc kubenswrapper[4821]: I1125 10:46:57.555621 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-srdjw" event={"ID":"68a5afc7-5e47-4bed-a6c2-8ff5b6b0236c","Type":"ContainerStarted","Data":"2c793cc86cb38df469836329e3d4d5652b95d368e90ae7040096dcd72bc0fe99"} Nov 25 10:46:59 crc kubenswrapper[4821]: I1125 10:46:59.718335 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-shhlw"] Nov 25 10:46:59 crc kubenswrapper[4821]: I1125 10:46:59.743046 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-gn85r"] Nov 25 10:46:59 crc kubenswrapper[4821]: I1125 10:46:59.744333 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-gn85r" Nov 25 10:46:59 crc kubenswrapper[4821]: I1125 10:46:59.752012 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-gn85r"] Nov 25 10:46:59 crc kubenswrapper[4821]: I1125 10:46:59.841148 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wgq8x\" (UniqueName: \"kubernetes.io/projected/47b41fd0-3e3d-4cf7-b8c9-1441ae9e2014-kube-api-access-wgq8x\") pod \"dnsmasq-dns-5ccc8479f9-gn85r\" (UID: \"47b41fd0-3e3d-4cf7-b8c9-1441ae9e2014\") " pod="openstack/dnsmasq-dns-5ccc8479f9-gn85r" Nov 25 10:46:59 crc kubenswrapper[4821]: I1125 10:46:59.841270 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/47b41fd0-3e3d-4cf7-b8c9-1441ae9e2014-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-gn85r\" (UID: \"47b41fd0-3e3d-4cf7-b8c9-1441ae9e2014\") " pod="openstack/dnsmasq-dns-5ccc8479f9-gn85r" Nov 25 10:46:59 crc kubenswrapper[4821]: I1125 10:46:59.841338 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47b41fd0-3e3d-4cf7-b8c9-1441ae9e2014-config\") pod \"dnsmasq-dns-5ccc8479f9-gn85r\" (UID: \"47b41fd0-3e3d-4cf7-b8c9-1441ae9e2014\") " pod="openstack/dnsmasq-dns-5ccc8479f9-gn85r" Nov 25 10:46:59 crc kubenswrapper[4821]: I1125 10:46:59.943856 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47b41fd0-3e3d-4cf7-b8c9-1441ae9e2014-config\") pod \"dnsmasq-dns-5ccc8479f9-gn85r\" (UID: \"47b41fd0-3e3d-4cf7-b8c9-1441ae9e2014\") " pod="openstack/dnsmasq-dns-5ccc8479f9-gn85r" Nov 25 10:46:59 crc kubenswrapper[4821]: I1125 10:46:59.943910 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wgq8x\" (UniqueName: \"kubernetes.io/projected/47b41fd0-3e3d-4cf7-b8c9-1441ae9e2014-kube-api-access-wgq8x\") pod \"dnsmasq-dns-5ccc8479f9-gn85r\" (UID: \"47b41fd0-3e3d-4cf7-b8c9-1441ae9e2014\") " pod="openstack/dnsmasq-dns-5ccc8479f9-gn85r" Nov 25 10:46:59 crc kubenswrapper[4821]: I1125 10:46:59.943961 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/47b41fd0-3e3d-4cf7-b8c9-1441ae9e2014-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-gn85r\" (UID: \"47b41fd0-3e3d-4cf7-b8c9-1441ae9e2014\") " pod="openstack/dnsmasq-dns-5ccc8479f9-gn85r" Nov 25 10:46:59 crc kubenswrapper[4821]: I1125 10:46:59.944854 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/47b41fd0-3e3d-4cf7-b8c9-1441ae9e2014-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-gn85r\" (UID: \"47b41fd0-3e3d-4cf7-b8c9-1441ae9e2014\") " pod="openstack/dnsmasq-dns-5ccc8479f9-gn85r" Nov 25 10:46:59 crc kubenswrapper[4821]: I1125 10:46:59.945469 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47b41fd0-3e3d-4cf7-b8c9-1441ae9e2014-config\") pod \"dnsmasq-dns-5ccc8479f9-gn85r\" (UID: \"47b41fd0-3e3d-4cf7-b8c9-1441ae9e2014\") " pod="openstack/dnsmasq-dns-5ccc8479f9-gn85r" Nov 25 10:46:59 crc kubenswrapper[4821]: I1125 10:46:59.971862 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wgq8x\" (UniqueName: \"kubernetes.io/projected/47b41fd0-3e3d-4cf7-b8c9-1441ae9e2014-kube-api-access-wgq8x\") pod \"dnsmasq-dns-5ccc8479f9-gn85r\" (UID: \"47b41fd0-3e3d-4cf7-b8c9-1441ae9e2014\") " pod="openstack/dnsmasq-dns-5ccc8479f9-gn85r" Nov 25 10:46:59 crc kubenswrapper[4821]: I1125 10:46:59.975524 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-srdjw"] Nov 25 10:47:00 crc kubenswrapper[4821]: I1125 10:47:00.019322 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-h5qkz"] Nov 25 10:47:00 crc kubenswrapper[4821]: I1125 10:47:00.020473 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-h5qkz" Nov 25 10:47:00 crc kubenswrapper[4821]: I1125 10:47:00.041748 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-h5qkz"] Nov 25 10:47:00 crc kubenswrapper[4821]: I1125 10:47:00.062382 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-gn85r" Nov 25 10:47:00 crc kubenswrapper[4821]: I1125 10:47:00.153523 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f289352e-7670-4503-b6bf-ccdaa45a7c36-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-h5qkz\" (UID: \"f289352e-7670-4503-b6bf-ccdaa45a7c36\") " pod="openstack/dnsmasq-dns-57d769cc4f-h5qkz" Nov 25 10:47:00 crc kubenswrapper[4821]: I1125 10:47:00.153668 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f289352e-7670-4503-b6bf-ccdaa45a7c36-config\") pod \"dnsmasq-dns-57d769cc4f-h5qkz\" (UID: \"f289352e-7670-4503-b6bf-ccdaa45a7c36\") " pod="openstack/dnsmasq-dns-57d769cc4f-h5qkz" Nov 25 10:47:00 crc kubenswrapper[4821]: I1125 10:47:00.153721 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgs6z\" (UniqueName: \"kubernetes.io/projected/f289352e-7670-4503-b6bf-ccdaa45a7c36-kube-api-access-qgs6z\") pod \"dnsmasq-dns-57d769cc4f-h5qkz\" (UID: \"f289352e-7670-4503-b6bf-ccdaa45a7c36\") " pod="openstack/dnsmasq-dns-57d769cc4f-h5qkz" Nov 25 10:47:00 crc kubenswrapper[4821]: I1125 10:47:00.256132 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f289352e-7670-4503-b6bf-ccdaa45a7c36-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-h5qkz\" (UID: \"f289352e-7670-4503-b6bf-ccdaa45a7c36\") " pod="openstack/dnsmasq-dns-57d769cc4f-h5qkz" Nov 25 10:47:00 crc kubenswrapper[4821]: I1125 10:47:00.256231 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f289352e-7670-4503-b6bf-ccdaa45a7c36-config\") pod \"dnsmasq-dns-57d769cc4f-h5qkz\" (UID: \"f289352e-7670-4503-b6bf-ccdaa45a7c36\") " pod="openstack/dnsmasq-dns-57d769cc4f-h5qkz" Nov 25 10:47:00 crc kubenswrapper[4821]: I1125 10:47:00.256273 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgs6z\" (UniqueName: \"kubernetes.io/projected/f289352e-7670-4503-b6bf-ccdaa45a7c36-kube-api-access-qgs6z\") pod \"dnsmasq-dns-57d769cc4f-h5qkz\" (UID: \"f289352e-7670-4503-b6bf-ccdaa45a7c36\") " pod="openstack/dnsmasq-dns-57d769cc4f-h5qkz" Nov 25 10:47:00 crc kubenswrapper[4821]: I1125 10:47:00.257919 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f289352e-7670-4503-b6bf-ccdaa45a7c36-config\") pod \"dnsmasq-dns-57d769cc4f-h5qkz\" (UID: \"f289352e-7670-4503-b6bf-ccdaa45a7c36\") " pod="openstack/dnsmasq-dns-57d769cc4f-h5qkz" Nov 25 10:47:00 crc kubenswrapper[4821]: I1125 10:47:00.258983 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f289352e-7670-4503-b6bf-ccdaa45a7c36-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-h5qkz\" (UID: \"f289352e-7670-4503-b6bf-ccdaa45a7c36\") " pod="openstack/dnsmasq-dns-57d769cc4f-h5qkz" Nov 25 10:47:00 crc kubenswrapper[4821]: I1125 10:47:00.278668 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgs6z\" (UniqueName: \"kubernetes.io/projected/f289352e-7670-4503-b6bf-ccdaa45a7c36-kube-api-access-qgs6z\") pod \"dnsmasq-dns-57d769cc4f-h5qkz\" (UID: \"f289352e-7670-4503-b6bf-ccdaa45a7c36\") " pod="openstack/dnsmasq-dns-57d769cc4f-h5qkz" Nov 25 10:47:00 crc kubenswrapper[4821]: I1125 10:47:00.338875 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-h5qkz" Nov 25 10:47:00 crc kubenswrapper[4821]: I1125 10:47:00.471658 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-gn85r"] Nov 25 10:47:00 crc kubenswrapper[4821]: I1125 10:47:00.483498 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8phxk" Nov 25 10:47:00 crc kubenswrapper[4821]: I1125 10:47:00.483581 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8phxk" Nov 25 10:47:00 crc kubenswrapper[4821]: I1125 10:47:00.533684 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8phxk" Nov 25 10:47:00 crc kubenswrapper[4821]: I1125 10:47:00.632666 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8phxk" Nov 25 10:47:00 crc kubenswrapper[4821]: I1125 10:47:00.764144 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8phxk"] Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.061329 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.062487 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.065301 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.065553 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.065751 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-vfvpv" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.066143 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.066289 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.067317 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.067622 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.074395 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.146363 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.151653 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.156607 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.156747 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.156772 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.156871 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.157017 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.157106 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-fwctx" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.157327 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.161193 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.169292 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/330a479e-3d76-4de3-bc84-1d7c2d65d265-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"330a479e-3d76-4de3-bc84-1d7c2d65d265\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.169336 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/330a479e-3d76-4de3-bc84-1d7c2d65d265-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"330a479e-3d76-4de3-bc84-1d7c2d65d265\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.169374 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/330a479e-3d76-4de3-bc84-1d7c2d65d265-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"330a479e-3d76-4de3-bc84-1d7c2d65d265\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.169414 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"330a479e-3d76-4de3-bc84-1d7c2d65d265\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.169439 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/330a479e-3d76-4de3-bc84-1d7c2d65d265-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"330a479e-3d76-4de3-bc84-1d7c2d65d265\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.169472 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/330a479e-3d76-4de3-bc84-1d7c2d65d265-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"330a479e-3d76-4de3-bc84-1d7c2d65d265\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.169488 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/330a479e-3d76-4de3-bc84-1d7c2d65d265-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"330a479e-3d76-4de3-bc84-1d7c2d65d265\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.169505 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/330a479e-3d76-4de3-bc84-1d7c2d65d265-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"330a479e-3d76-4de3-bc84-1d7c2d65d265\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.169531 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4tp7\" (UniqueName: \"kubernetes.io/projected/330a479e-3d76-4de3-bc84-1d7c2d65d265-kube-api-access-t4tp7\") pod \"rabbitmq-cell1-server-0\" (UID: \"330a479e-3d76-4de3-bc84-1d7c2d65d265\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.169546 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/330a479e-3d76-4de3-bc84-1d7c2d65d265-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"330a479e-3d76-4de3-bc84-1d7c2d65d265\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.169573 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/330a479e-3d76-4de3-bc84-1d7c2d65d265-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"330a479e-3d76-4de3-bc84-1d7c2d65d265\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.270980 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"330a479e-3d76-4de3-bc84-1d7c2d65d265\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.271031 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/330a479e-3d76-4de3-bc84-1d7c2d65d265-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"330a479e-3d76-4de3-bc84-1d7c2d65d265\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.271098 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2bbb754a-fff3-4ba2-ab75-68e48b626ab0-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"2bbb754a-fff3-4ba2-ab75-68e48b626ab0\") " pod="openstack/rabbitmq-server-0" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.271132 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2bbb754a-fff3-4ba2-ab75-68e48b626ab0-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"2bbb754a-fff3-4ba2-ab75-68e48b626ab0\") " pod="openstack/rabbitmq-server-0" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.271150 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/330a479e-3d76-4de3-bc84-1d7c2d65d265-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"330a479e-3d76-4de3-bc84-1d7c2d65d265\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.271186 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/330a479e-3d76-4de3-bc84-1d7c2d65d265-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"330a479e-3d76-4de3-bc84-1d7c2d65d265\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.271202 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2bbb754a-fff3-4ba2-ab75-68e48b626ab0-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"2bbb754a-fff3-4ba2-ab75-68e48b626ab0\") " pod="openstack/rabbitmq-server-0" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.271223 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/330a479e-3d76-4de3-bc84-1d7c2d65d265-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"330a479e-3d76-4de3-bc84-1d7c2d65d265\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.271274 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4tp7\" (UniqueName: \"kubernetes.io/projected/330a479e-3d76-4de3-bc84-1d7c2d65d265-kube-api-access-t4tp7\") pod \"rabbitmq-cell1-server-0\" (UID: \"330a479e-3d76-4de3-bc84-1d7c2d65d265\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.271291 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/330a479e-3d76-4de3-bc84-1d7c2d65d265-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"330a479e-3d76-4de3-bc84-1d7c2d65d265\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.271312 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2bbb754a-fff3-4ba2-ab75-68e48b626ab0-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"2bbb754a-fff3-4ba2-ab75-68e48b626ab0\") " pod="openstack/rabbitmq-server-0" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.271337 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/330a479e-3d76-4de3-bc84-1d7c2d65d265-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"330a479e-3d76-4de3-bc84-1d7c2d65d265\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.271361 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gpq9f\" (UniqueName: \"kubernetes.io/projected/2bbb754a-fff3-4ba2-ab75-68e48b626ab0-kube-api-access-gpq9f\") pod \"rabbitmq-server-0\" (UID: \"2bbb754a-fff3-4ba2-ab75-68e48b626ab0\") " pod="openstack/rabbitmq-server-0" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.271384 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2bbb754a-fff3-4ba2-ab75-68e48b626ab0-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"2bbb754a-fff3-4ba2-ab75-68e48b626ab0\") " pod="openstack/rabbitmq-server-0" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.271403 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2bbb754a-fff3-4ba2-ab75-68e48b626ab0-config-data\") pod \"rabbitmq-server-0\" (UID: \"2bbb754a-fff3-4ba2-ab75-68e48b626ab0\") " pod="openstack/rabbitmq-server-0" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.271425 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/330a479e-3d76-4de3-bc84-1d7c2d65d265-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"330a479e-3d76-4de3-bc84-1d7c2d65d265\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.271445 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2bbb754a-fff3-4ba2-ab75-68e48b626ab0-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"2bbb754a-fff3-4ba2-ab75-68e48b626ab0\") " pod="openstack/rabbitmq-server-0" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.271513 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/330a479e-3d76-4de3-bc84-1d7c2d65d265-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"330a479e-3d76-4de3-bc84-1d7c2d65d265\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.271593 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2bbb754a-fff3-4ba2-ab75-68e48b626ab0-pod-info\") pod \"rabbitmq-server-0\" (UID: \"2bbb754a-fff3-4ba2-ab75-68e48b626ab0\") " pod="openstack/rabbitmq-server-0" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.271628 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/330a479e-3d76-4de3-bc84-1d7c2d65d265-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"330a479e-3d76-4de3-bc84-1d7c2d65d265\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.271965 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/330a479e-3d76-4de3-bc84-1d7c2d65d265-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"330a479e-3d76-4de3-bc84-1d7c2d65d265\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.272011 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/330a479e-3d76-4de3-bc84-1d7c2d65d265-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"330a479e-3d76-4de3-bc84-1d7c2d65d265\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.272216 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/330a479e-3d76-4de3-bc84-1d7c2d65d265-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"330a479e-3d76-4de3-bc84-1d7c2d65d265\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.272440 4821 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"330a479e-3d76-4de3-bc84-1d7c2d65d265\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.272458 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"2bbb754a-fff3-4ba2-ab75-68e48b626ab0\") " pod="openstack/rabbitmq-server-0" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.272607 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2bbb754a-fff3-4ba2-ab75-68e48b626ab0-server-conf\") pod \"rabbitmq-server-0\" (UID: \"2bbb754a-fff3-4ba2-ab75-68e48b626ab0\") " pod="openstack/rabbitmq-server-0" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.273515 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/330a479e-3d76-4de3-bc84-1d7c2d65d265-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"330a479e-3d76-4de3-bc84-1d7c2d65d265\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.274274 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/330a479e-3d76-4de3-bc84-1d7c2d65d265-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"330a479e-3d76-4de3-bc84-1d7c2d65d265\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.277678 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/330a479e-3d76-4de3-bc84-1d7c2d65d265-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"330a479e-3d76-4de3-bc84-1d7c2d65d265\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.277703 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/330a479e-3d76-4de3-bc84-1d7c2d65d265-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"330a479e-3d76-4de3-bc84-1d7c2d65d265\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.278251 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/330a479e-3d76-4de3-bc84-1d7c2d65d265-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"330a479e-3d76-4de3-bc84-1d7c2d65d265\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.287217 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/330a479e-3d76-4de3-bc84-1d7c2d65d265-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"330a479e-3d76-4de3-bc84-1d7c2d65d265\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.289263 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4tp7\" (UniqueName: \"kubernetes.io/projected/330a479e-3d76-4de3-bc84-1d7c2d65d265-kube-api-access-t4tp7\") pod \"rabbitmq-cell1-server-0\" (UID: \"330a479e-3d76-4de3-bc84-1d7c2d65d265\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.294893 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"330a479e-3d76-4de3-bc84-1d7c2d65d265\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.374324 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2bbb754a-fff3-4ba2-ab75-68e48b626ab0-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"2bbb754a-fff3-4ba2-ab75-68e48b626ab0\") " pod="openstack/rabbitmq-server-0" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.374382 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gpq9f\" (UniqueName: \"kubernetes.io/projected/2bbb754a-fff3-4ba2-ab75-68e48b626ab0-kube-api-access-gpq9f\") pod \"rabbitmq-server-0\" (UID: \"2bbb754a-fff3-4ba2-ab75-68e48b626ab0\") " pod="openstack/rabbitmq-server-0" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.374414 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2bbb754a-fff3-4ba2-ab75-68e48b626ab0-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"2bbb754a-fff3-4ba2-ab75-68e48b626ab0\") " pod="openstack/rabbitmq-server-0" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.374434 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2bbb754a-fff3-4ba2-ab75-68e48b626ab0-config-data\") pod \"rabbitmq-server-0\" (UID: \"2bbb754a-fff3-4ba2-ab75-68e48b626ab0\") " pod="openstack/rabbitmq-server-0" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.374459 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2bbb754a-fff3-4ba2-ab75-68e48b626ab0-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"2bbb754a-fff3-4ba2-ab75-68e48b626ab0\") " pod="openstack/rabbitmq-server-0" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.374481 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2bbb754a-fff3-4ba2-ab75-68e48b626ab0-pod-info\") pod \"rabbitmq-server-0\" (UID: \"2bbb754a-fff3-4ba2-ab75-68e48b626ab0\") " pod="openstack/rabbitmq-server-0" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.374504 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"2bbb754a-fff3-4ba2-ab75-68e48b626ab0\") " pod="openstack/rabbitmq-server-0" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.374527 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2bbb754a-fff3-4ba2-ab75-68e48b626ab0-server-conf\") pod \"rabbitmq-server-0\" (UID: \"2bbb754a-fff3-4ba2-ab75-68e48b626ab0\") " pod="openstack/rabbitmq-server-0" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.374550 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2bbb754a-fff3-4ba2-ab75-68e48b626ab0-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"2bbb754a-fff3-4ba2-ab75-68e48b626ab0\") " pod="openstack/rabbitmq-server-0" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.374573 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2bbb754a-fff3-4ba2-ab75-68e48b626ab0-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"2bbb754a-fff3-4ba2-ab75-68e48b626ab0\") " pod="openstack/rabbitmq-server-0" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.374591 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2bbb754a-fff3-4ba2-ab75-68e48b626ab0-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"2bbb754a-fff3-4ba2-ab75-68e48b626ab0\") " pod="openstack/rabbitmq-server-0" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.375076 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2bbb754a-fff3-4ba2-ab75-68e48b626ab0-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"2bbb754a-fff3-4ba2-ab75-68e48b626ab0\") " pod="openstack/rabbitmq-server-0" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.375345 4821 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"2bbb754a-fff3-4ba2-ab75-68e48b626ab0\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/rabbitmq-server-0" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.376536 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2bbb754a-fff3-4ba2-ab75-68e48b626ab0-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"2bbb754a-fff3-4ba2-ab75-68e48b626ab0\") " pod="openstack/rabbitmq-server-0" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.377529 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2bbb754a-fff3-4ba2-ab75-68e48b626ab0-server-conf\") pod \"rabbitmq-server-0\" (UID: \"2bbb754a-fff3-4ba2-ab75-68e48b626ab0\") " pod="openstack/rabbitmq-server-0" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.377812 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2bbb754a-fff3-4ba2-ab75-68e48b626ab0-config-data\") pod \"rabbitmq-server-0\" (UID: \"2bbb754a-fff3-4ba2-ab75-68e48b626ab0\") " pod="openstack/rabbitmq-server-0" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.378297 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2bbb754a-fff3-4ba2-ab75-68e48b626ab0-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"2bbb754a-fff3-4ba2-ab75-68e48b626ab0\") " pod="openstack/rabbitmq-server-0" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.378638 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2bbb754a-fff3-4ba2-ab75-68e48b626ab0-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"2bbb754a-fff3-4ba2-ab75-68e48b626ab0\") " pod="openstack/rabbitmq-server-0" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.379339 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2bbb754a-fff3-4ba2-ab75-68e48b626ab0-pod-info\") pod \"rabbitmq-server-0\" (UID: \"2bbb754a-fff3-4ba2-ab75-68e48b626ab0\") " pod="openstack/rabbitmq-server-0" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.380618 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2bbb754a-fff3-4ba2-ab75-68e48b626ab0-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"2bbb754a-fff3-4ba2-ab75-68e48b626ab0\") " pod="openstack/rabbitmq-server-0" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.380702 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2bbb754a-fff3-4ba2-ab75-68e48b626ab0-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"2bbb754a-fff3-4ba2-ab75-68e48b626ab0\") " pod="openstack/rabbitmq-server-0" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.389606 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.395252 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gpq9f\" (UniqueName: \"kubernetes.io/projected/2bbb754a-fff3-4ba2-ab75-68e48b626ab0-kube-api-access-gpq9f\") pod \"rabbitmq-server-0\" (UID: \"2bbb754a-fff3-4ba2-ab75-68e48b626ab0\") " pod="openstack/rabbitmq-server-0" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.397621 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"2bbb754a-fff3-4ba2-ab75-68e48b626ab0\") " pod="openstack/rabbitmq-server-0" Nov 25 10:47:01 crc kubenswrapper[4821]: I1125 10:47:01.480910 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 25 10:47:02 crc kubenswrapper[4821]: I1125 10:47:02.597218 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-8phxk" podUID="e93e112b-d64e-4bab-b62d-cfee3e1a74eb" containerName="registry-server" containerID="cri-o://afb0d51c47960d814a1cea6065cbd96a29f2014cf8af36ff98b91e9fa73c658e" gracePeriod=2 Nov 25 10:47:02 crc kubenswrapper[4821]: I1125 10:47:02.713059 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Nov 25 10:47:02 crc kubenswrapper[4821]: I1125 10:47:02.714475 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 25 10:47:02 crc kubenswrapper[4821]: I1125 10:47:02.719181 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-hqz9w" Nov 25 10:47:02 crc kubenswrapper[4821]: I1125 10:47:02.720399 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Nov 25 10:47:02 crc kubenswrapper[4821]: I1125 10:47:02.720607 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Nov 25 10:47:02 crc kubenswrapper[4821]: I1125 10:47:02.720769 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Nov 25 10:47:02 crc kubenswrapper[4821]: I1125 10:47:02.727054 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Nov 25 10:47:02 crc kubenswrapper[4821]: I1125 10:47:02.727318 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 25 10:47:02 crc kubenswrapper[4821]: I1125 10:47:02.819573 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecb5eaf2-4178-43ec-bc21-0e05de8fc237-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"ecb5eaf2-4178-43ec-bc21-0e05de8fc237\") " pod="openstack/openstack-galera-0" Nov 25 10:47:02 crc kubenswrapper[4821]: I1125 10:47:02.819623 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-galera-0\" (UID: \"ecb5eaf2-4178-43ec-bc21-0e05de8fc237\") " pod="openstack/openstack-galera-0" Nov 25 10:47:02 crc kubenswrapper[4821]: I1125 10:47:02.819682 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ecb5eaf2-4178-43ec-bc21-0e05de8fc237-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"ecb5eaf2-4178-43ec-bc21-0e05de8fc237\") " pod="openstack/openstack-galera-0" Nov 25 10:47:02 crc kubenswrapper[4821]: I1125 10:47:02.819769 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b669f\" (UniqueName: \"kubernetes.io/projected/ecb5eaf2-4178-43ec-bc21-0e05de8fc237-kube-api-access-b669f\") pod \"openstack-galera-0\" (UID: \"ecb5eaf2-4178-43ec-bc21-0e05de8fc237\") " pod="openstack/openstack-galera-0" Nov 25 10:47:02 crc kubenswrapper[4821]: I1125 10:47:02.819798 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ecb5eaf2-4178-43ec-bc21-0e05de8fc237-kolla-config\") pod \"openstack-galera-0\" (UID: \"ecb5eaf2-4178-43ec-bc21-0e05de8fc237\") " pod="openstack/openstack-galera-0" Nov 25 10:47:02 crc kubenswrapper[4821]: I1125 10:47:02.819874 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ecb5eaf2-4178-43ec-bc21-0e05de8fc237-config-data-default\") pod \"openstack-galera-0\" (UID: \"ecb5eaf2-4178-43ec-bc21-0e05de8fc237\") " pod="openstack/openstack-galera-0" Nov 25 10:47:02 crc kubenswrapper[4821]: I1125 10:47:02.819937 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ecb5eaf2-4178-43ec-bc21-0e05de8fc237-operator-scripts\") pod \"openstack-galera-0\" (UID: \"ecb5eaf2-4178-43ec-bc21-0e05de8fc237\") " pod="openstack/openstack-galera-0" Nov 25 10:47:02 crc kubenswrapper[4821]: I1125 10:47:02.820011 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ecb5eaf2-4178-43ec-bc21-0e05de8fc237-config-data-generated\") pod \"openstack-galera-0\" (UID: \"ecb5eaf2-4178-43ec-bc21-0e05de8fc237\") " pod="openstack/openstack-galera-0" Nov 25 10:47:02 crc kubenswrapper[4821]: I1125 10:47:02.921383 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecb5eaf2-4178-43ec-bc21-0e05de8fc237-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"ecb5eaf2-4178-43ec-bc21-0e05de8fc237\") " pod="openstack/openstack-galera-0" Nov 25 10:47:02 crc kubenswrapper[4821]: I1125 10:47:02.921706 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-galera-0\" (UID: \"ecb5eaf2-4178-43ec-bc21-0e05de8fc237\") " pod="openstack/openstack-galera-0" Nov 25 10:47:02 crc kubenswrapper[4821]: I1125 10:47:02.921737 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ecb5eaf2-4178-43ec-bc21-0e05de8fc237-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"ecb5eaf2-4178-43ec-bc21-0e05de8fc237\") " pod="openstack/openstack-galera-0" Nov 25 10:47:02 crc kubenswrapper[4821]: I1125 10:47:02.921957 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ecb5eaf2-4178-43ec-bc21-0e05de8fc237-kolla-config\") pod \"openstack-galera-0\" (UID: \"ecb5eaf2-4178-43ec-bc21-0e05de8fc237\") " pod="openstack/openstack-galera-0" Nov 25 10:47:02 crc kubenswrapper[4821]: I1125 10:47:02.921976 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b669f\" (UniqueName: \"kubernetes.io/projected/ecb5eaf2-4178-43ec-bc21-0e05de8fc237-kube-api-access-b669f\") pod \"openstack-galera-0\" (UID: \"ecb5eaf2-4178-43ec-bc21-0e05de8fc237\") " pod="openstack/openstack-galera-0" Nov 25 10:47:02 crc kubenswrapper[4821]: I1125 10:47:02.922000 4821 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-galera-0\" (UID: \"ecb5eaf2-4178-43ec-bc21-0e05de8fc237\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/openstack-galera-0" Nov 25 10:47:02 crc kubenswrapper[4821]: I1125 10:47:02.921994 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ecb5eaf2-4178-43ec-bc21-0e05de8fc237-config-data-default\") pod \"openstack-galera-0\" (UID: \"ecb5eaf2-4178-43ec-bc21-0e05de8fc237\") " pod="openstack/openstack-galera-0" Nov 25 10:47:02 crc kubenswrapper[4821]: I1125 10:47:02.922372 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ecb5eaf2-4178-43ec-bc21-0e05de8fc237-operator-scripts\") pod \"openstack-galera-0\" (UID: \"ecb5eaf2-4178-43ec-bc21-0e05de8fc237\") " pod="openstack/openstack-galera-0" Nov 25 10:47:02 crc kubenswrapper[4821]: I1125 10:47:02.922404 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ecb5eaf2-4178-43ec-bc21-0e05de8fc237-config-data-generated\") pod \"openstack-galera-0\" (UID: \"ecb5eaf2-4178-43ec-bc21-0e05de8fc237\") " pod="openstack/openstack-galera-0" Nov 25 10:47:02 crc kubenswrapper[4821]: I1125 10:47:02.924235 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ecb5eaf2-4178-43ec-bc21-0e05de8fc237-config-data-generated\") pod \"openstack-galera-0\" (UID: \"ecb5eaf2-4178-43ec-bc21-0e05de8fc237\") " pod="openstack/openstack-galera-0" Nov 25 10:47:02 crc kubenswrapper[4821]: I1125 10:47:02.924580 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ecb5eaf2-4178-43ec-bc21-0e05de8fc237-config-data-default\") pod \"openstack-galera-0\" (UID: \"ecb5eaf2-4178-43ec-bc21-0e05de8fc237\") " pod="openstack/openstack-galera-0" Nov 25 10:47:02 crc kubenswrapper[4821]: I1125 10:47:02.925208 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ecb5eaf2-4178-43ec-bc21-0e05de8fc237-operator-scripts\") pod \"openstack-galera-0\" (UID: \"ecb5eaf2-4178-43ec-bc21-0e05de8fc237\") " pod="openstack/openstack-galera-0" Nov 25 10:47:02 crc kubenswrapper[4821]: I1125 10:47:02.925545 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ecb5eaf2-4178-43ec-bc21-0e05de8fc237-kolla-config\") pod \"openstack-galera-0\" (UID: \"ecb5eaf2-4178-43ec-bc21-0e05de8fc237\") " pod="openstack/openstack-galera-0" Nov 25 10:47:02 crc kubenswrapper[4821]: I1125 10:47:02.932987 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ecb5eaf2-4178-43ec-bc21-0e05de8fc237-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"ecb5eaf2-4178-43ec-bc21-0e05de8fc237\") " pod="openstack/openstack-galera-0" Nov 25 10:47:02 crc kubenswrapper[4821]: I1125 10:47:02.943415 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecb5eaf2-4178-43ec-bc21-0e05de8fc237-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"ecb5eaf2-4178-43ec-bc21-0e05de8fc237\") " pod="openstack/openstack-galera-0" Nov 25 10:47:02 crc kubenswrapper[4821]: I1125 10:47:02.944383 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b669f\" (UniqueName: \"kubernetes.io/projected/ecb5eaf2-4178-43ec-bc21-0e05de8fc237-kube-api-access-b669f\") pod \"openstack-galera-0\" (UID: \"ecb5eaf2-4178-43ec-bc21-0e05de8fc237\") " pod="openstack/openstack-galera-0" Nov 25 10:47:02 crc kubenswrapper[4821]: I1125 10:47:02.946251 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-galera-0\" (UID: \"ecb5eaf2-4178-43ec-bc21-0e05de8fc237\") " pod="openstack/openstack-galera-0" Nov 25 10:47:03 crc kubenswrapper[4821]: I1125 10:47:03.112498 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 25 10:47:03 crc kubenswrapper[4821]: I1125 10:47:03.616913 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-gn85r" event={"ID":"47b41fd0-3e3d-4cf7-b8c9-1441ae9e2014","Type":"ContainerStarted","Data":"4f21960c9d23afc3242da1057ea86fa8b849eba22c0b36e697e17207a6317d81"} Nov 25 10:47:03 crc kubenswrapper[4821]: I1125 10:47:03.640759 4821 generic.go:334] "Generic (PLEG): container finished" podID="e93e112b-d64e-4bab-b62d-cfee3e1a74eb" containerID="afb0d51c47960d814a1cea6065cbd96a29f2014cf8af36ff98b91e9fa73c658e" exitCode=0 Nov 25 10:47:03 crc kubenswrapper[4821]: I1125 10:47:03.640821 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8phxk" event={"ID":"e93e112b-d64e-4bab-b62d-cfee3e1a74eb","Type":"ContainerDied","Data":"afb0d51c47960d814a1cea6065cbd96a29f2014cf8af36ff98b91e9fa73c658e"} Nov 25 10:47:04 crc kubenswrapper[4821]: I1125 10:47:04.078622 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-h5qkz"] Nov 25 10:47:04 crc kubenswrapper[4821]: I1125 10:47:04.161952 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 25 10:47:04 crc kubenswrapper[4821]: I1125 10:47:04.163305 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 25 10:47:04 crc kubenswrapper[4821]: I1125 10:47:04.168265 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Nov 25 10:47:04 crc kubenswrapper[4821]: I1125 10:47:04.168396 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Nov 25 10:47:04 crc kubenswrapper[4821]: I1125 10:47:04.168423 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Nov 25 10:47:04 crc kubenswrapper[4821]: I1125 10:47:04.168544 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-dq2fm" Nov 25 10:47:04 crc kubenswrapper[4821]: I1125 10:47:04.172771 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 25 10:47:04 crc kubenswrapper[4821]: I1125 10:47:04.352439 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Nov 25 10:47:04 crc kubenswrapper[4821]: I1125 10:47:04.353732 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 25 10:47:04 crc kubenswrapper[4821]: I1125 10:47:04.356482 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/41c8c049-b466-4c19-944e-50a639491190-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"41c8c049-b466-4c19-944e-50a639491190\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:47:04 crc kubenswrapper[4821]: I1125 10:47:04.356672 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/41c8c049-b466-4c19-944e-50a639491190-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"41c8c049-b466-4c19-944e-50a639491190\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:47:04 crc kubenswrapper[4821]: I1125 10:47:04.356791 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/41c8c049-b466-4c19-944e-50a639491190-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"41c8c049-b466-4c19-944e-50a639491190\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:47:04 crc kubenswrapper[4821]: I1125 10:47:04.356856 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w8gpb\" (UniqueName: \"kubernetes.io/projected/41c8c049-b466-4c19-944e-50a639491190-kube-api-access-w8gpb\") pod \"openstack-cell1-galera-0\" (UID: \"41c8c049-b466-4c19-944e-50a639491190\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:47:04 crc kubenswrapper[4821]: I1125 10:47:04.356877 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/41c8c049-b466-4c19-944e-50a639491190-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"41c8c049-b466-4c19-944e-50a639491190\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:47:04 crc kubenswrapper[4821]: I1125 10:47:04.356896 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/41c8c049-b466-4c19-944e-50a639491190-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"41c8c049-b466-4c19-944e-50a639491190\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:47:04 crc kubenswrapper[4821]: I1125 10:47:04.357041 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41c8c049-b466-4c19-944e-50a639491190-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"41c8c049-b466-4c19-944e-50a639491190\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:47:04 crc kubenswrapper[4821]: I1125 10:47:04.357092 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-cell1-galera-0\" (UID: \"41c8c049-b466-4c19-944e-50a639491190\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:47:04 crc kubenswrapper[4821]: I1125 10:47:04.357256 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Nov 25 10:47:04 crc kubenswrapper[4821]: I1125 10:47:04.357360 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Nov 25 10:47:04 crc kubenswrapper[4821]: I1125 10:47:04.357463 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-9p7xt" Nov 25 10:47:04 crc kubenswrapper[4821]: I1125 10:47:04.367756 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 25 10:47:04 crc kubenswrapper[4821]: I1125 10:47:04.458152 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w8gpb\" (UniqueName: \"kubernetes.io/projected/41c8c049-b466-4c19-944e-50a639491190-kube-api-access-w8gpb\") pod \"openstack-cell1-galera-0\" (UID: \"41c8c049-b466-4c19-944e-50a639491190\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:47:04 crc kubenswrapper[4821]: I1125 10:47:04.458221 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/41c8c049-b466-4c19-944e-50a639491190-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"41c8c049-b466-4c19-944e-50a639491190\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:47:04 crc kubenswrapper[4821]: I1125 10:47:04.458247 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/41c8c049-b466-4c19-944e-50a639491190-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"41c8c049-b466-4c19-944e-50a639491190\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:47:04 crc kubenswrapper[4821]: I1125 10:47:04.458270 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8b9e8e7c-b055-47d3-874a-227d5f02432a-config-data\") pod \"memcached-0\" (UID: \"8b9e8e7c-b055-47d3-874a-227d5f02432a\") " pod="openstack/memcached-0" Nov 25 10:47:04 crc kubenswrapper[4821]: I1125 10:47:04.458289 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b9e8e7c-b055-47d3-874a-227d5f02432a-memcached-tls-certs\") pod \"memcached-0\" (UID: \"8b9e8e7c-b055-47d3-874a-227d5f02432a\") " pod="openstack/memcached-0" Nov 25 10:47:04 crc kubenswrapper[4821]: I1125 10:47:04.458305 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ns8r\" (UniqueName: \"kubernetes.io/projected/8b9e8e7c-b055-47d3-874a-227d5f02432a-kube-api-access-5ns8r\") pod \"memcached-0\" (UID: \"8b9e8e7c-b055-47d3-874a-227d5f02432a\") " pod="openstack/memcached-0" Nov 25 10:47:04 crc kubenswrapper[4821]: I1125 10:47:04.458351 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41c8c049-b466-4c19-944e-50a639491190-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"41c8c049-b466-4c19-944e-50a639491190\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:47:04 crc kubenswrapper[4821]: I1125 10:47:04.458377 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-cell1-galera-0\" (UID: \"41c8c049-b466-4c19-944e-50a639491190\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:47:04 crc kubenswrapper[4821]: I1125 10:47:04.458415 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/41c8c049-b466-4c19-944e-50a639491190-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"41c8c049-b466-4c19-944e-50a639491190\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:47:04 crc kubenswrapper[4821]: I1125 10:47:04.458434 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/41c8c049-b466-4c19-944e-50a639491190-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"41c8c049-b466-4c19-944e-50a639491190\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:47:04 crc kubenswrapper[4821]: I1125 10:47:04.458454 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8b9e8e7c-b055-47d3-874a-227d5f02432a-kolla-config\") pod \"memcached-0\" (UID: \"8b9e8e7c-b055-47d3-874a-227d5f02432a\") " pod="openstack/memcached-0" Nov 25 10:47:04 crc kubenswrapper[4821]: I1125 10:47:04.458476 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b9e8e7c-b055-47d3-874a-227d5f02432a-combined-ca-bundle\") pod \"memcached-0\" (UID: \"8b9e8e7c-b055-47d3-874a-227d5f02432a\") " pod="openstack/memcached-0" Nov 25 10:47:04 crc kubenswrapper[4821]: I1125 10:47:04.458496 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/41c8c049-b466-4c19-944e-50a639491190-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"41c8c049-b466-4c19-944e-50a639491190\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:47:04 crc kubenswrapper[4821]: I1125 10:47:04.459214 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/41c8c049-b466-4c19-944e-50a639491190-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"41c8c049-b466-4c19-944e-50a639491190\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:47:04 crc kubenswrapper[4821]: I1125 10:47:04.460009 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/41c8c049-b466-4c19-944e-50a639491190-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"41c8c049-b466-4c19-944e-50a639491190\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:47:04 crc kubenswrapper[4821]: I1125 10:47:04.460797 4821 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-cell1-galera-0\" (UID: \"41c8c049-b466-4c19-944e-50a639491190\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/openstack-cell1-galera-0" Nov 25 10:47:04 crc kubenswrapper[4821]: I1125 10:47:04.461216 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/41c8c049-b466-4c19-944e-50a639491190-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"41c8c049-b466-4c19-944e-50a639491190\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:47:04 crc kubenswrapper[4821]: I1125 10:47:04.461733 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/41c8c049-b466-4c19-944e-50a639491190-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"41c8c049-b466-4c19-944e-50a639491190\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:47:04 crc kubenswrapper[4821]: I1125 10:47:04.471138 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/41c8c049-b466-4c19-944e-50a639491190-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"41c8c049-b466-4c19-944e-50a639491190\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:47:04 crc kubenswrapper[4821]: I1125 10:47:04.472651 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41c8c049-b466-4c19-944e-50a639491190-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"41c8c049-b466-4c19-944e-50a639491190\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:47:04 crc kubenswrapper[4821]: I1125 10:47:04.475525 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w8gpb\" (UniqueName: \"kubernetes.io/projected/41c8c049-b466-4c19-944e-50a639491190-kube-api-access-w8gpb\") pod \"openstack-cell1-galera-0\" (UID: \"41c8c049-b466-4c19-944e-50a639491190\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:47:04 crc kubenswrapper[4821]: I1125 10:47:04.488589 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-cell1-galera-0\" (UID: \"41c8c049-b466-4c19-944e-50a639491190\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:47:04 crc kubenswrapper[4821]: I1125 10:47:04.559386 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8b9e8e7c-b055-47d3-874a-227d5f02432a-kolla-config\") pod \"memcached-0\" (UID: \"8b9e8e7c-b055-47d3-874a-227d5f02432a\") " pod="openstack/memcached-0" Nov 25 10:47:04 crc kubenswrapper[4821]: I1125 10:47:04.559432 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b9e8e7c-b055-47d3-874a-227d5f02432a-combined-ca-bundle\") pod \"memcached-0\" (UID: \"8b9e8e7c-b055-47d3-874a-227d5f02432a\") " pod="openstack/memcached-0" Nov 25 10:47:04 crc kubenswrapper[4821]: I1125 10:47:04.559473 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8b9e8e7c-b055-47d3-874a-227d5f02432a-config-data\") pod \"memcached-0\" (UID: \"8b9e8e7c-b055-47d3-874a-227d5f02432a\") " pod="openstack/memcached-0" Nov 25 10:47:04 crc kubenswrapper[4821]: I1125 10:47:04.559488 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b9e8e7c-b055-47d3-874a-227d5f02432a-memcached-tls-certs\") pod \"memcached-0\" (UID: \"8b9e8e7c-b055-47d3-874a-227d5f02432a\") " pod="openstack/memcached-0" Nov 25 10:47:04 crc kubenswrapper[4821]: I1125 10:47:04.559502 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ns8r\" (UniqueName: \"kubernetes.io/projected/8b9e8e7c-b055-47d3-874a-227d5f02432a-kube-api-access-5ns8r\") pod \"memcached-0\" (UID: \"8b9e8e7c-b055-47d3-874a-227d5f02432a\") " pod="openstack/memcached-0" Nov 25 10:47:04 crc kubenswrapper[4821]: I1125 10:47:04.560459 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8b9e8e7c-b055-47d3-874a-227d5f02432a-kolla-config\") pod \"memcached-0\" (UID: \"8b9e8e7c-b055-47d3-874a-227d5f02432a\") " pod="openstack/memcached-0" Nov 25 10:47:04 crc kubenswrapper[4821]: I1125 10:47:04.561809 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8b9e8e7c-b055-47d3-874a-227d5f02432a-config-data\") pod \"memcached-0\" (UID: \"8b9e8e7c-b055-47d3-874a-227d5f02432a\") " pod="openstack/memcached-0" Nov 25 10:47:04 crc kubenswrapper[4821]: I1125 10:47:04.563818 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b9e8e7c-b055-47d3-874a-227d5f02432a-combined-ca-bundle\") pod \"memcached-0\" (UID: \"8b9e8e7c-b055-47d3-874a-227d5f02432a\") " pod="openstack/memcached-0" Nov 25 10:47:04 crc kubenswrapper[4821]: I1125 10:47:04.579016 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b9e8e7c-b055-47d3-874a-227d5f02432a-memcached-tls-certs\") pod \"memcached-0\" (UID: \"8b9e8e7c-b055-47d3-874a-227d5f02432a\") " pod="openstack/memcached-0" Nov 25 10:47:04 crc kubenswrapper[4821]: I1125 10:47:04.583508 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ns8r\" (UniqueName: \"kubernetes.io/projected/8b9e8e7c-b055-47d3-874a-227d5f02432a-kube-api-access-5ns8r\") pod \"memcached-0\" (UID: \"8b9e8e7c-b055-47d3-874a-227d5f02432a\") " pod="openstack/memcached-0" Nov 25 10:47:04 crc kubenswrapper[4821]: I1125 10:47:04.681989 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 25 10:47:04 crc kubenswrapper[4821]: I1125 10:47:04.786727 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 25 10:47:06 crc kubenswrapper[4821]: I1125 10:47:06.401561 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 25 10:47:06 crc kubenswrapper[4821]: I1125 10:47:06.402749 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 25 10:47:06 crc kubenswrapper[4821]: I1125 10:47:06.405759 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-vxwcf" Nov 25 10:47:06 crc kubenswrapper[4821]: I1125 10:47:06.414873 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 25 10:47:06 crc kubenswrapper[4821]: I1125 10:47:06.592375 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwnx2\" (UniqueName: \"kubernetes.io/projected/e2b8639b-42d3-4de6-8d08-d5e2aac42391-kube-api-access-rwnx2\") pod \"kube-state-metrics-0\" (UID: \"e2b8639b-42d3-4de6-8d08-d5e2aac42391\") " pod="openstack/kube-state-metrics-0" Nov 25 10:47:06 crc kubenswrapper[4821]: I1125 10:47:06.693884 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rwnx2\" (UniqueName: \"kubernetes.io/projected/e2b8639b-42d3-4de6-8d08-d5e2aac42391-kube-api-access-rwnx2\") pod \"kube-state-metrics-0\" (UID: \"e2b8639b-42d3-4de6-8d08-d5e2aac42391\") " pod="openstack/kube-state-metrics-0" Nov 25 10:47:06 crc kubenswrapper[4821]: I1125 10:47:06.712384 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwnx2\" (UniqueName: \"kubernetes.io/projected/e2b8639b-42d3-4de6-8d08-d5e2aac42391-kube-api-access-rwnx2\") pod \"kube-state-metrics-0\" (UID: \"e2b8639b-42d3-4de6-8d08-d5e2aac42391\") " pod="openstack/kube-state-metrics-0" Nov 25 10:47:06 crc kubenswrapper[4821]: I1125 10:47:06.720893 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 25 10:47:08 crc kubenswrapper[4821]: W1125 10:47:08.029412 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf289352e_7670_4503_b6bf_ccdaa45a7c36.slice/crio-a08708b9197edf8c74a04f9d6c70e52120c11bdcb6e9d4c4e246a60c04eb6ddd WatchSource:0}: Error finding container a08708b9197edf8c74a04f9d6c70e52120c11bdcb6e9d4c4e246a60c04eb6ddd: Status 404 returned error can't find the container with id a08708b9197edf8c74a04f9d6c70e52120c11bdcb6e9d4c4e246a60c04eb6ddd Nov 25 10:47:08 crc kubenswrapper[4821]: I1125 10:47:08.093740 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8phxk" Nov 25 10:47:08 crc kubenswrapper[4821]: I1125 10:47:08.215070 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4l272\" (UniqueName: \"kubernetes.io/projected/e93e112b-d64e-4bab-b62d-cfee3e1a74eb-kube-api-access-4l272\") pod \"e93e112b-d64e-4bab-b62d-cfee3e1a74eb\" (UID: \"e93e112b-d64e-4bab-b62d-cfee3e1a74eb\") " Nov 25 10:47:08 crc kubenswrapper[4821]: I1125 10:47:08.215207 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e93e112b-d64e-4bab-b62d-cfee3e1a74eb-utilities\") pod \"e93e112b-d64e-4bab-b62d-cfee3e1a74eb\" (UID: \"e93e112b-d64e-4bab-b62d-cfee3e1a74eb\") " Nov 25 10:47:08 crc kubenswrapper[4821]: I1125 10:47:08.215239 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e93e112b-d64e-4bab-b62d-cfee3e1a74eb-catalog-content\") pod \"e93e112b-d64e-4bab-b62d-cfee3e1a74eb\" (UID: \"e93e112b-d64e-4bab-b62d-cfee3e1a74eb\") " Nov 25 10:47:08 crc kubenswrapper[4821]: I1125 10:47:08.216931 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e93e112b-d64e-4bab-b62d-cfee3e1a74eb-utilities" (OuterVolumeSpecName: "utilities") pod "e93e112b-d64e-4bab-b62d-cfee3e1a74eb" (UID: "e93e112b-d64e-4bab-b62d-cfee3e1a74eb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:47:08 crc kubenswrapper[4821]: I1125 10:47:08.219881 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e93e112b-d64e-4bab-b62d-cfee3e1a74eb-kube-api-access-4l272" (OuterVolumeSpecName: "kube-api-access-4l272") pod "e93e112b-d64e-4bab-b62d-cfee3e1a74eb" (UID: "e93e112b-d64e-4bab-b62d-cfee3e1a74eb"). InnerVolumeSpecName "kube-api-access-4l272". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:47:08 crc kubenswrapper[4821]: I1125 10:47:08.274092 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e93e112b-d64e-4bab-b62d-cfee3e1a74eb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e93e112b-d64e-4bab-b62d-cfee3e1a74eb" (UID: "e93e112b-d64e-4bab-b62d-cfee3e1a74eb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:47:08 crc kubenswrapper[4821]: I1125 10:47:08.316407 4821 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e93e112b-d64e-4bab-b62d-cfee3e1a74eb-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:08 crc kubenswrapper[4821]: I1125 10:47:08.316443 4821 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e93e112b-d64e-4bab-b62d-cfee3e1a74eb-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:08 crc kubenswrapper[4821]: I1125 10:47:08.316456 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4l272\" (UniqueName: \"kubernetes.io/projected/e93e112b-d64e-4bab-b62d-cfee3e1a74eb-kube-api-access-4l272\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:08 crc kubenswrapper[4821]: I1125 10:47:08.413843 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 25 10:47:08 crc kubenswrapper[4821]: I1125 10:47:08.684417 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-h5qkz" event={"ID":"f289352e-7670-4503-b6bf-ccdaa45a7c36","Type":"ContainerStarted","Data":"a08708b9197edf8c74a04f9d6c70e52120c11bdcb6e9d4c4e246a60c04eb6ddd"} Nov 25 10:47:08 crc kubenswrapper[4821]: I1125 10:47:08.686961 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8phxk" event={"ID":"e93e112b-d64e-4bab-b62d-cfee3e1a74eb","Type":"ContainerDied","Data":"97514406c4d204378d6867d011b122723a510d456aeb008ebbd1442cc6656fb1"} Nov 25 10:47:08 crc kubenswrapper[4821]: I1125 10:47:08.687039 4821 scope.go:117] "RemoveContainer" containerID="afb0d51c47960d814a1cea6065cbd96a29f2014cf8af36ff98b91e9fa73c658e" Nov 25 10:47:08 crc kubenswrapper[4821]: I1125 10:47:08.687347 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8phxk" Nov 25 10:47:08 crc kubenswrapper[4821]: I1125 10:47:08.725141 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8phxk"] Nov 25 10:47:08 crc kubenswrapper[4821]: I1125 10:47:08.732856 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-8phxk"] Nov 25 10:47:10 crc kubenswrapper[4821]: I1125 10:47:10.039990 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ltxc6"] Nov 25 10:47:10 crc kubenswrapper[4821]: E1125 10:47:10.041106 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e93e112b-d64e-4bab-b62d-cfee3e1a74eb" containerName="registry-server" Nov 25 10:47:10 crc kubenswrapper[4821]: I1125 10:47:10.041123 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="e93e112b-d64e-4bab-b62d-cfee3e1a74eb" containerName="registry-server" Nov 25 10:47:10 crc kubenswrapper[4821]: E1125 10:47:10.041158 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e93e112b-d64e-4bab-b62d-cfee3e1a74eb" containerName="extract-content" Nov 25 10:47:10 crc kubenswrapper[4821]: I1125 10:47:10.041183 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="e93e112b-d64e-4bab-b62d-cfee3e1a74eb" containerName="extract-content" Nov 25 10:47:10 crc kubenswrapper[4821]: E1125 10:47:10.041207 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e93e112b-d64e-4bab-b62d-cfee3e1a74eb" containerName="extract-utilities" Nov 25 10:47:10 crc kubenswrapper[4821]: I1125 10:47:10.041218 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="e93e112b-d64e-4bab-b62d-cfee3e1a74eb" containerName="extract-utilities" Nov 25 10:47:10 crc kubenswrapper[4821]: I1125 10:47:10.041404 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="e93e112b-d64e-4bab-b62d-cfee3e1a74eb" containerName="registry-server" Nov 25 10:47:10 crc kubenswrapper[4821]: I1125 10:47:10.042049 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ltxc6" Nov 25 10:47:10 crc kubenswrapper[4821]: I1125 10:47:10.044952 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-n2h5f" Nov 25 10:47:10 crc kubenswrapper[4821]: I1125 10:47:10.045257 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Nov 25 10:47:10 crc kubenswrapper[4821]: I1125 10:47:10.045415 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Nov 25 10:47:10 crc kubenswrapper[4821]: I1125 10:47:10.050385 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ltxc6"] Nov 25 10:47:10 crc kubenswrapper[4821]: I1125 10:47:10.086656 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-6mbr2"] Nov 25 10:47:10 crc kubenswrapper[4821]: I1125 10:47:10.089885 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-6mbr2" Nov 25 10:47:10 crc kubenswrapper[4821]: I1125 10:47:10.104990 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-6mbr2"] Nov 25 10:47:10 crc kubenswrapper[4821]: I1125 10:47:10.128794 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e93e112b-d64e-4bab-b62d-cfee3e1a74eb" path="/var/lib/kubelet/pods/e93e112b-d64e-4bab-b62d-cfee3e1a74eb/volumes" Nov 25 10:47:10 crc kubenswrapper[4821]: I1125 10:47:10.141559 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/47eaf58d-190b-4e75-9aac-11d7e51df259-var-run\") pod \"ovn-controller-ltxc6\" (UID: \"47eaf58d-190b-4e75-9aac-11d7e51df259\") " pod="openstack/ovn-controller-ltxc6" Nov 25 10:47:10 crc kubenswrapper[4821]: I1125 10:47:10.141612 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/47eaf58d-190b-4e75-9aac-11d7e51df259-ovn-controller-tls-certs\") pod \"ovn-controller-ltxc6\" (UID: \"47eaf58d-190b-4e75-9aac-11d7e51df259\") " pod="openstack/ovn-controller-ltxc6" Nov 25 10:47:10 crc kubenswrapper[4821]: I1125 10:47:10.141657 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/47eaf58d-190b-4e75-9aac-11d7e51df259-scripts\") pod \"ovn-controller-ltxc6\" (UID: \"47eaf58d-190b-4e75-9aac-11d7e51df259\") " pod="openstack/ovn-controller-ltxc6" Nov 25 10:47:10 crc kubenswrapper[4821]: I1125 10:47:10.141694 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8tsd\" (UniqueName: \"kubernetes.io/projected/47eaf58d-190b-4e75-9aac-11d7e51df259-kube-api-access-l8tsd\") pod \"ovn-controller-ltxc6\" (UID: \"47eaf58d-190b-4e75-9aac-11d7e51df259\") " pod="openstack/ovn-controller-ltxc6" Nov 25 10:47:10 crc kubenswrapper[4821]: I1125 10:47:10.141737 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/47eaf58d-190b-4e75-9aac-11d7e51df259-var-run-ovn\") pod \"ovn-controller-ltxc6\" (UID: \"47eaf58d-190b-4e75-9aac-11d7e51df259\") " pod="openstack/ovn-controller-ltxc6" Nov 25 10:47:10 crc kubenswrapper[4821]: I1125 10:47:10.141774 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/47eaf58d-190b-4e75-9aac-11d7e51df259-var-log-ovn\") pod \"ovn-controller-ltxc6\" (UID: \"47eaf58d-190b-4e75-9aac-11d7e51df259\") " pod="openstack/ovn-controller-ltxc6" Nov 25 10:47:10 crc kubenswrapper[4821]: I1125 10:47:10.141810 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47eaf58d-190b-4e75-9aac-11d7e51df259-combined-ca-bundle\") pod \"ovn-controller-ltxc6\" (UID: \"47eaf58d-190b-4e75-9aac-11d7e51df259\") " pod="openstack/ovn-controller-ltxc6" Nov 25 10:47:10 crc kubenswrapper[4821]: I1125 10:47:10.243845 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/72b307b1-f36e-424f-91ef-f7402b6d7535-etc-ovs\") pod \"ovn-controller-ovs-6mbr2\" (UID: \"72b307b1-f36e-424f-91ef-f7402b6d7535\") " pod="openstack/ovn-controller-ovs-6mbr2" Nov 25 10:47:10 crc kubenswrapper[4821]: I1125 10:47:10.243906 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47eaf58d-190b-4e75-9aac-11d7e51df259-combined-ca-bundle\") pod \"ovn-controller-ltxc6\" (UID: \"47eaf58d-190b-4e75-9aac-11d7e51df259\") " pod="openstack/ovn-controller-ltxc6" Nov 25 10:47:10 crc kubenswrapper[4821]: I1125 10:47:10.243941 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/72b307b1-f36e-424f-91ef-f7402b6d7535-var-run\") pod \"ovn-controller-ovs-6mbr2\" (UID: \"72b307b1-f36e-424f-91ef-f7402b6d7535\") " pod="openstack/ovn-controller-ovs-6mbr2" Nov 25 10:47:10 crc kubenswrapper[4821]: I1125 10:47:10.244051 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/47eaf58d-190b-4e75-9aac-11d7e51df259-var-run\") pod \"ovn-controller-ltxc6\" (UID: \"47eaf58d-190b-4e75-9aac-11d7e51df259\") " pod="openstack/ovn-controller-ltxc6" Nov 25 10:47:10 crc kubenswrapper[4821]: I1125 10:47:10.244095 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/47eaf58d-190b-4e75-9aac-11d7e51df259-ovn-controller-tls-certs\") pod \"ovn-controller-ltxc6\" (UID: \"47eaf58d-190b-4e75-9aac-11d7e51df259\") " pod="openstack/ovn-controller-ltxc6" Nov 25 10:47:10 crc kubenswrapper[4821]: I1125 10:47:10.244151 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/72b307b1-f36e-424f-91ef-f7402b6d7535-scripts\") pod \"ovn-controller-ovs-6mbr2\" (UID: \"72b307b1-f36e-424f-91ef-f7402b6d7535\") " pod="openstack/ovn-controller-ovs-6mbr2" Nov 25 10:47:10 crc kubenswrapper[4821]: I1125 10:47:10.244186 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/72b307b1-f36e-424f-91ef-f7402b6d7535-var-lib\") pod \"ovn-controller-ovs-6mbr2\" (UID: \"72b307b1-f36e-424f-91ef-f7402b6d7535\") " pod="openstack/ovn-controller-ovs-6mbr2" Nov 25 10:47:10 crc kubenswrapper[4821]: I1125 10:47:10.244209 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/47eaf58d-190b-4e75-9aac-11d7e51df259-scripts\") pod \"ovn-controller-ltxc6\" (UID: \"47eaf58d-190b-4e75-9aac-11d7e51df259\") " pod="openstack/ovn-controller-ltxc6" Nov 25 10:47:10 crc kubenswrapper[4821]: I1125 10:47:10.244236 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fcxnj\" (UniqueName: \"kubernetes.io/projected/72b307b1-f36e-424f-91ef-f7402b6d7535-kube-api-access-fcxnj\") pod \"ovn-controller-ovs-6mbr2\" (UID: \"72b307b1-f36e-424f-91ef-f7402b6d7535\") " pod="openstack/ovn-controller-ovs-6mbr2" Nov 25 10:47:10 crc kubenswrapper[4821]: I1125 10:47:10.244267 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8tsd\" (UniqueName: \"kubernetes.io/projected/47eaf58d-190b-4e75-9aac-11d7e51df259-kube-api-access-l8tsd\") pod \"ovn-controller-ltxc6\" (UID: \"47eaf58d-190b-4e75-9aac-11d7e51df259\") " pod="openstack/ovn-controller-ltxc6" Nov 25 10:47:10 crc kubenswrapper[4821]: I1125 10:47:10.244295 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/72b307b1-f36e-424f-91ef-f7402b6d7535-var-log\") pod \"ovn-controller-ovs-6mbr2\" (UID: \"72b307b1-f36e-424f-91ef-f7402b6d7535\") " pod="openstack/ovn-controller-ovs-6mbr2" Nov 25 10:47:10 crc kubenswrapper[4821]: I1125 10:47:10.244330 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/47eaf58d-190b-4e75-9aac-11d7e51df259-var-run-ovn\") pod \"ovn-controller-ltxc6\" (UID: \"47eaf58d-190b-4e75-9aac-11d7e51df259\") " pod="openstack/ovn-controller-ltxc6" Nov 25 10:47:10 crc kubenswrapper[4821]: I1125 10:47:10.244396 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/47eaf58d-190b-4e75-9aac-11d7e51df259-var-log-ovn\") pod \"ovn-controller-ltxc6\" (UID: \"47eaf58d-190b-4e75-9aac-11d7e51df259\") " pod="openstack/ovn-controller-ltxc6" Nov 25 10:47:10 crc kubenswrapper[4821]: I1125 10:47:10.244977 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/47eaf58d-190b-4e75-9aac-11d7e51df259-var-log-ovn\") pod \"ovn-controller-ltxc6\" (UID: \"47eaf58d-190b-4e75-9aac-11d7e51df259\") " pod="openstack/ovn-controller-ltxc6" Nov 25 10:47:10 crc kubenswrapper[4821]: I1125 10:47:10.245853 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/47eaf58d-190b-4e75-9aac-11d7e51df259-var-run-ovn\") pod \"ovn-controller-ltxc6\" (UID: \"47eaf58d-190b-4e75-9aac-11d7e51df259\") " pod="openstack/ovn-controller-ltxc6" Nov 25 10:47:10 crc kubenswrapper[4821]: I1125 10:47:10.245816 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/47eaf58d-190b-4e75-9aac-11d7e51df259-var-run\") pod \"ovn-controller-ltxc6\" (UID: \"47eaf58d-190b-4e75-9aac-11d7e51df259\") " pod="openstack/ovn-controller-ltxc6" Nov 25 10:47:10 crc kubenswrapper[4821]: I1125 10:47:10.248658 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Nov 25 10:47:10 crc kubenswrapper[4821]: I1125 10:47:10.249210 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Nov 25 10:47:10 crc kubenswrapper[4821]: I1125 10:47:10.257950 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/47eaf58d-190b-4e75-9aac-11d7e51df259-scripts\") pod \"ovn-controller-ltxc6\" (UID: \"47eaf58d-190b-4e75-9aac-11d7e51df259\") " pod="openstack/ovn-controller-ltxc6" Nov 25 10:47:10 crc kubenswrapper[4821]: I1125 10:47:10.266609 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/47eaf58d-190b-4e75-9aac-11d7e51df259-ovn-controller-tls-certs\") pod \"ovn-controller-ltxc6\" (UID: \"47eaf58d-190b-4e75-9aac-11d7e51df259\") " pod="openstack/ovn-controller-ltxc6" Nov 25 10:47:10 crc kubenswrapper[4821]: I1125 10:47:10.274662 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8tsd\" (UniqueName: \"kubernetes.io/projected/47eaf58d-190b-4e75-9aac-11d7e51df259-kube-api-access-l8tsd\") pod \"ovn-controller-ltxc6\" (UID: \"47eaf58d-190b-4e75-9aac-11d7e51df259\") " pod="openstack/ovn-controller-ltxc6" Nov 25 10:47:10 crc kubenswrapper[4821]: I1125 10:47:10.312421 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47eaf58d-190b-4e75-9aac-11d7e51df259-combined-ca-bundle\") pod \"ovn-controller-ltxc6\" (UID: \"47eaf58d-190b-4e75-9aac-11d7e51df259\") " pod="openstack/ovn-controller-ltxc6" Nov 25 10:47:10 crc kubenswrapper[4821]: I1125 10:47:10.352388 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fcxnj\" (UniqueName: \"kubernetes.io/projected/72b307b1-f36e-424f-91ef-f7402b6d7535-kube-api-access-fcxnj\") pod \"ovn-controller-ovs-6mbr2\" (UID: \"72b307b1-f36e-424f-91ef-f7402b6d7535\") " pod="openstack/ovn-controller-ovs-6mbr2" Nov 25 10:47:10 crc kubenswrapper[4821]: I1125 10:47:10.354602 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/72b307b1-f36e-424f-91ef-f7402b6d7535-var-log\") pod \"ovn-controller-ovs-6mbr2\" (UID: \"72b307b1-f36e-424f-91ef-f7402b6d7535\") " pod="openstack/ovn-controller-ovs-6mbr2" Nov 25 10:47:10 crc kubenswrapper[4821]: I1125 10:47:10.354735 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/72b307b1-f36e-424f-91ef-f7402b6d7535-etc-ovs\") pod \"ovn-controller-ovs-6mbr2\" (UID: \"72b307b1-f36e-424f-91ef-f7402b6d7535\") " pod="openstack/ovn-controller-ovs-6mbr2" Nov 25 10:47:10 crc kubenswrapper[4821]: I1125 10:47:10.354798 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/72b307b1-f36e-424f-91ef-f7402b6d7535-var-run\") pod \"ovn-controller-ovs-6mbr2\" (UID: \"72b307b1-f36e-424f-91ef-f7402b6d7535\") " pod="openstack/ovn-controller-ovs-6mbr2" Nov 25 10:47:10 crc kubenswrapper[4821]: I1125 10:47:10.354950 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/72b307b1-f36e-424f-91ef-f7402b6d7535-scripts\") pod \"ovn-controller-ovs-6mbr2\" (UID: \"72b307b1-f36e-424f-91ef-f7402b6d7535\") " pod="openstack/ovn-controller-ovs-6mbr2" Nov 25 10:47:10 crc kubenswrapper[4821]: I1125 10:47:10.354973 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/72b307b1-f36e-424f-91ef-f7402b6d7535-var-lib\") pod \"ovn-controller-ovs-6mbr2\" (UID: \"72b307b1-f36e-424f-91ef-f7402b6d7535\") " pod="openstack/ovn-controller-ovs-6mbr2" Nov 25 10:47:10 crc kubenswrapper[4821]: I1125 10:47:10.355230 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/72b307b1-f36e-424f-91ef-f7402b6d7535-var-lib\") pod \"ovn-controller-ovs-6mbr2\" (UID: \"72b307b1-f36e-424f-91ef-f7402b6d7535\") " pod="openstack/ovn-controller-ovs-6mbr2" Nov 25 10:47:10 crc kubenswrapper[4821]: I1125 10:47:10.355572 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/72b307b1-f36e-424f-91ef-f7402b6d7535-var-log\") pod \"ovn-controller-ovs-6mbr2\" (UID: \"72b307b1-f36e-424f-91ef-f7402b6d7535\") " pod="openstack/ovn-controller-ovs-6mbr2" Nov 25 10:47:10 crc kubenswrapper[4821]: I1125 10:47:10.355698 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/72b307b1-f36e-424f-91ef-f7402b6d7535-etc-ovs\") pod \"ovn-controller-ovs-6mbr2\" (UID: \"72b307b1-f36e-424f-91ef-f7402b6d7535\") " pod="openstack/ovn-controller-ovs-6mbr2" Nov 25 10:47:10 crc kubenswrapper[4821]: I1125 10:47:10.355739 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/72b307b1-f36e-424f-91ef-f7402b6d7535-var-run\") pod \"ovn-controller-ovs-6mbr2\" (UID: \"72b307b1-f36e-424f-91ef-f7402b6d7535\") " pod="openstack/ovn-controller-ovs-6mbr2" Nov 25 10:47:10 crc kubenswrapper[4821]: I1125 10:47:10.357399 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/72b307b1-f36e-424f-91ef-f7402b6d7535-scripts\") pod \"ovn-controller-ovs-6mbr2\" (UID: \"72b307b1-f36e-424f-91ef-f7402b6d7535\") " pod="openstack/ovn-controller-ovs-6mbr2" Nov 25 10:47:10 crc kubenswrapper[4821]: I1125 10:47:10.363242 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-n2h5f" Nov 25 10:47:10 crc kubenswrapper[4821]: I1125 10:47:10.370790 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ltxc6" Nov 25 10:47:10 crc kubenswrapper[4821]: I1125 10:47:10.374940 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fcxnj\" (UniqueName: \"kubernetes.io/projected/72b307b1-f36e-424f-91ef-f7402b6d7535-kube-api-access-fcxnj\") pod \"ovn-controller-ovs-6mbr2\" (UID: \"72b307b1-f36e-424f-91ef-f7402b6d7535\") " pod="openstack/ovn-controller-ovs-6mbr2" Nov 25 10:47:10 crc kubenswrapper[4821]: I1125 10:47:10.408491 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-6mbr2" Nov 25 10:47:10 crc kubenswrapper[4821]: I1125 10:47:10.932661 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 25 10:47:10 crc kubenswrapper[4821]: I1125 10:47:10.934325 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 25 10:47:10 crc kubenswrapper[4821]: I1125 10:47:10.937314 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-rklsp" Nov 25 10:47:10 crc kubenswrapper[4821]: I1125 10:47:10.937346 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Nov 25 10:47:10 crc kubenswrapper[4821]: I1125 10:47:10.937750 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Nov 25 10:47:10 crc kubenswrapper[4821]: I1125 10:47:10.937857 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Nov 25 10:47:10 crc kubenswrapper[4821]: I1125 10:47:10.937893 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Nov 25 10:47:10 crc kubenswrapper[4821]: I1125 10:47:10.940402 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 25 10:47:11 crc kubenswrapper[4821]: I1125 10:47:11.065678 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30500d7b-4e4b-4839-8a9d-0b05a0ebd1fb-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"30500d7b-4e4b-4839-8a9d-0b05a0ebd1fb\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:47:11 crc kubenswrapper[4821]: I1125 10:47:11.065747 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-nb-0\" (UID: \"30500d7b-4e4b-4839-8a9d-0b05a0ebd1fb\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:47:11 crc kubenswrapper[4821]: I1125 10:47:11.065801 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/30500d7b-4e4b-4839-8a9d-0b05a0ebd1fb-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"30500d7b-4e4b-4839-8a9d-0b05a0ebd1fb\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:47:11 crc kubenswrapper[4821]: I1125 10:47:11.065870 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/30500d7b-4e4b-4839-8a9d-0b05a0ebd1fb-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"30500d7b-4e4b-4839-8a9d-0b05a0ebd1fb\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:47:11 crc kubenswrapper[4821]: I1125 10:47:11.065902 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/30500d7b-4e4b-4839-8a9d-0b05a0ebd1fb-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"30500d7b-4e4b-4839-8a9d-0b05a0ebd1fb\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:47:11 crc kubenswrapper[4821]: I1125 10:47:11.065936 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxpwr\" (UniqueName: \"kubernetes.io/projected/30500d7b-4e4b-4839-8a9d-0b05a0ebd1fb-kube-api-access-xxpwr\") pod \"ovsdbserver-nb-0\" (UID: \"30500d7b-4e4b-4839-8a9d-0b05a0ebd1fb\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:47:11 crc kubenswrapper[4821]: I1125 10:47:11.065968 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/30500d7b-4e4b-4839-8a9d-0b05a0ebd1fb-config\") pod \"ovsdbserver-nb-0\" (UID: \"30500d7b-4e4b-4839-8a9d-0b05a0ebd1fb\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:47:11 crc kubenswrapper[4821]: I1125 10:47:11.065990 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/30500d7b-4e4b-4839-8a9d-0b05a0ebd1fb-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"30500d7b-4e4b-4839-8a9d-0b05a0ebd1fb\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:47:11 crc kubenswrapper[4821]: I1125 10:47:11.167033 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/30500d7b-4e4b-4839-8a9d-0b05a0ebd1fb-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"30500d7b-4e4b-4839-8a9d-0b05a0ebd1fb\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:47:11 crc kubenswrapper[4821]: I1125 10:47:11.167092 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/30500d7b-4e4b-4839-8a9d-0b05a0ebd1fb-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"30500d7b-4e4b-4839-8a9d-0b05a0ebd1fb\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:47:11 crc kubenswrapper[4821]: I1125 10:47:11.167116 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxpwr\" (UniqueName: \"kubernetes.io/projected/30500d7b-4e4b-4839-8a9d-0b05a0ebd1fb-kube-api-access-xxpwr\") pod \"ovsdbserver-nb-0\" (UID: \"30500d7b-4e4b-4839-8a9d-0b05a0ebd1fb\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:47:11 crc kubenswrapper[4821]: I1125 10:47:11.167144 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/30500d7b-4e4b-4839-8a9d-0b05a0ebd1fb-config\") pod \"ovsdbserver-nb-0\" (UID: \"30500d7b-4e4b-4839-8a9d-0b05a0ebd1fb\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:47:11 crc kubenswrapper[4821]: I1125 10:47:11.167222 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/30500d7b-4e4b-4839-8a9d-0b05a0ebd1fb-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"30500d7b-4e4b-4839-8a9d-0b05a0ebd1fb\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:47:11 crc kubenswrapper[4821]: I1125 10:47:11.167282 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30500d7b-4e4b-4839-8a9d-0b05a0ebd1fb-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"30500d7b-4e4b-4839-8a9d-0b05a0ebd1fb\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:47:11 crc kubenswrapper[4821]: I1125 10:47:11.167315 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-nb-0\" (UID: \"30500d7b-4e4b-4839-8a9d-0b05a0ebd1fb\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:47:11 crc kubenswrapper[4821]: I1125 10:47:11.167334 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/30500d7b-4e4b-4839-8a9d-0b05a0ebd1fb-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"30500d7b-4e4b-4839-8a9d-0b05a0ebd1fb\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:47:11 crc kubenswrapper[4821]: I1125 10:47:11.167531 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/30500d7b-4e4b-4839-8a9d-0b05a0ebd1fb-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"30500d7b-4e4b-4839-8a9d-0b05a0ebd1fb\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:47:11 crc kubenswrapper[4821]: I1125 10:47:11.168489 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/30500d7b-4e4b-4839-8a9d-0b05a0ebd1fb-config\") pod \"ovsdbserver-nb-0\" (UID: \"30500d7b-4e4b-4839-8a9d-0b05a0ebd1fb\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:47:11 crc kubenswrapper[4821]: I1125 10:47:11.168869 4821 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-nb-0\" (UID: \"30500d7b-4e4b-4839-8a9d-0b05a0ebd1fb\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/ovsdbserver-nb-0" Nov 25 10:47:11 crc kubenswrapper[4821]: I1125 10:47:11.169142 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/30500d7b-4e4b-4839-8a9d-0b05a0ebd1fb-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"30500d7b-4e4b-4839-8a9d-0b05a0ebd1fb\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:47:11 crc kubenswrapper[4821]: I1125 10:47:11.176950 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/30500d7b-4e4b-4839-8a9d-0b05a0ebd1fb-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"30500d7b-4e4b-4839-8a9d-0b05a0ebd1fb\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:47:11 crc kubenswrapper[4821]: I1125 10:47:11.177081 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/30500d7b-4e4b-4839-8a9d-0b05a0ebd1fb-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"30500d7b-4e4b-4839-8a9d-0b05a0ebd1fb\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:47:11 crc kubenswrapper[4821]: I1125 10:47:11.177417 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30500d7b-4e4b-4839-8a9d-0b05a0ebd1fb-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"30500d7b-4e4b-4839-8a9d-0b05a0ebd1fb\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:47:11 crc kubenswrapper[4821]: I1125 10:47:11.185381 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxpwr\" (UniqueName: \"kubernetes.io/projected/30500d7b-4e4b-4839-8a9d-0b05a0ebd1fb-kube-api-access-xxpwr\") pod \"ovsdbserver-nb-0\" (UID: \"30500d7b-4e4b-4839-8a9d-0b05a0ebd1fb\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:47:11 crc kubenswrapper[4821]: I1125 10:47:11.194397 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-nb-0\" (UID: \"30500d7b-4e4b-4839-8a9d-0b05a0ebd1fb\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:47:11 crc kubenswrapper[4821]: I1125 10:47:11.273199 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 25 10:47:13 crc kubenswrapper[4821]: W1125 10:47:13.297117 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod330a479e_3d76_4de3_bc84_1d7c2d65d265.slice/crio-79ce92a97646f24487ef3a12c92318542dd51e8f2b1e98173fcec59237602084 WatchSource:0}: Error finding container 79ce92a97646f24487ef3a12c92318542dd51e8f2b1e98173fcec59237602084: Status 404 returned error can't find the container with id 79ce92a97646f24487ef3a12c92318542dd51e8f2b1e98173fcec59237602084 Nov 25 10:47:13 crc kubenswrapper[4821]: I1125 10:47:13.739082 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"330a479e-3d76-4de3-bc84-1d7c2d65d265","Type":"ContainerStarted","Data":"79ce92a97646f24487ef3a12c92318542dd51e8f2b1e98173fcec59237602084"} Nov 25 10:47:13 crc kubenswrapper[4821]: I1125 10:47:13.755344 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 25 10:47:13 crc kubenswrapper[4821]: I1125 10:47:13.955199 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 25 10:47:13 crc kubenswrapper[4821]: I1125 10:47:13.957283 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 25 10:47:13 crc kubenswrapper[4821]: I1125 10:47:13.959644 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Nov 25 10:47:13 crc kubenswrapper[4821]: I1125 10:47:13.959645 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Nov 25 10:47:13 crc kubenswrapper[4821]: I1125 10:47:13.961747 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Nov 25 10:47:13 crc kubenswrapper[4821]: I1125 10:47:13.961962 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-mf46l" Nov 25 10:47:13 crc kubenswrapper[4821]: I1125 10:47:13.966395 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 25 10:47:14 crc kubenswrapper[4821]: I1125 10:47:14.114142 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-sb-0\" (UID: \"9aed573a-ad60-479a-ba27-8666d7dc4d9e\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:47:14 crc kubenswrapper[4821]: I1125 10:47:14.114206 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qtbxl\" (UniqueName: \"kubernetes.io/projected/9aed573a-ad60-479a-ba27-8666d7dc4d9e-kube-api-access-qtbxl\") pod \"ovsdbserver-sb-0\" (UID: \"9aed573a-ad60-479a-ba27-8666d7dc4d9e\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:47:14 crc kubenswrapper[4821]: I1125 10:47:14.114247 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9aed573a-ad60-479a-ba27-8666d7dc4d9e-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"9aed573a-ad60-479a-ba27-8666d7dc4d9e\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:47:14 crc kubenswrapper[4821]: I1125 10:47:14.114283 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9aed573a-ad60-479a-ba27-8666d7dc4d9e-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"9aed573a-ad60-479a-ba27-8666d7dc4d9e\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:47:14 crc kubenswrapper[4821]: I1125 10:47:14.114309 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9aed573a-ad60-479a-ba27-8666d7dc4d9e-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"9aed573a-ad60-479a-ba27-8666d7dc4d9e\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:47:14 crc kubenswrapper[4821]: I1125 10:47:14.114326 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/9aed573a-ad60-479a-ba27-8666d7dc4d9e-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"9aed573a-ad60-479a-ba27-8666d7dc4d9e\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:47:14 crc kubenswrapper[4821]: I1125 10:47:14.114360 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9aed573a-ad60-479a-ba27-8666d7dc4d9e-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"9aed573a-ad60-479a-ba27-8666d7dc4d9e\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:47:14 crc kubenswrapper[4821]: I1125 10:47:14.114391 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9aed573a-ad60-479a-ba27-8666d7dc4d9e-config\") pod \"ovsdbserver-sb-0\" (UID: \"9aed573a-ad60-479a-ba27-8666d7dc4d9e\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:47:14 crc kubenswrapper[4821]: I1125 10:47:14.216079 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9aed573a-ad60-479a-ba27-8666d7dc4d9e-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"9aed573a-ad60-479a-ba27-8666d7dc4d9e\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:47:14 crc kubenswrapper[4821]: I1125 10:47:14.216152 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9aed573a-ad60-479a-ba27-8666d7dc4d9e-config\") pod \"ovsdbserver-sb-0\" (UID: \"9aed573a-ad60-479a-ba27-8666d7dc4d9e\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:47:14 crc kubenswrapper[4821]: I1125 10:47:14.216210 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-sb-0\" (UID: \"9aed573a-ad60-479a-ba27-8666d7dc4d9e\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:47:14 crc kubenswrapper[4821]: I1125 10:47:14.216246 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qtbxl\" (UniqueName: \"kubernetes.io/projected/9aed573a-ad60-479a-ba27-8666d7dc4d9e-kube-api-access-qtbxl\") pod \"ovsdbserver-sb-0\" (UID: \"9aed573a-ad60-479a-ba27-8666d7dc4d9e\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:47:14 crc kubenswrapper[4821]: I1125 10:47:14.216319 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9aed573a-ad60-479a-ba27-8666d7dc4d9e-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"9aed573a-ad60-479a-ba27-8666d7dc4d9e\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:47:14 crc kubenswrapper[4821]: I1125 10:47:14.216363 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9aed573a-ad60-479a-ba27-8666d7dc4d9e-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"9aed573a-ad60-479a-ba27-8666d7dc4d9e\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:47:14 crc kubenswrapper[4821]: I1125 10:47:14.216395 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9aed573a-ad60-479a-ba27-8666d7dc4d9e-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"9aed573a-ad60-479a-ba27-8666d7dc4d9e\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:47:14 crc kubenswrapper[4821]: I1125 10:47:14.216414 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/9aed573a-ad60-479a-ba27-8666d7dc4d9e-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"9aed573a-ad60-479a-ba27-8666d7dc4d9e\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:47:14 crc kubenswrapper[4821]: I1125 10:47:14.216592 4821 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-sb-0\" (UID: \"9aed573a-ad60-479a-ba27-8666d7dc4d9e\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/ovsdbserver-sb-0" Nov 25 10:47:14 crc kubenswrapper[4821]: I1125 10:47:14.217194 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9aed573a-ad60-479a-ba27-8666d7dc4d9e-config\") pod \"ovsdbserver-sb-0\" (UID: \"9aed573a-ad60-479a-ba27-8666d7dc4d9e\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:47:14 crc kubenswrapper[4821]: I1125 10:47:14.217902 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/9aed573a-ad60-479a-ba27-8666d7dc4d9e-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"9aed573a-ad60-479a-ba27-8666d7dc4d9e\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:47:14 crc kubenswrapper[4821]: I1125 10:47:14.222198 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9aed573a-ad60-479a-ba27-8666d7dc4d9e-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"9aed573a-ad60-479a-ba27-8666d7dc4d9e\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:47:14 crc kubenswrapper[4821]: I1125 10:47:14.223767 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9aed573a-ad60-479a-ba27-8666d7dc4d9e-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"9aed573a-ad60-479a-ba27-8666d7dc4d9e\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:47:14 crc kubenswrapper[4821]: I1125 10:47:14.230007 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9aed573a-ad60-479a-ba27-8666d7dc4d9e-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"9aed573a-ad60-479a-ba27-8666d7dc4d9e\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:47:14 crc kubenswrapper[4821]: I1125 10:47:14.230524 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9aed573a-ad60-479a-ba27-8666d7dc4d9e-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"9aed573a-ad60-479a-ba27-8666d7dc4d9e\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:47:14 crc kubenswrapper[4821]: I1125 10:47:14.234229 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qtbxl\" (UniqueName: \"kubernetes.io/projected/9aed573a-ad60-479a-ba27-8666d7dc4d9e-kube-api-access-qtbxl\") pod \"ovsdbserver-sb-0\" (UID: \"9aed573a-ad60-479a-ba27-8666d7dc4d9e\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:47:14 crc kubenswrapper[4821]: I1125 10:47:14.243005 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-sb-0\" (UID: \"9aed573a-ad60-479a-ba27-8666d7dc4d9e\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:47:15 crc kubenswrapper[4821]: I1125 10:47:15.117965 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 25 10:47:15 crc kubenswrapper[4821]: I1125 10:47:15.961402 4821 scope.go:117] "RemoveContainer" containerID="37c74a014dadf4083716ab00ae6e0e26e84722e23c5ddd3ddc9d04b54a9a247d" Nov 25 10:47:16 crc kubenswrapper[4821]: E1125 10:47:16.029266 4821 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 25 10:47:16 crc kubenswrapper[4821]: E1125 10:47:16.029703 4821 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-99bb9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-shhlw_openstack(b7a3ac30-3925-493e-9e13-0f1a9a7e49c2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 10:47:16 crc kubenswrapper[4821]: E1125 10:47:16.034368 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-shhlw" podUID="b7a3ac30-3925-493e-9e13-0f1a9a7e49c2" Nov 25 10:47:16 crc kubenswrapper[4821]: E1125 10:47:16.077618 4821 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 25 10:47:16 crc kubenswrapper[4821]: E1125 10:47:16.078076 4821 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xqr4f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-srdjw_openstack(68a5afc7-5e47-4bed-a6c2-8ff5b6b0236c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 10:47:16 crc kubenswrapper[4821]: E1125 10:47:16.079441 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-srdjw" podUID="68a5afc7-5e47-4bed-a6c2-8ff5b6b0236c" Nov 25 10:47:16 crc kubenswrapper[4821]: I1125 10:47:16.150779 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"ecb5eaf2-4178-43ec-bc21-0e05de8fc237","Type":"ContainerStarted","Data":"8e8274789de288c3ed1cf2784513c716879cc63d876ae0363d8f4f4ae197cf08"} Nov 25 10:47:16 crc kubenswrapper[4821]: I1125 10:47:16.215803 4821 scope.go:117] "RemoveContainer" containerID="c8df2db54459c41b5438811f629d46a12cedcf62796b6e4dcdb43e6eefecedc6" Nov 25 10:47:16 crc kubenswrapper[4821]: I1125 10:47:16.599424 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 25 10:47:16 crc kubenswrapper[4821]: I1125 10:47:16.658627 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 25 10:47:16 crc kubenswrapper[4821]: I1125 10:47:16.685683 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 25 10:47:16 crc kubenswrapper[4821]: W1125 10:47:16.732440 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode2b8639b_42d3_4de6_8d08_d5e2aac42391.slice/crio-cd26802fbbd6b9df5a3b0dd72b2f0570f1d220bdc2538ea6a9db8e716dec4df0 WatchSource:0}: Error finding container cd26802fbbd6b9df5a3b0dd72b2f0570f1d220bdc2538ea6a9db8e716dec4df0: Status 404 returned error can't find the container with id cd26802fbbd6b9df5a3b0dd72b2f0570f1d220bdc2538ea6a9db8e716dec4df0 Nov 25 10:47:16 crc kubenswrapper[4821]: I1125 10:47:16.737220 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-shhlw" Nov 25 10:47:16 crc kubenswrapper[4821]: I1125 10:47:16.744241 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7a3ac30-3925-493e-9e13-0f1a9a7e49c2-config\") pod \"b7a3ac30-3925-493e-9e13-0f1a9a7e49c2\" (UID: \"b7a3ac30-3925-493e-9e13-0f1a9a7e49c2\") " Nov 25 10:47:16 crc kubenswrapper[4821]: I1125 10:47:16.744328 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-99bb9\" (UniqueName: \"kubernetes.io/projected/b7a3ac30-3925-493e-9e13-0f1a9a7e49c2-kube-api-access-99bb9\") pod \"b7a3ac30-3925-493e-9e13-0f1a9a7e49c2\" (UID: \"b7a3ac30-3925-493e-9e13-0f1a9a7e49c2\") " Nov 25 10:47:16 crc kubenswrapper[4821]: I1125 10:47:16.745963 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7a3ac30-3925-493e-9e13-0f1a9a7e49c2-config" (OuterVolumeSpecName: "config") pod "b7a3ac30-3925-493e-9e13-0f1a9a7e49c2" (UID: "b7a3ac30-3925-493e-9e13-0f1a9a7e49c2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:47:16 crc kubenswrapper[4821]: I1125 10:47:16.748651 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-srdjw" Nov 25 10:47:16 crc kubenswrapper[4821]: I1125 10:47:16.750537 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7a3ac30-3925-493e-9e13-0f1a9a7e49c2-kube-api-access-99bb9" (OuterVolumeSpecName: "kube-api-access-99bb9") pod "b7a3ac30-3925-493e-9e13-0f1a9a7e49c2" (UID: "b7a3ac30-3925-493e-9e13-0f1a9a7e49c2"). InnerVolumeSpecName "kube-api-access-99bb9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:47:16 crc kubenswrapper[4821]: I1125 10:47:16.846317 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xqr4f\" (UniqueName: \"kubernetes.io/projected/68a5afc7-5e47-4bed-a6c2-8ff5b6b0236c-kube-api-access-xqr4f\") pod \"68a5afc7-5e47-4bed-a6c2-8ff5b6b0236c\" (UID: \"68a5afc7-5e47-4bed-a6c2-8ff5b6b0236c\") " Nov 25 10:47:16 crc kubenswrapper[4821]: I1125 10:47:16.846513 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/68a5afc7-5e47-4bed-a6c2-8ff5b6b0236c-dns-svc\") pod \"68a5afc7-5e47-4bed-a6c2-8ff5b6b0236c\" (UID: \"68a5afc7-5e47-4bed-a6c2-8ff5b6b0236c\") " Nov 25 10:47:16 crc kubenswrapper[4821]: I1125 10:47:16.846603 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68a5afc7-5e47-4bed-a6c2-8ff5b6b0236c-config\") pod \"68a5afc7-5e47-4bed-a6c2-8ff5b6b0236c\" (UID: \"68a5afc7-5e47-4bed-a6c2-8ff5b6b0236c\") " Nov 25 10:47:16 crc kubenswrapper[4821]: I1125 10:47:16.846982 4821 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7a3ac30-3925-493e-9e13-0f1a9a7e49c2-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:16 crc kubenswrapper[4821]: I1125 10:47:16.847013 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-99bb9\" (UniqueName: \"kubernetes.io/projected/b7a3ac30-3925-493e-9e13-0f1a9a7e49c2-kube-api-access-99bb9\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:16 crc kubenswrapper[4821]: I1125 10:47:16.847282 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68a5afc7-5e47-4bed-a6c2-8ff5b6b0236c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "68a5afc7-5e47-4bed-a6c2-8ff5b6b0236c" (UID: "68a5afc7-5e47-4bed-a6c2-8ff5b6b0236c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:47:16 crc kubenswrapper[4821]: I1125 10:47:16.847437 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68a5afc7-5e47-4bed-a6c2-8ff5b6b0236c-config" (OuterVolumeSpecName: "config") pod "68a5afc7-5e47-4bed-a6c2-8ff5b6b0236c" (UID: "68a5afc7-5e47-4bed-a6c2-8ff5b6b0236c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:47:16 crc kubenswrapper[4821]: I1125 10:47:16.850153 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68a5afc7-5e47-4bed-a6c2-8ff5b6b0236c-kube-api-access-xqr4f" (OuterVolumeSpecName: "kube-api-access-xqr4f") pod "68a5afc7-5e47-4bed-a6c2-8ff5b6b0236c" (UID: "68a5afc7-5e47-4bed-a6c2-8ff5b6b0236c"). InnerVolumeSpecName "kube-api-access-xqr4f". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:47:16 crc kubenswrapper[4821]: I1125 10:47:16.951902 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xqr4f\" (UniqueName: \"kubernetes.io/projected/68a5afc7-5e47-4bed-a6c2-8ff5b6b0236c-kube-api-access-xqr4f\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:16 crc kubenswrapper[4821]: I1125 10:47:16.951946 4821 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/68a5afc7-5e47-4bed-a6c2-8ff5b6b0236c-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:16 crc kubenswrapper[4821]: I1125 10:47:16.951956 4821 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68a5afc7-5e47-4bed-a6c2-8ff5b6b0236c-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:17 crc kubenswrapper[4821]: I1125 10:47:17.049430 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 25 10:47:17 crc kubenswrapper[4821]: I1125 10:47:17.060546 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ltxc6"] Nov 25 10:47:17 crc kubenswrapper[4821]: I1125 10:47:17.143451 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 25 10:47:17 crc kubenswrapper[4821]: I1125 10:47:17.172703 4821 generic.go:334] "Generic (PLEG): container finished" podID="47b41fd0-3e3d-4cf7-b8c9-1441ae9e2014" containerID="b92f28a2accb85c00054ba7fcbfffaa462f22eb53623810f0e12676c3c13ca06" exitCode=0 Nov 25 10:47:17 crc kubenswrapper[4821]: I1125 10:47:17.172775 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-gn85r" event={"ID":"47b41fd0-3e3d-4cf7-b8c9-1441ae9e2014","Type":"ContainerDied","Data":"b92f28a2accb85c00054ba7fcbfffaa462f22eb53623810f0e12676c3c13ca06"} Nov 25 10:47:17 crc kubenswrapper[4821]: I1125 10:47:17.184446 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"9aed573a-ad60-479a-ba27-8666d7dc4d9e","Type":"ContainerStarted","Data":"354e33bfb18ac583a14798f04475a2021fbea33d33965d910b98968996e27e58"} Nov 25 10:47:17 crc kubenswrapper[4821]: I1125 10:47:17.186331 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"e2b8639b-42d3-4de6-8d08-d5e2aac42391","Type":"ContainerStarted","Data":"cd26802fbbd6b9df5a3b0dd72b2f0570f1d220bdc2538ea6a9db8e716dec4df0"} Nov 25 10:47:17 crc kubenswrapper[4821]: I1125 10:47:17.198827 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"41c8c049-b466-4c19-944e-50a639491190","Type":"ContainerStarted","Data":"c7246f53e5c2fe708e2c0bb8e044516e59691a2b58cf651438d44f4eaa049ffd"} Nov 25 10:47:17 crc kubenswrapper[4821]: I1125 10:47:17.200928 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-srdjw" Nov 25 10:47:17 crc kubenswrapper[4821]: I1125 10:47:17.200923 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-srdjw" event={"ID":"68a5afc7-5e47-4bed-a6c2-8ff5b6b0236c","Type":"ContainerDied","Data":"2c793cc86cb38df469836329e3d4d5652b95d368e90ae7040096dcd72bc0fe99"} Nov 25 10:47:17 crc kubenswrapper[4821]: I1125 10:47:17.204617 4821 generic.go:334] "Generic (PLEG): container finished" podID="f289352e-7670-4503-b6bf-ccdaa45a7c36" containerID="d0b34e2120b350fb56b94467176b871e29f1dd028a34ba33b0362146aef3e828" exitCode=0 Nov 25 10:47:17 crc kubenswrapper[4821]: I1125 10:47:17.204715 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-h5qkz" event={"ID":"f289352e-7670-4503-b6bf-ccdaa45a7c36","Type":"ContainerDied","Data":"d0b34e2120b350fb56b94467176b871e29f1dd028a34ba33b0362146aef3e828"} Nov 25 10:47:17 crc kubenswrapper[4821]: I1125 10:47:17.208946 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ltxc6" event={"ID":"47eaf58d-190b-4e75-9aac-11d7e51df259","Type":"ContainerStarted","Data":"2071cdb181e7629be63277faccb781f144470b960275f186fddf815658ca1d21"} Nov 25 10:47:17 crc kubenswrapper[4821]: I1125 10:47:17.210843 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"2bbb754a-fff3-4ba2-ab75-68e48b626ab0","Type":"ContainerStarted","Data":"e736050aac0d58ca110e71f33ba6e70d441ed7f3a727bd519a0b4115de93d021"} Nov 25 10:47:17 crc kubenswrapper[4821]: I1125 10:47:17.212068 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-shhlw" event={"ID":"b7a3ac30-3925-493e-9e13-0f1a9a7e49c2","Type":"ContainerDied","Data":"23de78eedabd01795c0b200e44b5388d4d4d3a58f73bd2db657ab9e800ea9af6"} Nov 25 10:47:17 crc kubenswrapper[4821]: I1125 10:47:17.212131 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-shhlw" Nov 25 10:47:17 crc kubenswrapper[4821]: I1125 10:47:17.215223 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"8b9e8e7c-b055-47d3-874a-227d5f02432a","Type":"ContainerStarted","Data":"28e00b76e5fdebe99e428811e7b6cd0134dada0ca6809fabde14781538e9049e"} Nov 25 10:47:17 crc kubenswrapper[4821]: I1125 10:47:17.292794 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-srdjw"] Nov 25 10:47:17 crc kubenswrapper[4821]: I1125 10:47:17.315589 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-srdjw"] Nov 25 10:47:17 crc kubenswrapper[4821]: I1125 10:47:17.331424 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-shhlw"] Nov 25 10:47:17 crc kubenswrapper[4821]: I1125 10:47:17.337059 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-shhlw"] Nov 25 10:47:17 crc kubenswrapper[4821]: I1125 10:47:17.689128 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-6mbr2"] Nov 25 10:47:18 crc kubenswrapper[4821]: I1125 10:47:18.130244 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68a5afc7-5e47-4bed-a6c2-8ff5b6b0236c" path="/var/lib/kubelet/pods/68a5afc7-5e47-4bed-a6c2-8ff5b6b0236c/volumes" Nov 25 10:47:18 crc kubenswrapper[4821]: I1125 10:47:18.130694 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7a3ac30-3925-493e-9e13-0f1a9a7e49c2" path="/var/lib/kubelet/pods/b7a3ac30-3925-493e-9e13-0f1a9a7e49c2/volumes" Nov 25 10:47:18 crc kubenswrapper[4821]: I1125 10:47:18.173567 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 25 10:47:18 crc kubenswrapper[4821]: I1125 10:47:18.225662 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-gn85r" event={"ID":"47b41fd0-3e3d-4cf7-b8c9-1441ae9e2014","Type":"ContainerStarted","Data":"8b6f0257d2bb41336663bbc498d363cb185610efcc20e807f10c16dcf0916f7c"} Nov 25 10:47:18 crc kubenswrapper[4821]: I1125 10:47:18.226311 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5ccc8479f9-gn85r" Nov 25 10:47:18 crc kubenswrapper[4821]: I1125 10:47:18.248968 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5ccc8479f9-gn85r" podStartSLOduration=6.625775405 podStartE2EDuration="19.248948079s" podCreationTimestamp="2025-11-25 10:46:59 +0000 UTC" firstStartedPulling="2025-11-25 10:47:03.610724773 +0000 UTC m=+894.147044610" lastFinishedPulling="2025-11-25 10:47:16.233897437 +0000 UTC m=+906.770217284" observedRunningTime="2025-11-25 10:47:18.241282665 +0000 UTC m=+908.777602532" watchObservedRunningTime="2025-11-25 10:47:18.248948079 +0000 UTC m=+908.785267926" Nov 25 10:47:19 crc kubenswrapper[4821]: I1125 10:47:19.233149 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-6mbr2" event={"ID":"72b307b1-f36e-424f-91ef-f7402b6d7535","Type":"ContainerStarted","Data":"581de49ed794fdf613eb1d502831c4c476aaf86403c2cb4b8541e47fd2075fdf"} Nov 25 10:47:19 crc kubenswrapper[4821]: I1125 10:47:19.234310 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"30500d7b-4e4b-4839-8a9d-0b05a0ebd1fb","Type":"ContainerStarted","Data":"0dcaf6032cd7e840375d48f7645f9a08c4aa296d67cddc7105929a678433200a"} Nov 25 10:47:23 crc kubenswrapper[4821]: I1125 10:47:23.270831 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-h5qkz" event={"ID":"f289352e-7670-4503-b6bf-ccdaa45a7c36","Type":"ContainerStarted","Data":"238bc5cdcac84072257aa4c9c41a4c1f1dbcc245eb1007ce0b9326f4683bab66"} Nov 25 10:47:23 crc kubenswrapper[4821]: I1125 10:47:23.271123 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-h5qkz" Nov 25 10:47:23 crc kubenswrapper[4821]: I1125 10:47:23.290606 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-h5qkz" podStartSLOduration=16.106727018 podStartE2EDuration="24.290589533s" podCreationTimestamp="2025-11-25 10:46:59 +0000 UTC" firstStartedPulling="2025-11-25 10:47:08.044394598 +0000 UTC m=+898.580714445" lastFinishedPulling="2025-11-25 10:47:16.228257113 +0000 UTC m=+906.764576960" observedRunningTime="2025-11-25 10:47:23.289986998 +0000 UTC m=+913.826306845" watchObservedRunningTime="2025-11-25 10:47:23.290589533 +0000 UTC m=+913.826909380" Nov 25 10:47:25 crc kubenswrapper[4821]: I1125 10:47:25.064309 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5ccc8479f9-gn85r" Nov 25 10:47:27 crc kubenswrapper[4821]: I1125 10:47:27.319024 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"41c8c049-b466-4c19-944e-50a639491190","Type":"ContainerStarted","Data":"24c78e44cfcdcef44fa30cedd0ade1994e19ddd2dc1a243cedbe1f8b201d490c"} Nov 25 10:47:27 crc kubenswrapper[4821]: I1125 10:47:27.326049 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"8b9e8e7c-b055-47d3-874a-227d5f02432a","Type":"ContainerStarted","Data":"c978e46612c58a34208701a08e43cba12e5f1a4d3bd52674068c5b271f3afba1"} Nov 25 10:47:27 crc kubenswrapper[4821]: I1125 10:47:27.326141 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Nov 25 10:47:27 crc kubenswrapper[4821]: I1125 10:47:27.328151 4821 generic.go:334] "Generic (PLEG): container finished" podID="72b307b1-f36e-424f-91ef-f7402b6d7535" containerID="6268bdf475806e4e601ccb029ab0f05130bebd55ab6b4900684341ae15a61d9b" exitCode=0 Nov 25 10:47:27 crc kubenswrapper[4821]: I1125 10:47:27.328297 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-6mbr2" event={"ID":"72b307b1-f36e-424f-91ef-f7402b6d7535","Type":"ContainerDied","Data":"6268bdf475806e4e601ccb029ab0f05130bebd55ab6b4900684341ae15a61d9b"} Nov 25 10:47:27 crc kubenswrapper[4821]: I1125 10:47:27.330990 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"9aed573a-ad60-479a-ba27-8666d7dc4d9e","Type":"ContainerStarted","Data":"8b925d6dbe4762f834016241b8231c45e969f5db03a96a8d2be3612011fe9d58"} Nov 25 10:47:27 crc kubenswrapper[4821]: I1125 10:47:27.333247 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"ecb5eaf2-4178-43ec-bc21-0e05de8fc237","Type":"ContainerStarted","Data":"a7cb2d363b40398cf4706d4d35fec6aa8a899ece43ba8cda84f01f6205335a58"} Nov 25 10:47:27 crc kubenswrapper[4821]: I1125 10:47:27.335972 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ltxc6" event={"ID":"47eaf58d-190b-4e75-9aac-11d7e51df259","Type":"ContainerStarted","Data":"ddd0571cb18ff2ee52786ba2b2ef6eb7f6a1ebafa0cafc48a0ee7c0cc1cbbde8"} Nov 25 10:47:27 crc kubenswrapper[4821]: I1125 10:47:27.336086 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ltxc6" Nov 25 10:47:27 crc kubenswrapper[4821]: I1125 10:47:27.342312 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"2bbb754a-fff3-4ba2-ab75-68e48b626ab0","Type":"ContainerStarted","Data":"ef45f64d359b5165b93856160cad982dc3eb70d24d3b0aed64df4b49a699da2b"} Nov 25 10:47:27 crc kubenswrapper[4821]: I1125 10:47:27.347099 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"e2b8639b-42d3-4de6-8d08-d5e2aac42391","Type":"ContainerStarted","Data":"9b16284d6f14b5581a449f355c6acf99510233d74e0787abff5bdfca3e4d45dd"} Nov 25 10:47:27 crc kubenswrapper[4821]: I1125 10:47:27.347319 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 25 10:47:27 crc kubenswrapper[4821]: I1125 10:47:27.350934 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"330a479e-3d76-4de3-bc84-1d7c2d65d265","Type":"ContainerStarted","Data":"658c4ee0f18233ace6320ae015dedf80011737bf96f761266c9995e11d4f5dc7"} Nov 25 10:47:27 crc kubenswrapper[4821]: I1125 10:47:27.353975 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"30500d7b-4e4b-4839-8a9d-0b05a0ebd1fb","Type":"ContainerStarted","Data":"5dbe8cfb08e6b7f7ef2fe3bebf8bb6bfe719a848499525ae4a4400716558c1e3"} Nov 25 10:47:27 crc kubenswrapper[4821]: I1125 10:47:27.364192 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ltxc6" podStartSLOduration=8.449170464 podStartE2EDuration="17.364153558s" podCreationTimestamp="2025-11-25 10:47:10 +0000 UTC" firstStartedPulling="2025-11-25 10:47:17.072623785 +0000 UTC m=+907.608943632" lastFinishedPulling="2025-11-25 10:47:25.987606879 +0000 UTC m=+916.523926726" observedRunningTime="2025-11-25 10:47:27.354634677 +0000 UTC m=+917.890954524" watchObservedRunningTime="2025-11-25 10:47:27.364153558 +0000 UTC m=+917.900473405" Nov 25 10:47:27 crc kubenswrapper[4821]: I1125 10:47:27.418823 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=14.553737157 podStartE2EDuration="23.418800934s" podCreationTimestamp="2025-11-25 10:47:04 +0000 UTC" firstStartedPulling="2025-11-25 10:47:17.073305653 +0000 UTC m=+907.609625510" lastFinishedPulling="2025-11-25 10:47:25.93836944 +0000 UTC m=+916.474689287" observedRunningTime="2025-11-25 10:47:27.408840282 +0000 UTC m=+917.945160129" watchObservedRunningTime="2025-11-25 10:47:27.418800934 +0000 UTC m=+917.955120781" Nov 25 10:47:27 crc kubenswrapper[4821]: I1125 10:47:27.464948 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=12.193337572 podStartE2EDuration="21.464926234s" podCreationTimestamp="2025-11-25 10:47:06 +0000 UTC" firstStartedPulling="2025-11-25 10:47:16.734501725 +0000 UTC m=+907.270821562" lastFinishedPulling="2025-11-25 10:47:26.006090387 +0000 UTC m=+916.542410224" observedRunningTime="2025-11-25 10:47:27.459925288 +0000 UTC m=+917.996245135" watchObservedRunningTime="2025-11-25 10:47:27.464926234 +0000 UTC m=+918.001246081" Nov 25 10:47:28 crc kubenswrapper[4821]: I1125 10:47:28.367185 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-6mbr2" event={"ID":"72b307b1-f36e-424f-91ef-f7402b6d7535","Type":"ContainerStarted","Data":"bcf7d1b64fd6986a1828a67cd65ca31b0c316a3853d9da1c31ba3fc7832cca3c"} Nov 25 10:47:28 crc kubenswrapper[4821]: I1125 10:47:28.367544 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-6mbr2" event={"ID":"72b307b1-f36e-424f-91ef-f7402b6d7535","Type":"ContainerStarted","Data":"8ffac985c8b5a70fb15ed9fc49777dbd5c3f91dec02aaab3ca5438a964d0cc04"} Nov 25 10:47:28 crc kubenswrapper[4821]: I1125 10:47:28.393877 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-6mbr2" podStartSLOduration=11.224234465 podStartE2EDuration="18.393832999s" podCreationTimestamp="2025-11-25 10:47:10 +0000 UTC" firstStartedPulling="2025-11-25 10:47:18.790889137 +0000 UTC m=+909.327208984" lastFinishedPulling="2025-11-25 10:47:25.960487671 +0000 UTC m=+916.496807518" observedRunningTime="2025-11-25 10:47:28.391433278 +0000 UTC m=+918.927753135" watchObservedRunningTime="2025-11-25 10:47:28.393832999 +0000 UTC m=+918.930152866" Nov 25 10:47:29 crc kubenswrapper[4821]: I1125 10:47:29.374747 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-6mbr2" Nov 25 10:47:29 crc kubenswrapper[4821]: I1125 10:47:29.375101 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-6mbr2" Nov 25 10:47:30 crc kubenswrapper[4821]: I1125 10:47:30.342785 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57d769cc4f-h5qkz" Nov 25 10:47:30 crc kubenswrapper[4821]: I1125 10:47:30.388074 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"9aed573a-ad60-479a-ba27-8666d7dc4d9e","Type":"ContainerStarted","Data":"9408fb23e2dfa62b025aa22a4e192d419b799c6238a68fa27efce7b0f3f2fd1d"} Nov 25 10:47:30 crc kubenswrapper[4821]: I1125 10:47:30.391834 4821 generic.go:334] "Generic (PLEG): container finished" podID="ecb5eaf2-4178-43ec-bc21-0e05de8fc237" containerID="a7cb2d363b40398cf4706d4d35fec6aa8a899ece43ba8cda84f01f6205335a58" exitCode=0 Nov 25 10:47:30 crc kubenswrapper[4821]: I1125 10:47:30.391955 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"ecb5eaf2-4178-43ec-bc21-0e05de8fc237","Type":"ContainerDied","Data":"a7cb2d363b40398cf4706d4d35fec6aa8a899ece43ba8cda84f01f6205335a58"} Nov 25 10:47:30 crc kubenswrapper[4821]: I1125 10:47:30.395022 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"30500d7b-4e4b-4839-8a9d-0b05a0ebd1fb","Type":"ContainerStarted","Data":"ad840a700e73cf8f2cd76b620341af9f6df301c8f67d168f639f6ea16723677e"} Nov 25 10:47:30 crc kubenswrapper[4821]: I1125 10:47:30.399301 4821 generic.go:334] "Generic (PLEG): container finished" podID="41c8c049-b466-4c19-944e-50a639491190" containerID="24c78e44cfcdcef44fa30cedd0ade1994e19ddd2dc1a243cedbe1f8b201d490c" exitCode=0 Nov 25 10:47:30 crc kubenswrapper[4821]: I1125 10:47:30.399707 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"41c8c049-b466-4c19-944e-50a639491190","Type":"ContainerDied","Data":"24c78e44cfcdcef44fa30cedd0ade1994e19ddd2dc1a243cedbe1f8b201d490c"} Nov 25 10:47:30 crc kubenswrapper[4821]: I1125 10:47:30.424352 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=6.097652372 podStartE2EDuration="18.424332607s" podCreationTimestamp="2025-11-25 10:47:12 +0000 UTC" firstStartedPulling="2025-11-25 10:47:17.157588735 +0000 UTC m=+907.693908582" lastFinishedPulling="2025-11-25 10:47:29.48426898 +0000 UTC m=+920.020588817" observedRunningTime="2025-11-25 10:47:30.423527106 +0000 UTC m=+920.959846973" watchObservedRunningTime="2025-11-25 10:47:30.424332607 +0000 UTC m=+920.960652454" Nov 25 10:47:30 crc kubenswrapper[4821]: I1125 10:47:30.479342 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-gn85r"] Nov 25 10:47:30 crc kubenswrapper[4821]: I1125 10:47:30.479604 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5ccc8479f9-gn85r" podUID="47b41fd0-3e3d-4cf7-b8c9-1441ae9e2014" containerName="dnsmasq-dns" containerID="cri-o://8b6f0257d2bb41336663bbc498d363cb185610efcc20e807f10c16dcf0916f7c" gracePeriod=10 Nov 25 10:47:30 crc kubenswrapper[4821]: I1125 10:47:30.491795 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=10.766841564 podStartE2EDuration="21.490132436s" podCreationTimestamp="2025-11-25 10:47:09 +0000 UTC" firstStartedPulling="2025-11-25 10:47:18.792349264 +0000 UTC m=+909.328669111" lastFinishedPulling="2025-11-25 10:47:29.515640136 +0000 UTC m=+920.051959983" observedRunningTime="2025-11-25 10:47:30.456412121 +0000 UTC m=+920.992732008" watchObservedRunningTime="2025-11-25 10:47:30.490132436 +0000 UTC m=+921.026452283" Nov 25 10:47:30 crc kubenswrapper[4821]: I1125 10:47:30.881090 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-gn85r" Nov 25 10:47:30 crc kubenswrapper[4821]: I1125 10:47:30.907304 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wgq8x\" (UniqueName: \"kubernetes.io/projected/47b41fd0-3e3d-4cf7-b8c9-1441ae9e2014-kube-api-access-wgq8x\") pod \"47b41fd0-3e3d-4cf7-b8c9-1441ae9e2014\" (UID: \"47b41fd0-3e3d-4cf7-b8c9-1441ae9e2014\") " Nov 25 10:47:30 crc kubenswrapper[4821]: I1125 10:47:30.907398 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47b41fd0-3e3d-4cf7-b8c9-1441ae9e2014-config\") pod \"47b41fd0-3e3d-4cf7-b8c9-1441ae9e2014\" (UID: \"47b41fd0-3e3d-4cf7-b8c9-1441ae9e2014\") " Nov 25 10:47:30 crc kubenswrapper[4821]: I1125 10:47:30.907442 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/47b41fd0-3e3d-4cf7-b8c9-1441ae9e2014-dns-svc\") pod \"47b41fd0-3e3d-4cf7-b8c9-1441ae9e2014\" (UID: \"47b41fd0-3e3d-4cf7-b8c9-1441ae9e2014\") " Nov 25 10:47:30 crc kubenswrapper[4821]: I1125 10:47:30.912617 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47b41fd0-3e3d-4cf7-b8c9-1441ae9e2014-kube-api-access-wgq8x" (OuterVolumeSpecName: "kube-api-access-wgq8x") pod "47b41fd0-3e3d-4cf7-b8c9-1441ae9e2014" (UID: "47b41fd0-3e3d-4cf7-b8c9-1441ae9e2014"). InnerVolumeSpecName "kube-api-access-wgq8x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:47:30 crc kubenswrapper[4821]: I1125 10:47:30.951125 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/47b41fd0-3e3d-4cf7-b8c9-1441ae9e2014-config" (OuterVolumeSpecName: "config") pod "47b41fd0-3e3d-4cf7-b8c9-1441ae9e2014" (UID: "47b41fd0-3e3d-4cf7-b8c9-1441ae9e2014"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:47:30 crc kubenswrapper[4821]: I1125 10:47:30.965888 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/47b41fd0-3e3d-4cf7-b8c9-1441ae9e2014-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "47b41fd0-3e3d-4cf7-b8c9-1441ae9e2014" (UID: "47b41fd0-3e3d-4cf7-b8c9-1441ae9e2014"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:47:31 crc kubenswrapper[4821]: I1125 10:47:31.008775 4821 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47b41fd0-3e3d-4cf7-b8c9-1441ae9e2014-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:31 crc kubenswrapper[4821]: I1125 10:47:31.008975 4821 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/47b41fd0-3e3d-4cf7-b8c9-1441ae9e2014-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:31 crc kubenswrapper[4821]: I1125 10:47:31.009041 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wgq8x\" (UniqueName: \"kubernetes.io/projected/47b41fd0-3e3d-4cf7-b8c9-1441ae9e2014-kube-api-access-wgq8x\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:31 crc kubenswrapper[4821]: I1125 10:47:31.273788 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Nov 25 10:47:31 crc kubenswrapper[4821]: I1125 10:47:31.408636 4821 generic.go:334] "Generic (PLEG): container finished" podID="47b41fd0-3e3d-4cf7-b8c9-1441ae9e2014" containerID="8b6f0257d2bb41336663bbc498d363cb185610efcc20e807f10c16dcf0916f7c" exitCode=0 Nov 25 10:47:31 crc kubenswrapper[4821]: I1125 10:47:31.408677 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-gn85r" Nov 25 10:47:31 crc kubenswrapper[4821]: I1125 10:47:31.408724 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-gn85r" event={"ID":"47b41fd0-3e3d-4cf7-b8c9-1441ae9e2014","Type":"ContainerDied","Data":"8b6f0257d2bb41336663bbc498d363cb185610efcc20e807f10c16dcf0916f7c"} Nov 25 10:47:31 crc kubenswrapper[4821]: I1125 10:47:31.408811 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-gn85r" event={"ID":"47b41fd0-3e3d-4cf7-b8c9-1441ae9e2014","Type":"ContainerDied","Data":"4f21960c9d23afc3242da1057ea86fa8b849eba22c0b36e697e17207a6317d81"} Nov 25 10:47:31 crc kubenswrapper[4821]: I1125 10:47:31.408831 4821 scope.go:117] "RemoveContainer" containerID="8b6f0257d2bb41336663bbc498d363cb185610efcc20e807f10c16dcf0916f7c" Nov 25 10:47:31 crc kubenswrapper[4821]: I1125 10:47:31.411061 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"ecb5eaf2-4178-43ec-bc21-0e05de8fc237","Type":"ContainerStarted","Data":"933e3f60ee7f6a44130f5e8edf0072b0c7cbd9eea388f50d7a78afbbf35569fb"} Nov 25 10:47:31 crc kubenswrapper[4821]: I1125 10:47:31.413438 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"41c8c049-b466-4c19-944e-50a639491190","Type":"ContainerStarted","Data":"b223332f6f17a2cea3c0b1ffe433d7f5e54fef518b9469adcda95d11008d9739"} Nov 25 10:47:31 crc kubenswrapper[4821]: I1125 10:47:31.440559 4821 scope.go:117] "RemoveContainer" containerID="b92f28a2accb85c00054ba7fcbfffaa462f22eb53623810f0e12676c3c13ca06" Nov 25 10:47:31 crc kubenswrapper[4821]: I1125 10:47:31.441089 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=23.957542992 podStartE2EDuration="30.441066788s" podCreationTimestamp="2025-11-25 10:47:01 +0000 UTC" firstStartedPulling="2025-11-25 10:47:16.016819791 +0000 UTC m=+906.553139638" lastFinishedPulling="2025-11-25 10:47:22.500343597 +0000 UTC m=+913.036663434" observedRunningTime="2025-11-25 10:47:31.43044369 +0000 UTC m=+921.966763547" watchObservedRunningTime="2025-11-25 10:47:31.441066788 +0000 UTC m=+921.977386655" Nov 25 10:47:31 crc kubenswrapper[4821]: I1125 10:47:31.456935 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=22.542268951 podStartE2EDuration="28.45691624s" podCreationTimestamp="2025-11-25 10:47:03 +0000 UTC" firstStartedPulling="2025-11-25 10:47:16.728607043 +0000 UTC m=+907.264926890" lastFinishedPulling="2025-11-25 10:47:22.643254332 +0000 UTC m=+913.179574179" observedRunningTime="2025-11-25 10:47:31.452425367 +0000 UTC m=+921.988745224" watchObservedRunningTime="2025-11-25 10:47:31.45691624 +0000 UTC m=+921.993236087" Nov 25 10:47:31 crc kubenswrapper[4821]: I1125 10:47:31.473794 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-gn85r"] Nov 25 10:47:31 crc kubenswrapper[4821]: I1125 10:47:31.476996 4821 scope.go:117] "RemoveContainer" containerID="8b6f0257d2bb41336663bbc498d363cb185610efcc20e807f10c16dcf0916f7c" Nov 25 10:47:31 crc kubenswrapper[4821]: E1125 10:47:31.477524 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b6f0257d2bb41336663bbc498d363cb185610efcc20e807f10c16dcf0916f7c\": container with ID starting with 8b6f0257d2bb41336663bbc498d363cb185610efcc20e807f10c16dcf0916f7c not found: ID does not exist" containerID="8b6f0257d2bb41336663bbc498d363cb185610efcc20e807f10c16dcf0916f7c" Nov 25 10:47:31 crc kubenswrapper[4821]: I1125 10:47:31.477574 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b6f0257d2bb41336663bbc498d363cb185610efcc20e807f10c16dcf0916f7c"} err="failed to get container status \"8b6f0257d2bb41336663bbc498d363cb185610efcc20e807f10c16dcf0916f7c\": rpc error: code = NotFound desc = could not find container \"8b6f0257d2bb41336663bbc498d363cb185610efcc20e807f10c16dcf0916f7c\": container with ID starting with 8b6f0257d2bb41336663bbc498d363cb185610efcc20e807f10c16dcf0916f7c not found: ID does not exist" Nov 25 10:47:31 crc kubenswrapper[4821]: I1125 10:47:31.477604 4821 scope.go:117] "RemoveContainer" containerID="b92f28a2accb85c00054ba7fcbfffaa462f22eb53623810f0e12676c3c13ca06" Nov 25 10:47:31 crc kubenswrapper[4821]: E1125 10:47:31.478011 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b92f28a2accb85c00054ba7fcbfffaa462f22eb53623810f0e12676c3c13ca06\": container with ID starting with b92f28a2accb85c00054ba7fcbfffaa462f22eb53623810f0e12676c3c13ca06 not found: ID does not exist" containerID="b92f28a2accb85c00054ba7fcbfffaa462f22eb53623810f0e12676c3c13ca06" Nov 25 10:47:31 crc kubenswrapper[4821]: I1125 10:47:31.478039 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b92f28a2accb85c00054ba7fcbfffaa462f22eb53623810f0e12676c3c13ca06"} err="failed to get container status \"b92f28a2accb85c00054ba7fcbfffaa462f22eb53623810f0e12676c3c13ca06\": rpc error: code = NotFound desc = could not find container \"b92f28a2accb85c00054ba7fcbfffaa462f22eb53623810f0e12676c3c13ca06\": container with ID starting with b92f28a2accb85c00054ba7fcbfffaa462f22eb53623810f0e12676c3c13ca06 not found: ID does not exist" Nov 25 10:47:31 crc kubenswrapper[4821]: I1125 10:47:31.479623 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-gn85r"] Nov 25 10:47:31 crc kubenswrapper[4821]: E1125 10:47:31.499752 4821 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod47b41fd0_3e3d_4cf7_b8c9_1441ae9e2014.slice/crio-4f21960c9d23afc3242da1057ea86fa8b849eba22c0b36e697e17207a6317d81\": RecentStats: unable to find data in memory cache]" Nov 25 10:47:32 crc kubenswrapper[4821]: I1125 10:47:32.132219 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47b41fd0-3e3d-4cf7-b8c9-1441ae9e2014" path="/var/lib/kubelet/pods/47b41fd0-3e3d-4cf7-b8c9-1441ae9e2014/volumes" Nov 25 10:47:32 crc kubenswrapper[4821]: I1125 10:47:32.274364 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Nov 25 10:47:32 crc kubenswrapper[4821]: I1125 10:47:32.312015 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Nov 25 10:47:32 crc kubenswrapper[4821]: I1125 10:47:32.462981 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Nov 25 10:47:32 crc kubenswrapper[4821]: I1125 10:47:32.754867 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-mr78v"] Nov 25 10:47:32 crc kubenswrapper[4821]: E1125 10:47:32.755208 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47b41fd0-3e3d-4cf7-b8c9-1441ae9e2014" containerName="dnsmasq-dns" Nov 25 10:47:32 crc kubenswrapper[4821]: I1125 10:47:32.755219 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="47b41fd0-3e3d-4cf7-b8c9-1441ae9e2014" containerName="dnsmasq-dns" Nov 25 10:47:32 crc kubenswrapper[4821]: E1125 10:47:32.755234 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47b41fd0-3e3d-4cf7-b8c9-1441ae9e2014" containerName="init" Nov 25 10:47:32 crc kubenswrapper[4821]: I1125 10:47:32.755240 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="47b41fd0-3e3d-4cf7-b8c9-1441ae9e2014" containerName="init" Nov 25 10:47:32 crc kubenswrapper[4821]: I1125 10:47:32.755384 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="47b41fd0-3e3d-4cf7-b8c9-1441ae9e2014" containerName="dnsmasq-dns" Nov 25 10:47:32 crc kubenswrapper[4821]: I1125 10:47:32.756134 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-mr78v" Nov 25 10:47:32 crc kubenswrapper[4821]: I1125 10:47:32.758240 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Nov 25 10:47:32 crc kubenswrapper[4821]: I1125 10:47:32.774852 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-mr78v"] Nov 25 10:47:32 crc kubenswrapper[4821]: I1125 10:47:32.837848 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90d5b9f5-56b8-4904-b66b-3b28fe35c2e0-config\") pod \"dnsmasq-dns-5bf47b49b7-mr78v\" (UID: \"90d5b9f5-56b8-4904-b66b-3b28fe35c2e0\") " pod="openstack/dnsmasq-dns-5bf47b49b7-mr78v" Nov 25 10:47:32 crc kubenswrapper[4821]: I1125 10:47:32.838694 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/90d5b9f5-56b8-4904-b66b-3b28fe35c2e0-ovsdbserver-nb\") pod \"dnsmasq-dns-5bf47b49b7-mr78v\" (UID: \"90d5b9f5-56b8-4904-b66b-3b28fe35c2e0\") " pod="openstack/dnsmasq-dns-5bf47b49b7-mr78v" Nov 25 10:47:32 crc kubenswrapper[4821]: I1125 10:47:32.838783 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/90d5b9f5-56b8-4904-b66b-3b28fe35c2e0-dns-svc\") pod \"dnsmasq-dns-5bf47b49b7-mr78v\" (UID: \"90d5b9f5-56b8-4904-b66b-3b28fe35c2e0\") " pod="openstack/dnsmasq-dns-5bf47b49b7-mr78v" Nov 25 10:47:32 crc kubenswrapper[4821]: I1125 10:47:32.838814 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b87hd\" (UniqueName: \"kubernetes.io/projected/90d5b9f5-56b8-4904-b66b-3b28fe35c2e0-kube-api-access-b87hd\") pod \"dnsmasq-dns-5bf47b49b7-mr78v\" (UID: \"90d5b9f5-56b8-4904-b66b-3b28fe35c2e0\") " pod="openstack/dnsmasq-dns-5bf47b49b7-mr78v" Nov 25 10:47:32 crc kubenswrapper[4821]: I1125 10:47:32.940188 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/90d5b9f5-56b8-4904-b66b-3b28fe35c2e0-ovsdbserver-nb\") pod \"dnsmasq-dns-5bf47b49b7-mr78v\" (UID: \"90d5b9f5-56b8-4904-b66b-3b28fe35c2e0\") " pod="openstack/dnsmasq-dns-5bf47b49b7-mr78v" Nov 25 10:47:32 crc kubenswrapper[4821]: I1125 10:47:32.940275 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/90d5b9f5-56b8-4904-b66b-3b28fe35c2e0-dns-svc\") pod \"dnsmasq-dns-5bf47b49b7-mr78v\" (UID: \"90d5b9f5-56b8-4904-b66b-3b28fe35c2e0\") " pod="openstack/dnsmasq-dns-5bf47b49b7-mr78v" Nov 25 10:47:32 crc kubenswrapper[4821]: I1125 10:47:32.940305 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b87hd\" (UniqueName: \"kubernetes.io/projected/90d5b9f5-56b8-4904-b66b-3b28fe35c2e0-kube-api-access-b87hd\") pod \"dnsmasq-dns-5bf47b49b7-mr78v\" (UID: \"90d5b9f5-56b8-4904-b66b-3b28fe35c2e0\") " pod="openstack/dnsmasq-dns-5bf47b49b7-mr78v" Nov 25 10:47:32 crc kubenswrapper[4821]: I1125 10:47:32.940382 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90d5b9f5-56b8-4904-b66b-3b28fe35c2e0-config\") pod \"dnsmasq-dns-5bf47b49b7-mr78v\" (UID: \"90d5b9f5-56b8-4904-b66b-3b28fe35c2e0\") " pod="openstack/dnsmasq-dns-5bf47b49b7-mr78v" Nov 25 10:47:32 crc kubenswrapper[4821]: I1125 10:47:32.941616 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90d5b9f5-56b8-4904-b66b-3b28fe35c2e0-config\") pod \"dnsmasq-dns-5bf47b49b7-mr78v\" (UID: \"90d5b9f5-56b8-4904-b66b-3b28fe35c2e0\") " pod="openstack/dnsmasq-dns-5bf47b49b7-mr78v" Nov 25 10:47:32 crc kubenswrapper[4821]: I1125 10:47:32.942474 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/90d5b9f5-56b8-4904-b66b-3b28fe35c2e0-dns-svc\") pod \"dnsmasq-dns-5bf47b49b7-mr78v\" (UID: \"90d5b9f5-56b8-4904-b66b-3b28fe35c2e0\") " pod="openstack/dnsmasq-dns-5bf47b49b7-mr78v" Nov 25 10:47:32 crc kubenswrapper[4821]: I1125 10:47:32.942510 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/90d5b9f5-56b8-4904-b66b-3b28fe35c2e0-ovsdbserver-nb\") pod \"dnsmasq-dns-5bf47b49b7-mr78v\" (UID: \"90d5b9f5-56b8-4904-b66b-3b28fe35c2e0\") " pod="openstack/dnsmasq-dns-5bf47b49b7-mr78v" Nov 25 10:47:32 crc kubenswrapper[4821]: I1125 10:47:32.947456 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-kkntf"] Nov 25 10:47:32 crc kubenswrapper[4821]: I1125 10:47:32.948582 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-kkntf" Nov 25 10:47:32 crc kubenswrapper[4821]: I1125 10:47:32.953570 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Nov 25 10:47:32 crc kubenswrapper[4821]: I1125 10:47:32.965576 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-kkntf"] Nov 25 10:47:32 crc kubenswrapper[4821]: I1125 10:47:32.969843 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b87hd\" (UniqueName: \"kubernetes.io/projected/90d5b9f5-56b8-4904-b66b-3b28fe35c2e0-kube-api-access-b87hd\") pod \"dnsmasq-dns-5bf47b49b7-mr78v\" (UID: \"90d5b9f5-56b8-4904-b66b-3b28fe35c2e0\") " pod="openstack/dnsmasq-dns-5bf47b49b7-mr78v" Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.041582 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/832f5849-e58a-407b-9f45-55c52cfbcd02-ovn-rundir\") pod \"ovn-controller-metrics-kkntf\" (UID: \"832f5849-e58a-407b-9f45-55c52cfbcd02\") " pod="openstack/ovn-controller-metrics-kkntf" Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.041721 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/832f5849-e58a-407b-9f45-55c52cfbcd02-config\") pod \"ovn-controller-metrics-kkntf\" (UID: \"832f5849-e58a-407b-9f45-55c52cfbcd02\") " pod="openstack/ovn-controller-metrics-kkntf" Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.041772 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/832f5849-e58a-407b-9f45-55c52cfbcd02-ovs-rundir\") pod \"ovn-controller-metrics-kkntf\" (UID: \"832f5849-e58a-407b-9f45-55c52cfbcd02\") " pod="openstack/ovn-controller-metrics-kkntf" Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.041807 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gpz5h\" (UniqueName: \"kubernetes.io/projected/832f5849-e58a-407b-9f45-55c52cfbcd02-kube-api-access-gpz5h\") pod \"ovn-controller-metrics-kkntf\" (UID: \"832f5849-e58a-407b-9f45-55c52cfbcd02\") " pod="openstack/ovn-controller-metrics-kkntf" Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.041986 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/832f5849-e58a-407b-9f45-55c52cfbcd02-combined-ca-bundle\") pod \"ovn-controller-metrics-kkntf\" (UID: \"832f5849-e58a-407b-9f45-55c52cfbcd02\") " pod="openstack/ovn-controller-metrics-kkntf" Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.042042 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/832f5849-e58a-407b-9f45-55c52cfbcd02-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-kkntf\" (UID: \"832f5849-e58a-407b-9f45-55c52cfbcd02\") " pod="openstack/ovn-controller-metrics-kkntf" Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.074587 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-mr78v" Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.114761 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.114812 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.118828 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.143816 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/832f5849-e58a-407b-9f45-55c52cfbcd02-ovn-rundir\") pod \"ovn-controller-metrics-kkntf\" (UID: \"832f5849-e58a-407b-9f45-55c52cfbcd02\") " pod="openstack/ovn-controller-metrics-kkntf" Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.143892 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/832f5849-e58a-407b-9f45-55c52cfbcd02-config\") pod \"ovn-controller-metrics-kkntf\" (UID: \"832f5849-e58a-407b-9f45-55c52cfbcd02\") " pod="openstack/ovn-controller-metrics-kkntf" Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.143933 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/832f5849-e58a-407b-9f45-55c52cfbcd02-ovs-rundir\") pod \"ovn-controller-metrics-kkntf\" (UID: \"832f5849-e58a-407b-9f45-55c52cfbcd02\") " pod="openstack/ovn-controller-metrics-kkntf" Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.143960 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gpz5h\" (UniqueName: \"kubernetes.io/projected/832f5849-e58a-407b-9f45-55c52cfbcd02-kube-api-access-gpz5h\") pod \"ovn-controller-metrics-kkntf\" (UID: \"832f5849-e58a-407b-9f45-55c52cfbcd02\") " pod="openstack/ovn-controller-metrics-kkntf" Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.143997 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/832f5849-e58a-407b-9f45-55c52cfbcd02-combined-ca-bundle\") pod \"ovn-controller-metrics-kkntf\" (UID: \"832f5849-e58a-407b-9f45-55c52cfbcd02\") " pod="openstack/ovn-controller-metrics-kkntf" Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.144014 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/832f5849-e58a-407b-9f45-55c52cfbcd02-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-kkntf\" (UID: \"832f5849-e58a-407b-9f45-55c52cfbcd02\") " pod="openstack/ovn-controller-metrics-kkntf" Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.144399 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/832f5849-e58a-407b-9f45-55c52cfbcd02-ovn-rundir\") pod \"ovn-controller-metrics-kkntf\" (UID: \"832f5849-e58a-407b-9f45-55c52cfbcd02\") " pod="openstack/ovn-controller-metrics-kkntf" Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.144476 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/832f5849-e58a-407b-9f45-55c52cfbcd02-ovs-rundir\") pod \"ovn-controller-metrics-kkntf\" (UID: \"832f5849-e58a-407b-9f45-55c52cfbcd02\") " pod="openstack/ovn-controller-metrics-kkntf" Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.145657 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/832f5849-e58a-407b-9f45-55c52cfbcd02-config\") pod \"ovn-controller-metrics-kkntf\" (UID: \"832f5849-e58a-407b-9f45-55c52cfbcd02\") " pod="openstack/ovn-controller-metrics-kkntf" Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.148426 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/832f5849-e58a-407b-9f45-55c52cfbcd02-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-kkntf\" (UID: \"832f5849-e58a-407b-9f45-55c52cfbcd02\") " pod="openstack/ovn-controller-metrics-kkntf" Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.151520 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/832f5849-e58a-407b-9f45-55c52cfbcd02-combined-ca-bundle\") pod \"ovn-controller-metrics-kkntf\" (UID: \"832f5849-e58a-407b-9f45-55c52cfbcd02\") " pod="openstack/ovn-controller-metrics-kkntf" Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.163591 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gpz5h\" (UniqueName: \"kubernetes.io/projected/832f5849-e58a-407b-9f45-55c52cfbcd02-kube-api-access-gpz5h\") pod \"ovn-controller-metrics-kkntf\" (UID: \"832f5849-e58a-407b-9f45-55c52cfbcd02\") " pod="openstack/ovn-controller-metrics-kkntf" Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.187154 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.267718 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-kkntf" Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.295260 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-mr78v"] Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.326731 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8554648995-l7fnk"] Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.332287 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-l7fnk" Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.336842 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.338809 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-l7fnk"] Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.431332 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.448951 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/140a7c0b-b837-43f7-a436-f13bcdd121b9-dns-svc\") pod \"dnsmasq-dns-8554648995-l7fnk\" (UID: \"140a7c0b-b837-43f7-a436-f13bcdd121b9\") " pod="openstack/dnsmasq-dns-8554648995-l7fnk" Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.449002 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gvbnm\" (UniqueName: \"kubernetes.io/projected/140a7c0b-b837-43f7-a436-f13bcdd121b9-kube-api-access-gvbnm\") pod \"dnsmasq-dns-8554648995-l7fnk\" (UID: \"140a7c0b-b837-43f7-a436-f13bcdd121b9\") " pod="openstack/dnsmasq-dns-8554648995-l7fnk" Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.449065 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/140a7c0b-b837-43f7-a436-f13bcdd121b9-config\") pod \"dnsmasq-dns-8554648995-l7fnk\" (UID: \"140a7c0b-b837-43f7-a436-f13bcdd121b9\") " pod="openstack/dnsmasq-dns-8554648995-l7fnk" Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.449186 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/140a7c0b-b837-43f7-a436-f13bcdd121b9-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-l7fnk\" (UID: \"140a7c0b-b837-43f7-a436-f13bcdd121b9\") " pod="openstack/dnsmasq-dns-8554648995-l7fnk" Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.449239 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/140a7c0b-b837-43f7-a436-f13bcdd121b9-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-l7fnk\" (UID: \"140a7c0b-b837-43f7-a436-f13bcdd121b9\") " pod="openstack/dnsmasq-dns-8554648995-l7fnk" Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.478553 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.551113 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/140a7c0b-b837-43f7-a436-f13bcdd121b9-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-l7fnk\" (UID: \"140a7c0b-b837-43f7-a436-f13bcdd121b9\") " pod="openstack/dnsmasq-dns-8554648995-l7fnk" Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.551224 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/140a7c0b-b837-43f7-a436-f13bcdd121b9-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-l7fnk\" (UID: \"140a7c0b-b837-43f7-a436-f13bcdd121b9\") " pod="openstack/dnsmasq-dns-8554648995-l7fnk" Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.551271 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gvbnm\" (UniqueName: \"kubernetes.io/projected/140a7c0b-b837-43f7-a436-f13bcdd121b9-kube-api-access-gvbnm\") pod \"dnsmasq-dns-8554648995-l7fnk\" (UID: \"140a7c0b-b837-43f7-a436-f13bcdd121b9\") " pod="openstack/dnsmasq-dns-8554648995-l7fnk" Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.551304 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/140a7c0b-b837-43f7-a436-f13bcdd121b9-dns-svc\") pod \"dnsmasq-dns-8554648995-l7fnk\" (UID: \"140a7c0b-b837-43f7-a436-f13bcdd121b9\") " pod="openstack/dnsmasq-dns-8554648995-l7fnk" Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.551383 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/140a7c0b-b837-43f7-a436-f13bcdd121b9-config\") pod \"dnsmasq-dns-8554648995-l7fnk\" (UID: \"140a7c0b-b837-43f7-a436-f13bcdd121b9\") " pod="openstack/dnsmasq-dns-8554648995-l7fnk" Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.552538 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/140a7c0b-b837-43f7-a436-f13bcdd121b9-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-l7fnk\" (UID: \"140a7c0b-b837-43f7-a436-f13bcdd121b9\") " pod="openstack/dnsmasq-dns-8554648995-l7fnk" Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.553326 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/140a7c0b-b837-43f7-a436-f13bcdd121b9-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-l7fnk\" (UID: \"140a7c0b-b837-43f7-a436-f13bcdd121b9\") " pod="openstack/dnsmasq-dns-8554648995-l7fnk" Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.553391 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/140a7c0b-b837-43f7-a436-f13bcdd121b9-config\") pod \"dnsmasq-dns-8554648995-l7fnk\" (UID: \"140a7c0b-b837-43f7-a436-f13bcdd121b9\") " pod="openstack/dnsmasq-dns-8554648995-l7fnk" Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.553725 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/140a7c0b-b837-43f7-a436-f13bcdd121b9-dns-svc\") pod \"dnsmasq-dns-8554648995-l7fnk\" (UID: \"140a7c0b-b837-43f7-a436-f13bcdd121b9\") " pod="openstack/dnsmasq-dns-8554648995-l7fnk" Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.581283 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gvbnm\" (UniqueName: \"kubernetes.io/projected/140a7c0b-b837-43f7-a436-f13bcdd121b9-kube-api-access-gvbnm\") pod \"dnsmasq-dns-8554648995-l7fnk\" (UID: \"140a7c0b-b837-43f7-a436-f13bcdd121b9\") " pod="openstack/dnsmasq-dns-8554648995-l7fnk" Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.581769 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-mr78v"] Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.661416 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-l7fnk" Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.738307 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.739847 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.743689 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.743939 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-vvg8r" Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.744281 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.744951 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.757890 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.805038 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-kkntf"] Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.855879 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/006cceec-8fb3-44a5-b3e1-50a01bc4b334-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"006cceec-8fb3-44a5-b3e1-50a01bc4b334\") " pod="openstack/ovn-northd-0" Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.855934 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/006cceec-8fb3-44a5-b3e1-50a01bc4b334-config\") pod \"ovn-northd-0\" (UID: \"006cceec-8fb3-44a5-b3e1-50a01bc4b334\") " pod="openstack/ovn-northd-0" Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.856088 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/006cceec-8fb3-44a5-b3e1-50a01bc4b334-scripts\") pod \"ovn-northd-0\" (UID: \"006cceec-8fb3-44a5-b3e1-50a01bc4b334\") " pod="openstack/ovn-northd-0" Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.856110 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/006cceec-8fb3-44a5-b3e1-50a01bc4b334-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"006cceec-8fb3-44a5-b3e1-50a01bc4b334\") " pod="openstack/ovn-northd-0" Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.856207 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sz6kd\" (UniqueName: \"kubernetes.io/projected/006cceec-8fb3-44a5-b3e1-50a01bc4b334-kube-api-access-sz6kd\") pod \"ovn-northd-0\" (UID: \"006cceec-8fb3-44a5-b3e1-50a01bc4b334\") " pod="openstack/ovn-northd-0" Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.856289 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/006cceec-8fb3-44a5-b3e1-50a01bc4b334-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"006cceec-8fb3-44a5-b3e1-50a01bc4b334\") " pod="openstack/ovn-northd-0" Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.856328 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/006cceec-8fb3-44a5-b3e1-50a01bc4b334-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"006cceec-8fb3-44a5-b3e1-50a01bc4b334\") " pod="openstack/ovn-northd-0" Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.958717 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/006cceec-8fb3-44a5-b3e1-50a01bc4b334-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"006cceec-8fb3-44a5-b3e1-50a01bc4b334\") " pod="openstack/ovn-northd-0" Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.958992 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/006cceec-8fb3-44a5-b3e1-50a01bc4b334-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"006cceec-8fb3-44a5-b3e1-50a01bc4b334\") " pod="openstack/ovn-northd-0" Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.959195 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/006cceec-8fb3-44a5-b3e1-50a01bc4b334-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"006cceec-8fb3-44a5-b3e1-50a01bc4b334\") " pod="openstack/ovn-northd-0" Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.959220 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/006cceec-8fb3-44a5-b3e1-50a01bc4b334-config\") pod \"ovn-northd-0\" (UID: \"006cceec-8fb3-44a5-b3e1-50a01bc4b334\") " pod="openstack/ovn-northd-0" Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.959270 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/006cceec-8fb3-44a5-b3e1-50a01bc4b334-scripts\") pod \"ovn-northd-0\" (UID: \"006cceec-8fb3-44a5-b3e1-50a01bc4b334\") " pod="openstack/ovn-northd-0" Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.959287 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/006cceec-8fb3-44a5-b3e1-50a01bc4b334-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"006cceec-8fb3-44a5-b3e1-50a01bc4b334\") " pod="openstack/ovn-northd-0" Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.959315 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sz6kd\" (UniqueName: \"kubernetes.io/projected/006cceec-8fb3-44a5-b3e1-50a01bc4b334-kube-api-access-sz6kd\") pod \"ovn-northd-0\" (UID: \"006cceec-8fb3-44a5-b3e1-50a01bc4b334\") " pod="openstack/ovn-northd-0" Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.961416 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/006cceec-8fb3-44a5-b3e1-50a01bc4b334-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"006cceec-8fb3-44a5-b3e1-50a01bc4b334\") " pod="openstack/ovn-northd-0" Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.962222 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/006cceec-8fb3-44a5-b3e1-50a01bc4b334-config\") pod \"ovn-northd-0\" (UID: \"006cceec-8fb3-44a5-b3e1-50a01bc4b334\") " pod="openstack/ovn-northd-0" Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.962367 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/006cceec-8fb3-44a5-b3e1-50a01bc4b334-scripts\") pod \"ovn-northd-0\" (UID: \"006cceec-8fb3-44a5-b3e1-50a01bc4b334\") " pod="openstack/ovn-northd-0" Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.964922 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/006cceec-8fb3-44a5-b3e1-50a01bc4b334-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"006cceec-8fb3-44a5-b3e1-50a01bc4b334\") " pod="openstack/ovn-northd-0" Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.966661 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/006cceec-8fb3-44a5-b3e1-50a01bc4b334-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"006cceec-8fb3-44a5-b3e1-50a01bc4b334\") " pod="openstack/ovn-northd-0" Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.969269 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/006cceec-8fb3-44a5-b3e1-50a01bc4b334-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"006cceec-8fb3-44a5-b3e1-50a01bc4b334\") " pod="openstack/ovn-northd-0" Nov 25 10:47:33 crc kubenswrapper[4821]: I1125 10:47:33.975853 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sz6kd\" (UniqueName: \"kubernetes.io/projected/006cceec-8fb3-44a5-b3e1-50a01bc4b334-kube-api-access-sz6kd\") pod \"ovn-northd-0\" (UID: \"006cceec-8fb3-44a5-b3e1-50a01bc4b334\") " pod="openstack/ovn-northd-0" Nov 25 10:47:34 crc kubenswrapper[4821]: I1125 10:47:34.059688 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 25 10:47:34 crc kubenswrapper[4821]: I1125 10:47:34.190308 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-l7fnk"] Nov 25 10:47:34 crc kubenswrapper[4821]: I1125 10:47:34.437650 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-kkntf" event={"ID":"832f5849-e58a-407b-9f45-55c52cfbcd02","Type":"ContainerStarted","Data":"ff362a776c38c52c270bff9892e4e53948c72205483b4d0af9e592263c226ec3"} Nov 25 10:47:34 crc kubenswrapper[4821]: I1125 10:47:34.437967 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-kkntf" event={"ID":"832f5849-e58a-407b-9f45-55c52cfbcd02","Type":"ContainerStarted","Data":"40999f7ffefa7e48d2c91e2cafd40511774c7d2947cb67d38f763bb3d8627c0b"} Nov 25 10:47:34 crc kubenswrapper[4821]: I1125 10:47:34.440746 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-l7fnk" event={"ID":"140a7c0b-b837-43f7-a436-f13bcdd121b9","Type":"ContainerStarted","Data":"75b3c9835d7ea64b3028a7a711e863997e6f6f8400479a89e324b5a948258155"} Nov 25 10:47:34 crc kubenswrapper[4821]: I1125 10:47:34.440766 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-l7fnk" event={"ID":"140a7c0b-b837-43f7-a436-f13bcdd121b9","Type":"ContainerStarted","Data":"c8dac4b2882be1438f030e415dff6ae88c561eaf3a8deddb3d074e68af1aaa51"} Nov 25 10:47:34 crc kubenswrapper[4821]: I1125 10:47:34.442593 4821 generic.go:334] "Generic (PLEG): container finished" podID="90d5b9f5-56b8-4904-b66b-3b28fe35c2e0" containerID="18bf29a97b52e256c472325cb819bf93dbf600eb9b6977229650951379d6fae6" exitCode=0 Nov 25 10:47:34 crc kubenswrapper[4821]: I1125 10:47:34.442658 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-mr78v" event={"ID":"90d5b9f5-56b8-4904-b66b-3b28fe35c2e0","Type":"ContainerDied","Data":"18bf29a97b52e256c472325cb819bf93dbf600eb9b6977229650951379d6fae6"} Nov 25 10:47:34 crc kubenswrapper[4821]: I1125 10:47:34.442687 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-mr78v" event={"ID":"90d5b9f5-56b8-4904-b66b-3b28fe35c2e0","Type":"ContainerStarted","Data":"b4653fd0a95fac17c7f69586ec98c52acd9d9da42d9e74780b270035ab2e29c7"} Nov 25 10:47:34 crc kubenswrapper[4821]: I1125 10:47:34.459083 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-kkntf" podStartSLOduration=2.459058657 podStartE2EDuration="2.459058657s" podCreationTimestamp="2025-11-25 10:47:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:47:34.455682641 +0000 UTC m=+924.992002498" watchObservedRunningTime="2025-11-25 10:47:34.459058657 +0000 UTC m=+924.995378534" Nov 25 10:47:34 crc kubenswrapper[4821]: I1125 10:47:34.509230 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 25 10:47:34 crc kubenswrapper[4821]: I1125 10:47:34.683818 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Nov 25 10:47:34 crc kubenswrapper[4821]: I1125 10:47:34.714156 4821 patch_prober.go:28] interesting pod/machine-config-daemon-2krbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:47:34 crc kubenswrapper[4821]: I1125 10:47:34.714212 4821 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:47:34 crc kubenswrapper[4821]: I1125 10:47:34.715952 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-mr78v" Nov 25 10:47:34 crc kubenswrapper[4821]: I1125 10:47:34.787245 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Nov 25 10:47:34 crc kubenswrapper[4821]: I1125 10:47:34.787284 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Nov 25 10:47:34 crc kubenswrapper[4821]: I1125 10:47:34.787771 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90d5b9f5-56b8-4904-b66b-3b28fe35c2e0-config\") pod \"90d5b9f5-56b8-4904-b66b-3b28fe35c2e0\" (UID: \"90d5b9f5-56b8-4904-b66b-3b28fe35c2e0\") " Nov 25 10:47:34 crc kubenswrapper[4821]: I1125 10:47:34.787809 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b87hd\" (UniqueName: \"kubernetes.io/projected/90d5b9f5-56b8-4904-b66b-3b28fe35c2e0-kube-api-access-b87hd\") pod \"90d5b9f5-56b8-4904-b66b-3b28fe35c2e0\" (UID: \"90d5b9f5-56b8-4904-b66b-3b28fe35c2e0\") " Nov 25 10:47:34 crc kubenswrapper[4821]: I1125 10:47:34.787884 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/90d5b9f5-56b8-4904-b66b-3b28fe35c2e0-ovsdbserver-nb\") pod \"90d5b9f5-56b8-4904-b66b-3b28fe35c2e0\" (UID: \"90d5b9f5-56b8-4904-b66b-3b28fe35c2e0\") " Nov 25 10:47:34 crc kubenswrapper[4821]: I1125 10:47:34.787931 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/90d5b9f5-56b8-4904-b66b-3b28fe35c2e0-dns-svc\") pod \"90d5b9f5-56b8-4904-b66b-3b28fe35c2e0\" (UID: \"90d5b9f5-56b8-4904-b66b-3b28fe35c2e0\") " Nov 25 10:47:34 crc kubenswrapper[4821]: I1125 10:47:34.791889 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90d5b9f5-56b8-4904-b66b-3b28fe35c2e0-kube-api-access-b87hd" (OuterVolumeSpecName: "kube-api-access-b87hd") pod "90d5b9f5-56b8-4904-b66b-3b28fe35c2e0" (UID: "90d5b9f5-56b8-4904-b66b-3b28fe35c2e0"). InnerVolumeSpecName "kube-api-access-b87hd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:47:34 crc kubenswrapper[4821]: I1125 10:47:34.814155 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/90d5b9f5-56b8-4904-b66b-3b28fe35c2e0-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "90d5b9f5-56b8-4904-b66b-3b28fe35c2e0" (UID: "90d5b9f5-56b8-4904-b66b-3b28fe35c2e0"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:47:34 crc kubenswrapper[4821]: I1125 10:47:34.816334 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/90d5b9f5-56b8-4904-b66b-3b28fe35c2e0-config" (OuterVolumeSpecName: "config") pod "90d5b9f5-56b8-4904-b66b-3b28fe35c2e0" (UID: "90d5b9f5-56b8-4904-b66b-3b28fe35c2e0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:47:34 crc kubenswrapper[4821]: I1125 10:47:34.821435 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/90d5b9f5-56b8-4904-b66b-3b28fe35c2e0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "90d5b9f5-56b8-4904-b66b-3b28fe35c2e0" (UID: "90d5b9f5-56b8-4904-b66b-3b28fe35c2e0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:47:34 crc kubenswrapper[4821]: I1125 10:47:34.889118 4821 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/90d5b9f5-56b8-4904-b66b-3b28fe35c2e0-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:34 crc kubenswrapper[4821]: I1125 10:47:34.889146 4821 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/90d5b9f5-56b8-4904-b66b-3b28fe35c2e0-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:34 crc kubenswrapper[4821]: I1125 10:47:34.889156 4821 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90d5b9f5-56b8-4904-b66b-3b28fe35c2e0-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:34 crc kubenswrapper[4821]: I1125 10:47:34.889186 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b87hd\" (UniqueName: \"kubernetes.io/projected/90d5b9f5-56b8-4904-b66b-3b28fe35c2e0-kube-api-access-b87hd\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:35 crc kubenswrapper[4821]: I1125 10:47:35.464731 4821 generic.go:334] "Generic (PLEG): container finished" podID="140a7c0b-b837-43f7-a436-f13bcdd121b9" containerID="75b3c9835d7ea64b3028a7a711e863997e6f6f8400479a89e324b5a948258155" exitCode=0 Nov 25 10:47:35 crc kubenswrapper[4821]: I1125 10:47:35.464911 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-l7fnk" event={"ID":"140a7c0b-b837-43f7-a436-f13bcdd121b9","Type":"ContainerDied","Data":"75b3c9835d7ea64b3028a7a711e863997e6f6f8400479a89e324b5a948258155"} Nov 25 10:47:35 crc kubenswrapper[4821]: I1125 10:47:35.468533 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"006cceec-8fb3-44a5-b3e1-50a01bc4b334","Type":"ContainerStarted","Data":"546c5ed9a95cc1f91a265bdbb88d92d178f67f045c1289e3c82bdacae7204bb9"} Nov 25 10:47:35 crc kubenswrapper[4821]: I1125 10:47:35.471458 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-mr78v" Nov 25 10:47:35 crc kubenswrapper[4821]: I1125 10:47:35.476312 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-mr78v" event={"ID":"90d5b9f5-56b8-4904-b66b-3b28fe35c2e0","Type":"ContainerDied","Data":"b4653fd0a95fac17c7f69586ec98c52acd9d9da42d9e74780b270035ab2e29c7"} Nov 25 10:47:35 crc kubenswrapper[4821]: I1125 10:47:35.476392 4821 scope.go:117] "RemoveContainer" containerID="18bf29a97b52e256c472325cb819bf93dbf600eb9b6977229650951379d6fae6" Nov 25 10:47:35 crc kubenswrapper[4821]: I1125 10:47:35.542783 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-mr78v"] Nov 25 10:47:35 crc kubenswrapper[4821]: I1125 10:47:35.547502 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-mr78v"] Nov 25 10:47:35 crc kubenswrapper[4821]: I1125 10:47:35.731428 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Nov 25 10:47:35 crc kubenswrapper[4821]: I1125 10:47:35.837826 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Nov 25 10:47:36 crc kubenswrapper[4821]: I1125 10:47:36.127382 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90d5b9f5-56b8-4904-b66b-3b28fe35c2e0" path="/var/lib/kubelet/pods/90d5b9f5-56b8-4904-b66b-3b28fe35c2e0/volumes" Nov 25 10:47:36 crc kubenswrapper[4821]: I1125 10:47:36.481212 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-l7fnk" event={"ID":"140a7c0b-b837-43f7-a436-f13bcdd121b9","Type":"ContainerStarted","Data":"152029880e04f0cb6f017bf75909e6af89ab5a26094cbfcf69a4db74291fd82e"} Nov 25 10:47:36 crc kubenswrapper[4821]: I1125 10:47:36.481367 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8554648995-l7fnk" Nov 25 10:47:36 crc kubenswrapper[4821]: I1125 10:47:36.486515 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"006cceec-8fb3-44a5-b3e1-50a01bc4b334","Type":"ContainerStarted","Data":"df21c503875e936383501bab2f1622643ec1fbd35cc2773e0caccb5a4382810c"} Nov 25 10:47:36 crc kubenswrapper[4821]: I1125 10:47:36.486570 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"006cceec-8fb3-44a5-b3e1-50a01bc4b334","Type":"ContainerStarted","Data":"910d040877b84b4a4ce836980138e0dca3c6280c0044cc27b1c5e231a5dfb9f8"} Nov 25 10:47:36 crc kubenswrapper[4821]: I1125 10:47:36.486618 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Nov 25 10:47:36 crc kubenswrapper[4821]: I1125 10:47:36.510066 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8554648995-l7fnk" podStartSLOduration=3.510046665 podStartE2EDuration="3.510046665s" podCreationTimestamp="2025-11-25 10:47:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:47:36.506928615 +0000 UTC m=+927.043248502" watchObservedRunningTime="2025-11-25 10:47:36.510046665 +0000 UTC m=+927.046366512" Nov 25 10:47:36 crc kubenswrapper[4821]: I1125 10:47:36.531841 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.301425405 podStartE2EDuration="3.531817297s" podCreationTimestamp="2025-11-25 10:47:33 +0000 UTC" firstStartedPulling="2025-11-25 10:47:34.522976478 +0000 UTC m=+925.059296325" lastFinishedPulling="2025-11-25 10:47:35.75336837 +0000 UTC m=+926.289688217" observedRunningTime="2025-11-25 10:47:36.524193634 +0000 UTC m=+927.060513481" watchObservedRunningTime="2025-11-25 10:47:36.531817297 +0000 UTC m=+927.068137154" Nov 25 10:47:36 crc kubenswrapper[4821]: I1125 10:47:36.741336 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 25 10:47:36 crc kubenswrapper[4821]: I1125 10:47:36.746709 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-l7fnk"] Nov 25 10:47:36 crc kubenswrapper[4821]: I1125 10:47:36.791561 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-wctvh"] Nov 25 10:47:36 crc kubenswrapper[4821]: E1125 10:47:36.791921 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90d5b9f5-56b8-4904-b66b-3b28fe35c2e0" containerName="init" Nov 25 10:47:36 crc kubenswrapper[4821]: I1125 10:47:36.791937 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="90d5b9f5-56b8-4904-b66b-3b28fe35c2e0" containerName="init" Nov 25 10:47:36 crc kubenswrapper[4821]: I1125 10:47:36.792102 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="90d5b9f5-56b8-4904-b66b-3b28fe35c2e0" containerName="init" Nov 25 10:47:36 crc kubenswrapper[4821]: I1125 10:47:36.792891 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-wctvh" Nov 25 10:47:36 crc kubenswrapper[4821]: I1125 10:47:36.814712 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-wctvh"] Nov 25 10:47:36 crc kubenswrapper[4821]: I1125 10:47:36.924307 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0333edd1-89f3-4ff0-9da3-544fac7620bb-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-wctvh\" (UID: \"0333edd1-89f3-4ff0-9da3-544fac7620bb\") " pod="openstack/dnsmasq-dns-b8fbc5445-wctvh" Nov 25 10:47:36 crc kubenswrapper[4821]: I1125 10:47:36.924407 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-759mz\" (UniqueName: \"kubernetes.io/projected/0333edd1-89f3-4ff0-9da3-544fac7620bb-kube-api-access-759mz\") pod \"dnsmasq-dns-b8fbc5445-wctvh\" (UID: \"0333edd1-89f3-4ff0-9da3-544fac7620bb\") " pod="openstack/dnsmasq-dns-b8fbc5445-wctvh" Nov 25 10:47:36 crc kubenswrapper[4821]: I1125 10:47:36.924471 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0333edd1-89f3-4ff0-9da3-544fac7620bb-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-wctvh\" (UID: \"0333edd1-89f3-4ff0-9da3-544fac7620bb\") " pod="openstack/dnsmasq-dns-b8fbc5445-wctvh" Nov 25 10:47:36 crc kubenswrapper[4821]: I1125 10:47:36.924498 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0333edd1-89f3-4ff0-9da3-544fac7620bb-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-wctvh\" (UID: \"0333edd1-89f3-4ff0-9da3-544fac7620bb\") " pod="openstack/dnsmasq-dns-b8fbc5445-wctvh" Nov 25 10:47:36 crc kubenswrapper[4821]: I1125 10:47:36.924526 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0333edd1-89f3-4ff0-9da3-544fac7620bb-config\") pod \"dnsmasq-dns-b8fbc5445-wctvh\" (UID: \"0333edd1-89f3-4ff0-9da3-544fac7620bb\") " pod="openstack/dnsmasq-dns-b8fbc5445-wctvh" Nov 25 10:47:37 crc kubenswrapper[4821]: I1125 10:47:37.025670 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-759mz\" (UniqueName: \"kubernetes.io/projected/0333edd1-89f3-4ff0-9da3-544fac7620bb-kube-api-access-759mz\") pod \"dnsmasq-dns-b8fbc5445-wctvh\" (UID: \"0333edd1-89f3-4ff0-9da3-544fac7620bb\") " pod="openstack/dnsmasq-dns-b8fbc5445-wctvh" Nov 25 10:47:37 crc kubenswrapper[4821]: I1125 10:47:37.025734 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0333edd1-89f3-4ff0-9da3-544fac7620bb-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-wctvh\" (UID: \"0333edd1-89f3-4ff0-9da3-544fac7620bb\") " pod="openstack/dnsmasq-dns-b8fbc5445-wctvh" Nov 25 10:47:37 crc kubenswrapper[4821]: I1125 10:47:37.025752 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0333edd1-89f3-4ff0-9da3-544fac7620bb-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-wctvh\" (UID: \"0333edd1-89f3-4ff0-9da3-544fac7620bb\") " pod="openstack/dnsmasq-dns-b8fbc5445-wctvh" Nov 25 10:47:37 crc kubenswrapper[4821]: I1125 10:47:37.025773 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0333edd1-89f3-4ff0-9da3-544fac7620bb-config\") pod \"dnsmasq-dns-b8fbc5445-wctvh\" (UID: \"0333edd1-89f3-4ff0-9da3-544fac7620bb\") " pod="openstack/dnsmasq-dns-b8fbc5445-wctvh" Nov 25 10:47:37 crc kubenswrapper[4821]: I1125 10:47:37.025861 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0333edd1-89f3-4ff0-9da3-544fac7620bb-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-wctvh\" (UID: \"0333edd1-89f3-4ff0-9da3-544fac7620bb\") " pod="openstack/dnsmasq-dns-b8fbc5445-wctvh" Nov 25 10:47:37 crc kubenswrapper[4821]: I1125 10:47:37.026841 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0333edd1-89f3-4ff0-9da3-544fac7620bb-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-wctvh\" (UID: \"0333edd1-89f3-4ff0-9da3-544fac7620bb\") " pod="openstack/dnsmasq-dns-b8fbc5445-wctvh" Nov 25 10:47:37 crc kubenswrapper[4821]: I1125 10:47:37.027587 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0333edd1-89f3-4ff0-9da3-544fac7620bb-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-wctvh\" (UID: \"0333edd1-89f3-4ff0-9da3-544fac7620bb\") " pod="openstack/dnsmasq-dns-b8fbc5445-wctvh" Nov 25 10:47:37 crc kubenswrapper[4821]: I1125 10:47:37.028092 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0333edd1-89f3-4ff0-9da3-544fac7620bb-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-wctvh\" (UID: \"0333edd1-89f3-4ff0-9da3-544fac7620bb\") " pod="openstack/dnsmasq-dns-b8fbc5445-wctvh" Nov 25 10:47:37 crc kubenswrapper[4821]: I1125 10:47:37.028598 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0333edd1-89f3-4ff0-9da3-544fac7620bb-config\") pod \"dnsmasq-dns-b8fbc5445-wctvh\" (UID: \"0333edd1-89f3-4ff0-9da3-544fac7620bb\") " pod="openstack/dnsmasq-dns-b8fbc5445-wctvh" Nov 25 10:47:37 crc kubenswrapper[4821]: I1125 10:47:37.043781 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-759mz\" (UniqueName: \"kubernetes.io/projected/0333edd1-89f3-4ff0-9da3-544fac7620bb-kube-api-access-759mz\") pod \"dnsmasq-dns-b8fbc5445-wctvh\" (UID: \"0333edd1-89f3-4ff0-9da3-544fac7620bb\") " pod="openstack/dnsmasq-dns-b8fbc5445-wctvh" Nov 25 10:47:37 crc kubenswrapper[4821]: I1125 10:47:37.118098 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-wctvh" Nov 25 10:47:37 crc kubenswrapper[4821]: I1125 10:47:37.566607 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-wctvh"] Nov 25 10:47:37 crc kubenswrapper[4821]: W1125 10:47:37.573382 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0333edd1_89f3_4ff0_9da3_544fac7620bb.slice/crio-7bfb195a271d2a394c92d33cf0882fc4bde496bec9c182da2ebc130f98daf5fb WatchSource:0}: Error finding container 7bfb195a271d2a394c92d33cf0882fc4bde496bec9c182da2ebc130f98daf5fb: Status 404 returned error can't find the container with id 7bfb195a271d2a394c92d33cf0882fc4bde496bec9c182da2ebc130f98daf5fb Nov 25 10:47:37 crc kubenswrapper[4821]: I1125 10:47:37.888468 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Nov 25 10:47:37 crc kubenswrapper[4821]: I1125 10:47:37.893704 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 25 10:47:37 crc kubenswrapper[4821]: I1125 10:47:37.896702 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Nov 25 10:47:37 crc kubenswrapper[4821]: I1125 10:47:37.896748 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Nov 25 10:47:37 crc kubenswrapper[4821]: I1125 10:47:37.896926 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-bfshm" Nov 25 10:47:37 crc kubenswrapper[4821]: I1125 10:47:37.896968 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Nov 25 10:47:37 crc kubenswrapper[4821]: I1125 10:47:37.912680 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Nov 25 10:47:38 crc kubenswrapper[4821]: I1125 10:47:38.050113 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/0ffb0da2-affb-4a95-87f3-eac442dc754d-cache\") pod \"swift-storage-0\" (UID: \"0ffb0da2-affb-4a95-87f3-eac442dc754d\") " pod="openstack/swift-storage-0" Nov 25 10:47:38 crc kubenswrapper[4821]: I1125 10:47:38.050213 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0ffb0da2-affb-4a95-87f3-eac442dc754d-etc-swift\") pod \"swift-storage-0\" (UID: \"0ffb0da2-affb-4a95-87f3-eac442dc754d\") " pod="openstack/swift-storage-0" Nov 25 10:47:38 crc kubenswrapper[4821]: I1125 10:47:38.050262 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/0ffb0da2-affb-4a95-87f3-eac442dc754d-lock\") pod \"swift-storage-0\" (UID: \"0ffb0da2-affb-4a95-87f3-eac442dc754d\") " pod="openstack/swift-storage-0" Nov 25 10:47:38 crc kubenswrapper[4821]: I1125 10:47:38.050296 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xr9g4\" (UniqueName: \"kubernetes.io/projected/0ffb0da2-affb-4a95-87f3-eac442dc754d-kube-api-access-xr9g4\") pod \"swift-storage-0\" (UID: \"0ffb0da2-affb-4a95-87f3-eac442dc754d\") " pod="openstack/swift-storage-0" Nov 25 10:47:38 crc kubenswrapper[4821]: I1125 10:47:38.050322 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"swift-storage-0\" (UID: \"0ffb0da2-affb-4a95-87f3-eac442dc754d\") " pod="openstack/swift-storage-0" Nov 25 10:47:38 crc kubenswrapper[4821]: I1125 10:47:38.151154 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/0ffb0da2-affb-4a95-87f3-eac442dc754d-cache\") pod \"swift-storage-0\" (UID: \"0ffb0da2-affb-4a95-87f3-eac442dc754d\") " pod="openstack/swift-storage-0" Nov 25 10:47:38 crc kubenswrapper[4821]: I1125 10:47:38.151271 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0ffb0da2-affb-4a95-87f3-eac442dc754d-etc-swift\") pod \"swift-storage-0\" (UID: \"0ffb0da2-affb-4a95-87f3-eac442dc754d\") " pod="openstack/swift-storage-0" Nov 25 10:47:38 crc kubenswrapper[4821]: I1125 10:47:38.151321 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/0ffb0da2-affb-4a95-87f3-eac442dc754d-lock\") pod \"swift-storage-0\" (UID: \"0ffb0da2-affb-4a95-87f3-eac442dc754d\") " pod="openstack/swift-storage-0" Nov 25 10:47:38 crc kubenswrapper[4821]: I1125 10:47:38.151372 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xr9g4\" (UniqueName: \"kubernetes.io/projected/0ffb0da2-affb-4a95-87f3-eac442dc754d-kube-api-access-xr9g4\") pod \"swift-storage-0\" (UID: \"0ffb0da2-affb-4a95-87f3-eac442dc754d\") " pod="openstack/swift-storage-0" Nov 25 10:47:38 crc kubenswrapper[4821]: I1125 10:47:38.151394 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"swift-storage-0\" (UID: \"0ffb0da2-affb-4a95-87f3-eac442dc754d\") " pod="openstack/swift-storage-0" Nov 25 10:47:38 crc kubenswrapper[4821]: E1125 10:47:38.151459 4821 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 25 10:47:38 crc kubenswrapper[4821]: E1125 10:47:38.151478 4821 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 25 10:47:38 crc kubenswrapper[4821]: E1125 10:47:38.151530 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/0ffb0da2-affb-4a95-87f3-eac442dc754d-etc-swift podName:0ffb0da2-affb-4a95-87f3-eac442dc754d nodeName:}" failed. No retries permitted until 2025-11-25 10:47:38.651511255 +0000 UTC m=+929.187831102 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/0ffb0da2-affb-4a95-87f3-eac442dc754d-etc-swift") pod "swift-storage-0" (UID: "0ffb0da2-affb-4a95-87f3-eac442dc754d") : configmap "swift-ring-files" not found Nov 25 10:47:38 crc kubenswrapper[4821]: I1125 10:47:38.151911 4821 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"swift-storage-0\" (UID: \"0ffb0da2-affb-4a95-87f3-eac442dc754d\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/swift-storage-0" Nov 25 10:47:38 crc kubenswrapper[4821]: I1125 10:47:38.152806 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/0ffb0da2-affb-4a95-87f3-eac442dc754d-lock\") pod \"swift-storage-0\" (UID: \"0ffb0da2-affb-4a95-87f3-eac442dc754d\") " pod="openstack/swift-storage-0" Nov 25 10:47:38 crc kubenswrapper[4821]: I1125 10:47:38.152828 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/0ffb0da2-affb-4a95-87f3-eac442dc754d-cache\") pod \"swift-storage-0\" (UID: \"0ffb0da2-affb-4a95-87f3-eac442dc754d\") " pod="openstack/swift-storage-0" Nov 25 10:47:38 crc kubenswrapper[4821]: I1125 10:47:38.168357 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xr9g4\" (UniqueName: \"kubernetes.io/projected/0ffb0da2-affb-4a95-87f3-eac442dc754d-kube-api-access-xr9g4\") pod \"swift-storage-0\" (UID: \"0ffb0da2-affb-4a95-87f3-eac442dc754d\") " pod="openstack/swift-storage-0" Nov 25 10:47:38 crc kubenswrapper[4821]: I1125 10:47:38.190896 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"swift-storage-0\" (UID: \"0ffb0da2-affb-4a95-87f3-eac442dc754d\") " pod="openstack/swift-storage-0" Nov 25 10:47:38 crc kubenswrapper[4821]: I1125 10:47:38.433946 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-gmrpw"] Nov 25 10:47:38 crc kubenswrapper[4821]: I1125 10:47:38.435126 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-gmrpw" Nov 25 10:47:38 crc kubenswrapper[4821]: I1125 10:47:38.437085 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Nov 25 10:47:38 crc kubenswrapper[4821]: I1125 10:47:38.437318 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Nov 25 10:47:38 crc kubenswrapper[4821]: I1125 10:47:38.439061 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Nov 25 10:47:38 crc kubenswrapper[4821]: I1125 10:47:38.441522 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-gmrpw"] Nov 25 10:47:38 crc kubenswrapper[4821]: I1125 10:47:38.502003 4821 generic.go:334] "Generic (PLEG): container finished" podID="0333edd1-89f3-4ff0-9da3-544fac7620bb" containerID="1b15650f1234d447a4a820c0d5f0a318e138ce1c4873e82e6ba0d4fb074ad408" exitCode=0 Nov 25 10:47:38 crc kubenswrapper[4821]: I1125 10:47:38.502050 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-wctvh" event={"ID":"0333edd1-89f3-4ff0-9da3-544fac7620bb","Type":"ContainerDied","Data":"1b15650f1234d447a4a820c0d5f0a318e138ce1c4873e82e6ba0d4fb074ad408"} Nov 25 10:47:38 crc kubenswrapper[4821]: I1125 10:47:38.502086 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-wctvh" event={"ID":"0333edd1-89f3-4ff0-9da3-544fac7620bb","Type":"ContainerStarted","Data":"7bfb195a271d2a394c92d33cf0882fc4bde496bec9c182da2ebc130f98daf5fb"} Nov 25 10:47:38 crc kubenswrapper[4821]: I1125 10:47:38.502375 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8554648995-l7fnk" podUID="140a7c0b-b837-43f7-a436-f13bcdd121b9" containerName="dnsmasq-dns" containerID="cri-o://152029880e04f0cb6f017bf75909e6af89ab5a26094cbfcf69a4db74291fd82e" gracePeriod=10 Nov 25 10:47:38 crc kubenswrapper[4821]: I1125 10:47:38.560539 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1aa8233d-1fe4-44ec-acc6-0d5548c70d59-ring-data-devices\") pod \"swift-ring-rebalance-gmrpw\" (UID: \"1aa8233d-1fe4-44ec-acc6-0d5548c70d59\") " pod="openstack/swift-ring-rebalance-gmrpw" Nov 25 10:47:38 crc kubenswrapper[4821]: I1125 10:47:38.560646 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1aa8233d-1fe4-44ec-acc6-0d5548c70d59-dispersionconf\") pod \"swift-ring-rebalance-gmrpw\" (UID: \"1aa8233d-1fe4-44ec-acc6-0d5548c70d59\") " pod="openstack/swift-ring-rebalance-gmrpw" Nov 25 10:47:38 crc kubenswrapper[4821]: I1125 10:47:38.560706 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1aa8233d-1fe4-44ec-acc6-0d5548c70d59-etc-swift\") pod \"swift-ring-rebalance-gmrpw\" (UID: \"1aa8233d-1fe4-44ec-acc6-0d5548c70d59\") " pod="openstack/swift-ring-rebalance-gmrpw" Nov 25 10:47:38 crc kubenswrapper[4821]: I1125 10:47:38.560805 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1aa8233d-1fe4-44ec-acc6-0d5548c70d59-combined-ca-bundle\") pod \"swift-ring-rebalance-gmrpw\" (UID: \"1aa8233d-1fe4-44ec-acc6-0d5548c70d59\") " pod="openstack/swift-ring-rebalance-gmrpw" Nov 25 10:47:38 crc kubenswrapper[4821]: I1125 10:47:38.560857 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1aa8233d-1fe4-44ec-acc6-0d5548c70d59-swiftconf\") pod \"swift-ring-rebalance-gmrpw\" (UID: \"1aa8233d-1fe4-44ec-acc6-0d5548c70d59\") " pod="openstack/swift-ring-rebalance-gmrpw" Nov 25 10:47:38 crc kubenswrapper[4821]: I1125 10:47:38.560884 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-php28\" (UniqueName: \"kubernetes.io/projected/1aa8233d-1fe4-44ec-acc6-0d5548c70d59-kube-api-access-php28\") pod \"swift-ring-rebalance-gmrpw\" (UID: \"1aa8233d-1fe4-44ec-acc6-0d5548c70d59\") " pod="openstack/swift-ring-rebalance-gmrpw" Nov 25 10:47:38 crc kubenswrapper[4821]: I1125 10:47:38.560960 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1aa8233d-1fe4-44ec-acc6-0d5548c70d59-scripts\") pod \"swift-ring-rebalance-gmrpw\" (UID: \"1aa8233d-1fe4-44ec-acc6-0d5548c70d59\") " pod="openstack/swift-ring-rebalance-gmrpw" Nov 25 10:47:38 crc kubenswrapper[4821]: I1125 10:47:38.663230 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1aa8233d-1fe4-44ec-acc6-0d5548c70d59-combined-ca-bundle\") pod \"swift-ring-rebalance-gmrpw\" (UID: \"1aa8233d-1fe4-44ec-acc6-0d5548c70d59\") " pod="openstack/swift-ring-rebalance-gmrpw" Nov 25 10:47:38 crc kubenswrapper[4821]: I1125 10:47:38.663556 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1aa8233d-1fe4-44ec-acc6-0d5548c70d59-swiftconf\") pod \"swift-ring-rebalance-gmrpw\" (UID: \"1aa8233d-1fe4-44ec-acc6-0d5548c70d59\") " pod="openstack/swift-ring-rebalance-gmrpw" Nov 25 10:47:38 crc kubenswrapper[4821]: I1125 10:47:38.663579 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-php28\" (UniqueName: \"kubernetes.io/projected/1aa8233d-1fe4-44ec-acc6-0d5548c70d59-kube-api-access-php28\") pod \"swift-ring-rebalance-gmrpw\" (UID: \"1aa8233d-1fe4-44ec-acc6-0d5548c70d59\") " pod="openstack/swift-ring-rebalance-gmrpw" Nov 25 10:47:38 crc kubenswrapper[4821]: I1125 10:47:38.663598 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1aa8233d-1fe4-44ec-acc6-0d5548c70d59-scripts\") pod \"swift-ring-rebalance-gmrpw\" (UID: \"1aa8233d-1fe4-44ec-acc6-0d5548c70d59\") " pod="openstack/swift-ring-rebalance-gmrpw" Nov 25 10:47:38 crc kubenswrapper[4821]: I1125 10:47:38.663642 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1aa8233d-1fe4-44ec-acc6-0d5548c70d59-ring-data-devices\") pod \"swift-ring-rebalance-gmrpw\" (UID: \"1aa8233d-1fe4-44ec-acc6-0d5548c70d59\") " pod="openstack/swift-ring-rebalance-gmrpw" Nov 25 10:47:38 crc kubenswrapper[4821]: I1125 10:47:38.663683 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1aa8233d-1fe4-44ec-acc6-0d5548c70d59-dispersionconf\") pod \"swift-ring-rebalance-gmrpw\" (UID: \"1aa8233d-1fe4-44ec-acc6-0d5548c70d59\") " pod="openstack/swift-ring-rebalance-gmrpw" Nov 25 10:47:38 crc kubenswrapper[4821]: I1125 10:47:38.663708 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0ffb0da2-affb-4a95-87f3-eac442dc754d-etc-swift\") pod \"swift-storage-0\" (UID: \"0ffb0da2-affb-4a95-87f3-eac442dc754d\") " pod="openstack/swift-storage-0" Nov 25 10:47:38 crc kubenswrapper[4821]: I1125 10:47:38.663736 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1aa8233d-1fe4-44ec-acc6-0d5548c70d59-etc-swift\") pod \"swift-ring-rebalance-gmrpw\" (UID: \"1aa8233d-1fe4-44ec-acc6-0d5548c70d59\") " pod="openstack/swift-ring-rebalance-gmrpw" Nov 25 10:47:38 crc kubenswrapper[4821]: I1125 10:47:38.665220 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1aa8233d-1fe4-44ec-acc6-0d5548c70d59-ring-data-devices\") pod \"swift-ring-rebalance-gmrpw\" (UID: \"1aa8233d-1fe4-44ec-acc6-0d5548c70d59\") " pod="openstack/swift-ring-rebalance-gmrpw" Nov 25 10:47:38 crc kubenswrapper[4821]: I1125 10:47:38.665251 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1aa8233d-1fe4-44ec-acc6-0d5548c70d59-etc-swift\") pod \"swift-ring-rebalance-gmrpw\" (UID: \"1aa8233d-1fe4-44ec-acc6-0d5548c70d59\") " pod="openstack/swift-ring-rebalance-gmrpw" Nov 25 10:47:38 crc kubenswrapper[4821]: I1125 10:47:38.665358 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1aa8233d-1fe4-44ec-acc6-0d5548c70d59-scripts\") pod \"swift-ring-rebalance-gmrpw\" (UID: \"1aa8233d-1fe4-44ec-acc6-0d5548c70d59\") " pod="openstack/swift-ring-rebalance-gmrpw" Nov 25 10:47:38 crc kubenswrapper[4821]: E1125 10:47:38.665396 4821 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 25 10:47:38 crc kubenswrapper[4821]: E1125 10:47:38.665415 4821 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 25 10:47:38 crc kubenswrapper[4821]: E1125 10:47:38.665472 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/0ffb0da2-affb-4a95-87f3-eac442dc754d-etc-swift podName:0ffb0da2-affb-4a95-87f3-eac442dc754d nodeName:}" failed. No retries permitted until 2025-11-25 10:47:39.665455262 +0000 UTC m=+930.201775169 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/0ffb0da2-affb-4a95-87f3-eac442dc754d-etc-swift") pod "swift-storage-0" (UID: "0ffb0da2-affb-4a95-87f3-eac442dc754d") : configmap "swift-ring-files" not found Nov 25 10:47:38 crc kubenswrapper[4821]: I1125 10:47:38.669886 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1aa8233d-1fe4-44ec-acc6-0d5548c70d59-combined-ca-bundle\") pod \"swift-ring-rebalance-gmrpw\" (UID: \"1aa8233d-1fe4-44ec-acc6-0d5548c70d59\") " pod="openstack/swift-ring-rebalance-gmrpw" Nov 25 10:47:38 crc kubenswrapper[4821]: I1125 10:47:38.671889 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1aa8233d-1fe4-44ec-acc6-0d5548c70d59-dispersionconf\") pod \"swift-ring-rebalance-gmrpw\" (UID: \"1aa8233d-1fe4-44ec-acc6-0d5548c70d59\") " pod="openstack/swift-ring-rebalance-gmrpw" Nov 25 10:47:38 crc kubenswrapper[4821]: I1125 10:47:38.672876 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1aa8233d-1fe4-44ec-acc6-0d5548c70d59-swiftconf\") pod \"swift-ring-rebalance-gmrpw\" (UID: \"1aa8233d-1fe4-44ec-acc6-0d5548c70d59\") " pod="openstack/swift-ring-rebalance-gmrpw" Nov 25 10:47:38 crc kubenswrapper[4821]: I1125 10:47:38.684260 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-php28\" (UniqueName: \"kubernetes.io/projected/1aa8233d-1fe4-44ec-acc6-0d5548c70d59-kube-api-access-php28\") pod \"swift-ring-rebalance-gmrpw\" (UID: \"1aa8233d-1fe4-44ec-acc6-0d5548c70d59\") " pod="openstack/swift-ring-rebalance-gmrpw" Nov 25 10:47:38 crc kubenswrapper[4821]: I1125 10:47:38.758874 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-gmrpw" Nov 25 10:47:38 crc kubenswrapper[4821]: I1125 10:47:38.965501 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-l7fnk" Nov 25 10:47:39 crc kubenswrapper[4821]: I1125 10:47:39.009693 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Nov 25 10:47:39 crc kubenswrapper[4821]: I1125 10:47:39.067793 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/140a7c0b-b837-43f7-a436-f13bcdd121b9-dns-svc\") pod \"140a7c0b-b837-43f7-a436-f13bcdd121b9\" (UID: \"140a7c0b-b837-43f7-a436-f13bcdd121b9\") " Nov 25 10:47:39 crc kubenswrapper[4821]: I1125 10:47:39.067931 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/140a7c0b-b837-43f7-a436-f13bcdd121b9-ovsdbserver-sb\") pod \"140a7c0b-b837-43f7-a436-f13bcdd121b9\" (UID: \"140a7c0b-b837-43f7-a436-f13bcdd121b9\") " Nov 25 10:47:39 crc kubenswrapper[4821]: I1125 10:47:39.067987 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/140a7c0b-b837-43f7-a436-f13bcdd121b9-config\") pod \"140a7c0b-b837-43f7-a436-f13bcdd121b9\" (UID: \"140a7c0b-b837-43f7-a436-f13bcdd121b9\") " Nov 25 10:47:39 crc kubenswrapper[4821]: I1125 10:47:39.068009 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/140a7c0b-b837-43f7-a436-f13bcdd121b9-ovsdbserver-nb\") pod \"140a7c0b-b837-43f7-a436-f13bcdd121b9\" (UID: \"140a7c0b-b837-43f7-a436-f13bcdd121b9\") " Nov 25 10:47:39 crc kubenswrapper[4821]: I1125 10:47:39.068045 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gvbnm\" (UniqueName: \"kubernetes.io/projected/140a7c0b-b837-43f7-a436-f13bcdd121b9-kube-api-access-gvbnm\") pod \"140a7c0b-b837-43f7-a436-f13bcdd121b9\" (UID: \"140a7c0b-b837-43f7-a436-f13bcdd121b9\") " Nov 25 10:47:39 crc kubenswrapper[4821]: I1125 10:47:39.074420 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/140a7c0b-b837-43f7-a436-f13bcdd121b9-kube-api-access-gvbnm" (OuterVolumeSpecName: "kube-api-access-gvbnm") pod "140a7c0b-b837-43f7-a436-f13bcdd121b9" (UID: "140a7c0b-b837-43f7-a436-f13bcdd121b9"). InnerVolumeSpecName "kube-api-access-gvbnm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:47:39 crc kubenswrapper[4821]: I1125 10:47:39.099511 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Nov 25 10:47:39 crc kubenswrapper[4821]: I1125 10:47:39.118142 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/140a7c0b-b837-43f7-a436-f13bcdd121b9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "140a7c0b-b837-43f7-a436-f13bcdd121b9" (UID: "140a7c0b-b837-43f7-a436-f13bcdd121b9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:47:39 crc kubenswrapper[4821]: I1125 10:47:39.123520 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/140a7c0b-b837-43f7-a436-f13bcdd121b9-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "140a7c0b-b837-43f7-a436-f13bcdd121b9" (UID: "140a7c0b-b837-43f7-a436-f13bcdd121b9"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:47:39 crc kubenswrapper[4821]: I1125 10:47:39.124688 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/140a7c0b-b837-43f7-a436-f13bcdd121b9-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "140a7c0b-b837-43f7-a436-f13bcdd121b9" (UID: "140a7c0b-b837-43f7-a436-f13bcdd121b9"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:47:39 crc kubenswrapper[4821]: I1125 10:47:39.124832 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/140a7c0b-b837-43f7-a436-f13bcdd121b9-config" (OuterVolumeSpecName: "config") pod "140a7c0b-b837-43f7-a436-f13bcdd121b9" (UID: "140a7c0b-b837-43f7-a436-f13bcdd121b9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:47:39 crc kubenswrapper[4821]: I1125 10:47:39.169715 4821 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/140a7c0b-b837-43f7-a436-f13bcdd121b9-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:39 crc kubenswrapper[4821]: I1125 10:47:39.169742 4821 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/140a7c0b-b837-43f7-a436-f13bcdd121b9-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:39 crc kubenswrapper[4821]: I1125 10:47:39.169754 4821 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/140a7c0b-b837-43f7-a436-f13bcdd121b9-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:39 crc kubenswrapper[4821]: I1125 10:47:39.169765 4821 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/140a7c0b-b837-43f7-a436-f13bcdd121b9-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:39 crc kubenswrapper[4821]: I1125 10:47:39.169776 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gvbnm\" (UniqueName: \"kubernetes.io/projected/140a7c0b-b837-43f7-a436-f13bcdd121b9-kube-api-access-gvbnm\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:39 crc kubenswrapper[4821]: W1125 10:47:39.235446 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1aa8233d_1fe4_44ec_acc6_0d5548c70d59.slice/crio-d4053685b0a60e6d2adcf8455c3d95f804252d698b630071dae91597fb96ff6f WatchSource:0}: Error finding container d4053685b0a60e6d2adcf8455c3d95f804252d698b630071dae91597fb96ff6f: Status 404 returned error can't find the container with id d4053685b0a60e6d2adcf8455c3d95f804252d698b630071dae91597fb96ff6f Nov 25 10:47:39 crc kubenswrapper[4821]: I1125 10:47:39.246868 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-gmrpw"] Nov 25 10:47:39 crc kubenswrapper[4821]: I1125 10:47:39.510614 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-gmrpw" event={"ID":"1aa8233d-1fe4-44ec-acc6-0d5548c70d59","Type":"ContainerStarted","Data":"d4053685b0a60e6d2adcf8455c3d95f804252d698b630071dae91597fb96ff6f"} Nov 25 10:47:39 crc kubenswrapper[4821]: I1125 10:47:39.512467 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-wctvh" event={"ID":"0333edd1-89f3-4ff0-9da3-544fac7620bb","Type":"ContainerStarted","Data":"b07e78295a4d2230639263bc614ce41918d2a2e30f14bafc5a14bae2081a653b"} Nov 25 10:47:39 crc kubenswrapper[4821]: I1125 10:47:39.512561 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b8fbc5445-wctvh" Nov 25 10:47:39 crc kubenswrapper[4821]: I1125 10:47:39.514284 4821 generic.go:334] "Generic (PLEG): container finished" podID="140a7c0b-b837-43f7-a436-f13bcdd121b9" containerID="152029880e04f0cb6f017bf75909e6af89ab5a26094cbfcf69a4db74291fd82e" exitCode=0 Nov 25 10:47:39 crc kubenswrapper[4821]: I1125 10:47:39.514335 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-l7fnk" Nov 25 10:47:39 crc kubenswrapper[4821]: I1125 10:47:39.514347 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-l7fnk" event={"ID":"140a7c0b-b837-43f7-a436-f13bcdd121b9","Type":"ContainerDied","Data":"152029880e04f0cb6f017bf75909e6af89ab5a26094cbfcf69a4db74291fd82e"} Nov 25 10:47:39 crc kubenswrapper[4821]: I1125 10:47:39.514384 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-l7fnk" event={"ID":"140a7c0b-b837-43f7-a436-f13bcdd121b9","Type":"ContainerDied","Data":"c8dac4b2882be1438f030e415dff6ae88c561eaf3a8deddb3d074e68af1aaa51"} Nov 25 10:47:39 crc kubenswrapper[4821]: I1125 10:47:39.514404 4821 scope.go:117] "RemoveContainer" containerID="152029880e04f0cb6f017bf75909e6af89ab5a26094cbfcf69a4db74291fd82e" Nov 25 10:47:39 crc kubenswrapper[4821]: I1125 10:47:39.534069 4821 scope.go:117] "RemoveContainer" containerID="75b3c9835d7ea64b3028a7a711e863997e6f6f8400479a89e324b5a948258155" Nov 25 10:47:39 crc kubenswrapper[4821]: I1125 10:47:39.538217 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-b8fbc5445-wctvh" podStartSLOduration=3.538203721 podStartE2EDuration="3.538203721s" podCreationTimestamp="2025-11-25 10:47:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:47:39.536104708 +0000 UTC m=+930.072424555" watchObservedRunningTime="2025-11-25 10:47:39.538203721 +0000 UTC m=+930.074523568" Nov 25 10:47:39 crc kubenswrapper[4821]: I1125 10:47:39.554303 4821 scope.go:117] "RemoveContainer" containerID="152029880e04f0cb6f017bf75909e6af89ab5a26094cbfcf69a4db74291fd82e" Nov 25 10:47:39 crc kubenswrapper[4821]: E1125 10:47:39.554956 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"152029880e04f0cb6f017bf75909e6af89ab5a26094cbfcf69a4db74291fd82e\": container with ID starting with 152029880e04f0cb6f017bf75909e6af89ab5a26094cbfcf69a4db74291fd82e not found: ID does not exist" containerID="152029880e04f0cb6f017bf75909e6af89ab5a26094cbfcf69a4db74291fd82e" Nov 25 10:47:39 crc kubenswrapper[4821]: I1125 10:47:39.554995 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"152029880e04f0cb6f017bf75909e6af89ab5a26094cbfcf69a4db74291fd82e"} err="failed to get container status \"152029880e04f0cb6f017bf75909e6af89ab5a26094cbfcf69a4db74291fd82e\": rpc error: code = NotFound desc = could not find container \"152029880e04f0cb6f017bf75909e6af89ab5a26094cbfcf69a4db74291fd82e\": container with ID starting with 152029880e04f0cb6f017bf75909e6af89ab5a26094cbfcf69a4db74291fd82e not found: ID does not exist" Nov 25 10:47:39 crc kubenswrapper[4821]: I1125 10:47:39.555023 4821 scope.go:117] "RemoveContainer" containerID="75b3c9835d7ea64b3028a7a711e863997e6f6f8400479a89e324b5a948258155" Nov 25 10:47:39 crc kubenswrapper[4821]: I1125 10:47:39.556598 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-l7fnk"] Nov 25 10:47:39 crc kubenswrapper[4821]: E1125 10:47:39.560801 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75b3c9835d7ea64b3028a7a711e863997e6f6f8400479a89e324b5a948258155\": container with ID starting with 75b3c9835d7ea64b3028a7a711e863997e6f6f8400479a89e324b5a948258155 not found: ID does not exist" containerID="75b3c9835d7ea64b3028a7a711e863997e6f6f8400479a89e324b5a948258155" Nov 25 10:47:39 crc kubenswrapper[4821]: I1125 10:47:39.560845 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75b3c9835d7ea64b3028a7a711e863997e6f6f8400479a89e324b5a948258155"} err="failed to get container status \"75b3c9835d7ea64b3028a7a711e863997e6f6f8400479a89e324b5a948258155\": rpc error: code = NotFound desc = could not find container \"75b3c9835d7ea64b3028a7a711e863997e6f6f8400479a89e324b5a948258155\": container with ID starting with 75b3c9835d7ea64b3028a7a711e863997e6f6f8400479a89e324b5a948258155 not found: ID does not exist" Nov 25 10:47:39 crc kubenswrapper[4821]: I1125 10:47:39.563333 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8554648995-l7fnk"] Nov 25 10:47:39 crc kubenswrapper[4821]: I1125 10:47:39.677929 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0ffb0da2-affb-4a95-87f3-eac442dc754d-etc-swift\") pod \"swift-storage-0\" (UID: \"0ffb0da2-affb-4a95-87f3-eac442dc754d\") " pod="openstack/swift-storage-0" Nov 25 10:47:39 crc kubenswrapper[4821]: E1125 10:47:39.678188 4821 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 25 10:47:39 crc kubenswrapper[4821]: E1125 10:47:39.678220 4821 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 25 10:47:39 crc kubenswrapper[4821]: E1125 10:47:39.678267 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/0ffb0da2-affb-4a95-87f3-eac442dc754d-etc-swift podName:0ffb0da2-affb-4a95-87f3-eac442dc754d nodeName:}" failed. No retries permitted until 2025-11-25 10:47:41.678249084 +0000 UTC m=+932.214568981 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/0ffb0da2-affb-4a95-87f3-eac442dc754d-etc-swift") pod "swift-storage-0" (UID: "0ffb0da2-affb-4a95-87f3-eac442dc754d") : configmap "swift-ring-files" not found Nov 25 10:47:40 crc kubenswrapper[4821]: I1125 10:47:40.135262 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="140a7c0b-b837-43f7-a436-f13bcdd121b9" path="/var/lib/kubelet/pods/140a7c0b-b837-43f7-a436-f13bcdd121b9/volumes" Nov 25 10:47:41 crc kubenswrapper[4821]: I1125 10:47:41.716570 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0ffb0da2-affb-4a95-87f3-eac442dc754d-etc-swift\") pod \"swift-storage-0\" (UID: \"0ffb0da2-affb-4a95-87f3-eac442dc754d\") " pod="openstack/swift-storage-0" Nov 25 10:47:41 crc kubenswrapper[4821]: E1125 10:47:41.716793 4821 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 25 10:47:41 crc kubenswrapper[4821]: E1125 10:47:41.717027 4821 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 25 10:47:41 crc kubenswrapper[4821]: E1125 10:47:41.717086 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/0ffb0da2-affb-4a95-87f3-eac442dc754d-etc-swift podName:0ffb0da2-affb-4a95-87f3-eac442dc754d nodeName:}" failed. No retries permitted until 2025-11-25 10:47:45.717070793 +0000 UTC m=+936.253390640 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/0ffb0da2-affb-4a95-87f3-eac442dc754d-etc-swift") pod "swift-storage-0" (UID: "0ffb0da2-affb-4a95-87f3-eac442dc754d") : configmap "swift-ring-files" not found Nov 25 10:47:44 crc kubenswrapper[4821]: I1125 10:47:44.317051 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-7774-account-create-xq572"] Nov 25 10:47:44 crc kubenswrapper[4821]: E1125 10:47:44.318825 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="140a7c0b-b837-43f7-a436-f13bcdd121b9" containerName="init" Nov 25 10:47:44 crc kubenswrapper[4821]: I1125 10:47:44.319085 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="140a7c0b-b837-43f7-a436-f13bcdd121b9" containerName="init" Nov 25 10:47:44 crc kubenswrapper[4821]: E1125 10:47:44.319155 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="140a7c0b-b837-43f7-a436-f13bcdd121b9" containerName="dnsmasq-dns" Nov 25 10:47:44 crc kubenswrapper[4821]: I1125 10:47:44.320301 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="140a7c0b-b837-43f7-a436-f13bcdd121b9" containerName="dnsmasq-dns" Nov 25 10:47:44 crc kubenswrapper[4821]: I1125 10:47:44.320518 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="140a7c0b-b837-43f7-a436-f13bcdd121b9" containerName="dnsmasq-dns" Nov 25 10:47:44 crc kubenswrapper[4821]: I1125 10:47:44.321104 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7774-account-create-xq572" Nov 25 10:47:44 crc kubenswrapper[4821]: I1125 10:47:44.323655 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Nov 25 10:47:44 crc kubenswrapper[4821]: I1125 10:47:44.334082 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7774-account-create-xq572"] Nov 25 10:47:44 crc kubenswrapper[4821]: I1125 10:47:44.382308 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-kt7h7"] Nov 25 10:47:44 crc kubenswrapper[4821]: I1125 10:47:44.383479 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-kt7h7" Nov 25 10:47:44 crc kubenswrapper[4821]: I1125 10:47:44.393917 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-kt7h7"] Nov 25 10:47:44 crc kubenswrapper[4821]: I1125 10:47:44.463901 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4hz8\" (UniqueName: \"kubernetes.io/projected/802c0ca8-b806-4218-86a0-e88b3325a762-kube-api-access-m4hz8\") pod \"keystone-7774-account-create-xq572\" (UID: \"802c0ca8-b806-4218-86a0-e88b3325a762\") " pod="openstack/keystone-7774-account-create-xq572" Nov 25 10:47:44 crc kubenswrapper[4821]: I1125 10:47:44.464000 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tg9cs\" (UniqueName: \"kubernetes.io/projected/f10473c7-e8cd-4534-8232-7cb6a97be557-kube-api-access-tg9cs\") pod \"keystone-db-create-kt7h7\" (UID: \"f10473c7-e8cd-4534-8232-7cb6a97be557\") " pod="openstack/keystone-db-create-kt7h7" Nov 25 10:47:44 crc kubenswrapper[4821]: I1125 10:47:44.464037 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f10473c7-e8cd-4534-8232-7cb6a97be557-operator-scripts\") pod \"keystone-db-create-kt7h7\" (UID: \"f10473c7-e8cd-4534-8232-7cb6a97be557\") " pod="openstack/keystone-db-create-kt7h7" Nov 25 10:47:44 crc kubenswrapper[4821]: I1125 10:47:44.464393 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/802c0ca8-b806-4218-86a0-e88b3325a762-operator-scripts\") pod \"keystone-7774-account-create-xq572\" (UID: \"802c0ca8-b806-4218-86a0-e88b3325a762\") " pod="openstack/keystone-7774-account-create-xq572" Nov 25 10:47:44 crc kubenswrapper[4821]: I1125 10:47:44.566022 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f10473c7-e8cd-4534-8232-7cb6a97be557-operator-scripts\") pod \"keystone-db-create-kt7h7\" (UID: \"f10473c7-e8cd-4534-8232-7cb6a97be557\") " pod="openstack/keystone-db-create-kt7h7" Nov 25 10:47:44 crc kubenswrapper[4821]: I1125 10:47:44.566114 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/802c0ca8-b806-4218-86a0-e88b3325a762-operator-scripts\") pod \"keystone-7774-account-create-xq572\" (UID: \"802c0ca8-b806-4218-86a0-e88b3325a762\") " pod="openstack/keystone-7774-account-create-xq572" Nov 25 10:47:44 crc kubenswrapper[4821]: I1125 10:47:44.566199 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4hz8\" (UniqueName: \"kubernetes.io/projected/802c0ca8-b806-4218-86a0-e88b3325a762-kube-api-access-m4hz8\") pod \"keystone-7774-account-create-xq572\" (UID: \"802c0ca8-b806-4218-86a0-e88b3325a762\") " pod="openstack/keystone-7774-account-create-xq572" Nov 25 10:47:44 crc kubenswrapper[4821]: I1125 10:47:44.566258 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tg9cs\" (UniqueName: \"kubernetes.io/projected/f10473c7-e8cd-4534-8232-7cb6a97be557-kube-api-access-tg9cs\") pod \"keystone-db-create-kt7h7\" (UID: \"f10473c7-e8cd-4534-8232-7cb6a97be557\") " pod="openstack/keystone-db-create-kt7h7" Nov 25 10:47:44 crc kubenswrapper[4821]: I1125 10:47:44.568195 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f10473c7-e8cd-4534-8232-7cb6a97be557-operator-scripts\") pod \"keystone-db-create-kt7h7\" (UID: \"f10473c7-e8cd-4534-8232-7cb6a97be557\") " pod="openstack/keystone-db-create-kt7h7" Nov 25 10:47:44 crc kubenswrapper[4821]: I1125 10:47:44.571020 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/802c0ca8-b806-4218-86a0-e88b3325a762-operator-scripts\") pod \"keystone-7774-account-create-xq572\" (UID: \"802c0ca8-b806-4218-86a0-e88b3325a762\") " pod="openstack/keystone-7774-account-create-xq572" Nov 25 10:47:44 crc kubenswrapper[4821]: I1125 10:47:44.585317 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-kxgp9"] Nov 25 10:47:44 crc kubenswrapper[4821]: I1125 10:47:44.586582 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-kxgp9" Nov 25 10:47:44 crc kubenswrapper[4821]: I1125 10:47:44.593145 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-kxgp9"] Nov 25 10:47:44 crc kubenswrapper[4821]: I1125 10:47:44.594977 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tg9cs\" (UniqueName: \"kubernetes.io/projected/f10473c7-e8cd-4534-8232-7cb6a97be557-kube-api-access-tg9cs\") pod \"keystone-db-create-kt7h7\" (UID: \"f10473c7-e8cd-4534-8232-7cb6a97be557\") " pod="openstack/keystone-db-create-kt7h7" Nov 25 10:47:44 crc kubenswrapper[4821]: I1125 10:47:44.594983 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4hz8\" (UniqueName: \"kubernetes.io/projected/802c0ca8-b806-4218-86a0-e88b3325a762-kube-api-access-m4hz8\") pod \"keystone-7774-account-create-xq572\" (UID: \"802c0ca8-b806-4218-86a0-e88b3325a762\") " pod="openstack/keystone-7774-account-create-xq572" Nov 25 10:47:44 crc kubenswrapper[4821]: I1125 10:47:44.663659 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7774-account-create-xq572" Nov 25 10:47:44 crc kubenswrapper[4821]: I1125 10:47:44.667496 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a9ee739b-0ed7-43aa-b6f1-a779cb718758-operator-scripts\") pod \"placement-db-create-kxgp9\" (UID: \"a9ee739b-0ed7-43aa-b6f1-a779cb718758\") " pod="openstack/placement-db-create-kxgp9" Nov 25 10:47:44 crc kubenswrapper[4821]: I1125 10:47:44.667621 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-48pp5\" (UniqueName: \"kubernetes.io/projected/a9ee739b-0ed7-43aa-b6f1-a779cb718758-kube-api-access-48pp5\") pod \"placement-db-create-kxgp9\" (UID: \"a9ee739b-0ed7-43aa-b6f1-a779cb718758\") " pod="openstack/placement-db-create-kxgp9" Nov 25 10:47:44 crc kubenswrapper[4821]: I1125 10:47:44.688270 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-9c2a-account-create-r76sx"] Nov 25 10:47:44 crc kubenswrapper[4821]: I1125 10:47:44.689321 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-9c2a-account-create-r76sx" Nov 25 10:47:44 crc kubenswrapper[4821]: I1125 10:47:44.691699 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Nov 25 10:47:44 crc kubenswrapper[4821]: I1125 10:47:44.701970 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-kt7h7" Nov 25 10:47:44 crc kubenswrapper[4821]: I1125 10:47:44.702809 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-9c2a-account-create-r76sx"] Nov 25 10:47:44 crc kubenswrapper[4821]: I1125 10:47:44.770511 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a9ee739b-0ed7-43aa-b6f1-a779cb718758-operator-scripts\") pod \"placement-db-create-kxgp9\" (UID: \"a9ee739b-0ed7-43aa-b6f1-a779cb718758\") " pod="openstack/placement-db-create-kxgp9" Nov 25 10:47:44 crc kubenswrapper[4821]: I1125 10:47:44.771106 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j72cd\" (UniqueName: \"kubernetes.io/projected/52803d63-d5ab-42c6-a17e-fcef8e92ea44-kube-api-access-j72cd\") pod \"placement-9c2a-account-create-r76sx\" (UID: \"52803d63-d5ab-42c6-a17e-fcef8e92ea44\") " pod="openstack/placement-9c2a-account-create-r76sx" Nov 25 10:47:44 crc kubenswrapper[4821]: I1125 10:47:44.771188 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/52803d63-d5ab-42c6-a17e-fcef8e92ea44-operator-scripts\") pod \"placement-9c2a-account-create-r76sx\" (UID: \"52803d63-d5ab-42c6-a17e-fcef8e92ea44\") " pod="openstack/placement-9c2a-account-create-r76sx" Nov 25 10:47:44 crc kubenswrapper[4821]: I1125 10:47:44.771278 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-48pp5\" (UniqueName: \"kubernetes.io/projected/a9ee739b-0ed7-43aa-b6f1-a779cb718758-kube-api-access-48pp5\") pod \"placement-db-create-kxgp9\" (UID: \"a9ee739b-0ed7-43aa-b6f1-a779cb718758\") " pod="openstack/placement-db-create-kxgp9" Nov 25 10:47:44 crc kubenswrapper[4821]: I1125 10:47:44.772219 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a9ee739b-0ed7-43aa-b6f1-a779cb718758-operator-scripts\") pod \"placement-db-create-kxgp9\" (UID: \"a9ee739b-0ed7-43aa-b6f1-a779cb718758\") " pod="openstack/placement-db-create-kxgp9" Nov 25 10:47:44 crc kubenswrapper[4821]: I1125 10:47:44.791004 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-48pp5\" (UniqueName: \"kubernetes.io/projected/a9ee739b-0ed7-43aa-b6f1-a779cb718758-kube-api-access-48pp5\") pod \"placement-db-create-kxgp9\" (UID: \"a9ee739b-0ed7-43aa-b6f1-a779cb718758\") " pod="openstack/placement-db-create-kxgp9" Nov 25 10:47:44 crc kubenswrapper[4821]: I1125 10:47:44.822998 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-dzlxn"] Nov 25 10:47:44 crc kubenswrapper[4821]: I1125 10:47:44.824997 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-dzlxn" Nov 25 10:47:44 crc kubenswrapper[4821]: I1125 10:47:44.829709 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-dzlxn"] Nov 25 10:47:44 crc kubenswrapper[4821]: I1125 10:47:44.877453 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j72cd\" (UniqueName: \"kubernetes.io/projected/52803d63-d5ab-42c6-a17e-fcef8e92ea44-kube-api-access-j72cd\") pod \"placement-9c2a-account-create-r76sx\" (UID: \"52803d63-d5ab-42c6-a17e-fcef8e92ea44\") " pod="openstack/placement-9c2a-account-create-r76sx" Nov 25 10:47:44 crc kubenswrapper[4821]: I1125 10:47:44.877534 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/52803d63-d5ab-42c6-a17e-fcef8e92ea44-operator-scripts\") pod \"placement-9c2a-account-create-r76sx\" (UID: \"52803d63-d5ab-42c6-a17e-fcef8e92ea44\") " pod="openstack/placement-9c2a-account-create-r76sx" Nov 25 10:47:44 crc kubenswrapper[4821]: I1125 10:47:44.878348 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/52803d63-d5ab-42c6-a17e-fcef8e92ea44-operator-scripts\") pod \"placement-9c2a-account-create-r76sx\" (UID: \"52803d63-d5ab-42c6-a17e-fcef8e92ea44\") " pod="openstack/placement-9c2a-account-create-r76sx" Nov 25 10:47:44 crc kubenswrapper[4821]: I1125 10:47:44.899276 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j72cd\" (UniqueName: \"kubernetes.io/projected/52803d63-d5ab-42c6-a17e-fcef8e92ea44-kube-api-access-j72cd\") pod \"placement-9c2a-account-create-r76sx\" (UID: \"52803d63-d5ab-42c6-a17e-fcef8e92ea44\") " pod="openstack/placement-9c2a-account-create-r76sx" Nov 25 10:47:44 crc kubenswrapper[4821]: I1125 10:47:44.899350 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-5ef2-account-create-mmmzg"] Nov 25 10:47:44 crc kubenswrapper[4821]: I1125 10:47:44.900394 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-5ef2-account-create-mmmzg" Nov 25 10:47:44 crc kubenswrapper[4821]: I1125 10:47:44.902966 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Nov 25 10:47:44 crc kubenswrapper[4821]: I1125 10:47:44.918105 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-5ef2-account-create-mmmzg"] Nov 25 10:47:44 crc kubenswrapper[4821]: I1125 10:47:44.935073 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-kxgp9" Nov 25 10:47:44 crc kubenswrapper[4821]: I1125 10:47:44.980517 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9528m\" (UniqueName: \"kubernetes.io/projected/7ed5a7f1-1e7b-438a-b083-f7f45b28443a-kube-api-access-9528m\") pod \"glance-db-create-dzlxn\" (UID: \"7ed5a7f1-1e7b-438a-b083-f7f45b28443a\") " pod="openstack/glance-db-create-dzlxn" Nov 25 10:47:44 crc kubenswrapper[4821]: I1125 10:47:44.980789 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0685db28-2cf1-4e13-b6d8-d37983c03d42-operator-scripts\") pod \"glance-5ef2-account-create-mmmzg\" (UID: \"0685db28-2cf1-4e13-b6d8-d37983c03d42\") " pod="openstack/glance-5ef2-account-create-mmmzg" Nov 25 10:47:44 crc kubenswrapper[4821]: I1125 10:47:44.980843 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ed5a7f1-1e7b-438a-b083-f7f45b28443a-operator-scripts\") pod \"glance-db-create-dzlxn\" (UID: \"7ed5a7f1-1e7b-438a-b083-f7f45b28443a\") " pod="openstack/glance-db-create-dzlxn" Nov 25 10:47:44 crc kubenswrapper[4821]: I1125 10:47:44.980946 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-drtsw\" (UniqueName: \"kubernetes.io/projected/0685db28-2cf1-4e13-b6d8-d37983c03d42-kube-api-access-drtsw\") pod \"glance-5ef2-account-create-mmmzg\" (UID: \"0685db28-2cf1-4e13-b6d8-d37983c03d42\") " pod="openstack/glance-5ef2-account-create-mmmzg" Nov 25 10:47:45 crc kubenswrapper[4821]: I1125 10:47:45.082594 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9528m\" (UniqueName: \"kubernetes.io/projected/7ed5a7f1-1e7b-438a-b083-f7f45b28443a-kube-api-access-9528m\") pod \"glance-db-create-dzlxn\" (UID: \"7ed5a7f1-1e7b-438a-b083-f7f45b28443a\") " pod="openstack/glance-db-create-dzlxn" Nov 25 10:47:45 crc kubenswrapper[4821]: I1125 10:47:45.082656 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0685db28-2cf1-4e13-b6d8-d37983c03d42-operator-scripts\") pod \"glance-5ef2-account-create-mmmzg\" (UID: \"0685db28-2cf1-4e13-b6d8-d37983c03d42\") " pod="openstack/glance-5ef2-account-create-mmmzg" Nov 25 10:47:45 crc kubenswrapper[4821]: I1125 10:47:45.082679 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ed5a7f1-1e7b-438a-b083-f7f45b28443a-operator-scripts\") pod \"glance-db-create-dzlxn\" (UID: \"7ed5a7f1-1e7b-438a-b083-f7f45b28443a\") " pod="openstack/glance-db-create-dzlxn" Nov 25 10:47:45 crc kubenswrapper[4821]: I1125 10:47:45.082705 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-drtsw\" (UniqueName: \"kubernetes.io/projected/0685db28-2cf1-4e13-b6d8-d37983c03d42-kube-api-access-drtsw\") pod \"glance-5ef2-account-create-mmmzg\" (UID: \"0685db28-2cf1-4e13-b6d8-d37983c03d42\") " pod="openstack/glance-5ef2-account-create-mmmzg" Nov 25 10:47:45 crc kubenswrapper[4821]: I1125 10:47:45.083478 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0685db28-2cf1-4e13-b6d8-d37983c03d42-operator-scripts\") pod \"glance-5ef2-account-create-mmmzg\" (UID: \"0685db28-2cf1-4e13-b6d8-d37983c03d42\") " pod="openstack/glance-5ef2-account-create-mmmzg" Nov 25 10:47:45 crc kubenswrapper[4821]: I1125 10:47:45.083684 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ed5a7f1-1e7b-438a-b083-f7f45b28443a-operator-scripts\") pod \"glance-db-create-dzlxn\" (UID: \"7ed5a7f1-1e7b-438a-b083-f7f45b28443a\") " pod="openstack/glance-db-create-dzlxn" Nov 25 10:47:45 crc kubenswrapper[4821]: I1125 10:47:45.099298 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9528m\" (UniqueName: \"kubernetes.io/projected/7ed5a7f1-1e7b-438a-b083-f7f45b28443a-kube-api-access-9528m\") pod \"glance-db-create-dzlxn\" (UID: \"7ed5a7f1-1e7b-438a-b083-f7f45b28443a\") " pod="openstack/glance-db-create-dzlxn" Nov 25 10:47:45 crc kubenswrapper[4821]: I1125 10:47:45.111147 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-drtsw\" (UniqueName: \"kubernetes.io/projected/0685db28-2cf1-4e13-b6d8-d37983c03d42-kube-api-access-drtsw\") pod \"glance-5ef2-account-create-mmmzg\" (UID: \"0685db28-2cf1-4e13-b6d8-d37983c03d42\") " pod="openstack/glance-5ef2-account-create-mmmzg" Nov 25 10:47:45 crc kubenswrapper[4821]: I1125 10:47:45.140566 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-9c2a-account-create-r76sx" Nov 25 10:47:45 crc kubenswrapper[4821]: I1125 10:47:45.145943 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-kt7h7"] Nov 25 10:47:45 crc kubenswrapper[4821]: I1125 10:47:45.152240 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-dzlxn" Nov 25 10:47:45 crc kubenswrapper[4821]: W1125 10:47:45.153140 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf10473c7_e8cd_4534_8232_7cb6a97be557.slice/crio-73a980f8874116d8a525d99fbae3ad2bef612e79148ec91802a5f48d975b5a06 WatchSource:0}: Error finding container 73a980f8874116d8a525d99fbae3ad2bef612e79148ec91802a5f48d975b5a06: Status 404 returned error can't find the container with id 73a980f8874116d8a525d99fbae3ad2bef612e79148ec91802a5f48d975b5a06 Nov 25 10:47:45 crc kubenswrapper[4821]: I1125 10:47:45.162113 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7774-account-create-xq572"] Nov 25 10:47:45 crc kubenswrapper[4821]: W1125 10:47:45.183775 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod802c0ca8_b806_4218_86a0_e88b3325a762.slice/crio-e79fd27579ed80f38657046cc5061948106ccee8e9d685d87947695aed89cc80 WatchSource:0}: Error finding container e79fd27579ed80f38657046cc5061948106ccee8e9d685d87947695aed89cc80: Status 404 returned error can't find the container with id e79fd27579ed80f38657046cc5061948106ccee8e9d685d87947695aed89cc80 Nov 25 10:47:45 crc kubenswrapper[4821]: I1125 10:47:45.219608 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-5ef2-account-create-mmmzg" Nov 25 10:47:45 crc kubenswrapper[4821]: I1125 10:47:45.381646 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-kxgp9"] Nov 25 10:47:45 crc kubenswrapper[4821]: W1125 10:47:45.406686 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda9ee739b_0ed7_43aa_b6f1_a779cb718758.slice/crio-a8f86e90ffd1260d19e65e3fff5245b9f89c7d4adee1df176322748222d104e5 WatchSource:0}: Error finding container a8f86e90ffd1260d19e65e3fff5245b9f89c7d4adee1df176322748222d104e5: Status 404 returned error can't find the container with id a8f86e90ffd1260d19e65e3fff5245b9f89c7d4adee1df176322748222d104e5 Nov 25 10:47:45 crc kubenswrapper[4821]: I1125 10:47:45.570377 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-kt7h7" event={"ID":"f10473c7-e8cd-4534-8232-7cb6a97be557","Type":"ContainerStarted","Data":"2c49eaddef93d8519ac175e2b0c354fb919e1f6745d0a5c19c8cb6a79c6a1fc1"} Nov 25 10:47:45 crc kubenswrapper[4821]: I1125 10:47:45.570425 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-kt7h7" event={"ID":"f10473c7-e8cd-4534-8232-7cb6a97be557","Type":"ContainerStarted","Data":"73a980f8874116d8a525d99fbae3ad2bef612e79148ec91802a5f48d975b5a06"} Nov 25 10:47:45 crc kubenswrapper[4821]: I1125 10:47:45.575310 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7774-account-create-xq572" event={"ID":"802c0ca8-b806-4218-86a0-e88b3325a762","Type":"ContainerStarted","Data":"5266b3949d733e0f91f36d6c1b21457c8f847bfd3498f3f8137f27d730d1b9bb"} Nov 25 10:47:45 crc kubenswrapper[4821]: I1125 10:47:45.575354 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7774-account-create-xq572" event={"ID":"802c0ca8-b806-4218-86a0-e88b3325a762","Type":"ContainerStarted","Data":"e79fd27579ed80f38657046cc5061948106ccee8e9d685d87947695aed89cc80"} Nov 25 10:47:45 crc kubenswrapper[4821]: I1125 10:47:45.577336 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-kxgp9" event={"ID":"a9ee739b-0ed7-43aa-b6f1-a779cb718758","Type":"ContainerStarted","Data":"a8f86e90ffd1260d19e65e3fff5245b9f89c7d4adee1df176322748222d104e5"} Nov 25 10:47:45 crc kubenswrapper[4821]: I1125 10:47:45.591919 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-create-kt7h7" podStartSLOduration=1.591900448 podStartE2EDuration="1.591900448s" podCreationTimestamp="2025-11-25 10:47:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:47:45.587336072 +0000 UTC m=+936.123655939" watchObservedRunningTime="2025-11-25 10:47:45.591900448 +0000 UTC m=+936.128220305" Nov 25 10:47:45 crc kubenswrapper[4821]: I1125 10:47:45.618411 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-7774-account-create-xq572" podStartSLOduration=1.61838851 podStartE2EDuration="1.61838851s" podCreationTimestamp="2025-11-25 10:47:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:47:45.607587285 +0000 UTC m=+936.143907142" watchObservedRunningTime="2025-11-25 10:47:45.61838851 +0000 UTC m=+936.154708377" Nov 25 10:47:45 crc kubenswrapper[4821]: I1125 10:47:45.693410 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-dzlxn"] Nov 25 10:47:45 crc kubenswrapper[4821]: I1125 10:47:45.735275 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-9c2a-account-create-r76sx"] Nov 25 10:47:45 crc kubenswrapper[4821]: I1125 10:47:45.742053 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-5ef2-account-create-mmmzg"] Nov 25 10:47:45 crc kubenswrapper[4821]: I1125 10:47:45.794633 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0ffb0da2-affb-4a95-87f3-eac442dc754d-etc-swift\") pod \"swift-storage-0\" (UID: \"0ffb0da2-affb-4a95-87f3-eac442dc754d\") " pod="openstack/swift-storage-0" Nov 25 10:47:45 crc kubenswrapper[4821]: E1125 10:47:45.795101 4821 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 25 10:47:45 crc kubenswrapper[4821]: E1125 10:47:45.795156 4821 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 25 10:47:45 crc kubenswrapper[4821]: E1125 10:47:45.795271 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/0ffb0da2-affb-4a95-87f3-eac442dc754d-etc-swift podName:0ffb0da2-affb-4a95-87f3-eac442dc754d nodeName:}" failed. No retries permitted until 2025-11-25 10:47:53.795240936 +0000 UTC m=+944.331560783 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/0ffb0da2-affb-4a95-87f3-eac442dc754d-etc-swift") pod "swift-storage-0" (UID: "0ffb0da2-affb-4a95-87f3-eac442dc754d") : configmap "swift-ring-files" not found Nov 25 10:47:46 crc kubenswrapper[4821]: I1125 10:47:46.587527 4821 generic.go:334] "Generic (PLEG): container finished" podID="f10473c7-e8cd-4534-8232-7cb6a97be557" containerID="2c49eaddef93d8519ac175e2b0c354fb919e1f6745d0a5c19c8cb6a79c6a1fc1" exitCode=0 Nov 25 10:47:46 crc kubenswrapper[4821]: I1125 10:47:46.587649 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-kt7h7" event={"ID":"f10473c7-e8cd-4534-8232-7cb6a97be557","Type":"ContainerDied","Data":"2c49eaddef93d8519ac175e2b0c354fb919e1f6745d0a5c19c8cb6a79c6a1fc1"} Nov 25 10:47:46 crc kubenswrapper[4821]: I1125 10:47:46.591843 4821 generic.go:334] "Generic (PLEG): container finished" podID="802c0ca8-b806-4218-86a0-e88b3325a762" containerID="5266b3949d733e0f91f36d6c1b21457c8f847bfd3498f3f8137f27d730d1b9bb" exitCode=0 Nov 25 10:47:46 crc kubenswrapper[4821]: I1125 10:47:46.591910 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7774-account-create-xq572" event={"ID":"802c0ca8-b806-4218-86a0-e88b3325a762","Type":"ContainerDied","Data":"5266b3949d733e0f91f36d6c1b21457c8f847bfd3498f3f8137f27d730d1b9bb"} Nov 25 10:47:46 crc kubenswrapper[4821]: I1125 10:47:46.595133 4821 generic.go:334] "Generic (PLEG): container finished" podID="a9ee739b-0ed7-43aa-b6f1-a779cb718758" containerID="5fe0a53af69a3e9008fa221903a96105bc4aeebd297467181649e66df5d458d6" exitCode=0 Nov 25 10:47:46 crc kubenswrapper[4821]: I1125 10:47:46.595270 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-kxgp9" event={"ID":"a9ee739b-0ed7-43aa-b6f1-a779cb718758","Type":"ContainerDied","Data":"5fe0a53af69a3e9008fa221903a96105bc4aeebd297467181649e66df5d458d6"} Nov 25 10:47:47 crc kubenswrapper[4821]: I1125 10:47:47.121195 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-b8fbc5445-wctvh" Nov 25 10:47:47 crc kubenswrapper[4821]: I1125 10:47:47.199914 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-h5qkz"] Nov 25 10:47:47 crc kubenswrapper[4821]: I1125 10:47:47.200788 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-h5qkz" podUID="f289352e-7670-4503-b6bf-ccdaa45a7c36" containerName="dnsmasq-dns" containerID="cri-o://238bc5cdcac84072257aa4c9c41a4c1f1dbcc245eb1007ce0b9326f4683bab66" gracePeriod=10 Nov 25 10:47:47 crc kubenswrapper[4821]: I1125 10:47:47.607418 4821 generic.go:334] "Generic (PLEG): container finished" podID="f289352e-7670-4503-b6bf-ccdaa45a7c36" containerID="238bc5cdcac84072257aa4c9c41a4c1f1dbcc245eb1007ce0b9326f4683bab66" exitCode=0 Nov 25 10:47:47 crc kubenswrapper[4821]: I1125 10:47:47.607555 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-h5qkz" event={"ID":"f289352e-7670-4503-b6bf-ccdaa45a7c36","Type":"ContainerDied","Data":"238bc5cdcac84072257aa4c9c41a4c1f1dbcc245eb1007ce0b9326f4683bab66"} Nov 25 10:47:47 crc kubenswrapper[4821]: I1125 10:47:47.983447 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-kxgp9" Nov 25 10:47:48 crc kubenswrapper[4821]: I1125 10:47:48.149567 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7774-account-create-xq572" Nov 25 10:47:48 crc kubenswrapper[4821]: I1125 10:47:48.150906 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a9ee739b-0ed7-43aa-b6f1-a779cb718758-operator-scripts\") pod \"a9ee739b-0ed7-43aa-b6f1-a779cb718758\" (UID: \"a9ee739b-0ed7-43aa-b6f1-a779cb718758\") " Nov 25 10:47:48 crc kubenswrapper[4821]: I1125 10:47:48.150963 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-48pp5\" (UniqueName: \"kubernetes.io/projected/a9ee739b-0ed7-43aa-b6f1-a779cb718758-kube-api-access-48pp5\") pod \"a9ee739b-0ed7-43aa-b6f1-a779cb718758\" (UID: \"a9ee739b-0ed7-43aa-b6f1-a779cb718758\") " Nov 25 10:47:48 crc kubenswrapper[4821]: I1125 10:47:48.151868 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9ee739b-0ed7-43aa-b6f1-a779cb718758-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a9ee739b-0ed7-43aa-b6f1-a779cb718758" (UID: "a9ee739b-0ed7-43aa-b6f1-a779cb718758"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:47:48 crc kubenswrapper[4821]: I1125 10:47:48.155848 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9ee739b-0ed7-43aa-b6f1-a779cb718758-kube-api-access-48pp5" (OuterVolumeSpecName: "kube-api-access-48pp5") pod "a9ee739b-0ed7-43aa-b6f1-a779cb718758" (UID: "a9ee739b-0ed7-43aa-b6f1-a779cb718758"). InnerVolumeSpecName "kube-api-access-48pp5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:47:48 crc kubenswrapper[4821]: I1125 10:47:48.166963 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-h5qkz" Nov 25 10:47:48 crc kubenswrapper[4821]: I1125 10:47:48.192429 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-kt7h7" Nov 25 10:47:48 crc kubenswrapper[4821]: I1125 10:47:48.252227 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/802c0ca8-b806-4218-86a0-e88b3325a762-operator-scripts\") pod \"802c0ca8-b806-4218-86a0-e88b3325a762\" (UID: \"802c0ca8-b806-4218-86a0-e88b3325a762\") " Nov 25 10:47:48 crc kubenswrapper[4821]: I1125 10:47:48.252447 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m4hz8\" (UniqueName: \"kubernetes.io/projected/802c0ca8-b806-4218-86a0-e88b3325a762-kube-api-access-m4hz8\") pod \"802c0ca8-b806-4218-86a0-e88b3325a762\" (UID: \"802c0ca8-b806-4218-86a0-e88b3325a762\") " Nov 25 10:47:48 crc kubenswrapper[4821]: I1125 10:47:48.252799 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-48pp5\" (UniqueName: \"kubernetes.io/projected/a9ee739b-0ed7-43aa-b6f1-a779cb718758-kube-api-access-48pp5\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:48 crc kubenswrapper[4821]: I1125 10:47:48.252823 4821 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a9ee739b-0ed7-43aa-b6f1-a779cb718758-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:48 crc kubenswrapper[4821]: I1125 10:47:48.253144 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/802c0ca8-b806-4218-86a0-e88b3325a762-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "802c0ca8-b806-4218-86a0-e88b3325a762" (UID: "802c0ca8-b806-4218-86a0-e88b3325a762"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:47:48 crc kubenswrapper[4821]: I1125 10:47:48.256399 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/802c0ca8-b806-4218-86a0-e88b3325a762-kube-api-access-m4hz8" (OuterVolumeSpecName: "kube-api-access-m4hz8") pod "802c0ca8-b806-4218-86a0-e88b3325a762" (UID: "802c0ca8-b806-4218-86a0-e88b3325a762"). InnerVolumeSpecName "kube-api-access-m4hz8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:47:48 crc kubenswrapper[4821]: I1125 10:47:48.354107 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qgs6z\" (UniqueName: \"kubernetes.io/projected/f289352e-7670-4503-b6bf-ccdaa45a7c36-kube-api-access-qgs6z\") pod \"f289352e-7670-4503-b6bf-ccdaa45a7c36\" (UID: \"f289352e-7670-4503-b6bf-ccdaa45a7c36\") " Nov 25 10:47:48 crc kubenswrapper[4821]: I1125 10:47:48.354182 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f10473c7-e8cd-4534-8232-7cb6a97be557-operator-scripts\") pod \"f10473c7-e8cd-4534-8232-7cb6a97be557\" (UID: \"f10473c7-e8cd-4534-8232-7cb6a97be557\") " Nov 25 10:47:48 crc kubenswrapper[4821]: I1125 10:47:48.354313 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f289352e-7670-4503-b6bf-ccdaa45a7c36-config\") pod \"f289352e-7670-4503-b6bf-ccdaa45a7c36\" (UID: \"f289352e-7670-4503-b6bf-ccdaa45a7c36\") " Nov 25 10:47:48 crc kubenswrapper[4821]: I1125 10:47:48.354371 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tg9cs\" (UniqueName: \"kubernetes.io/projected/f10473c7-e8cd-4534-8232-7cb6a97be557-kube-api-access-tg9cs\") pod \"f10473c7-e8cd-4534-8232-7cb6a97be557\" (UID: \"f10473c7-e8cd-4534-8232-7cb6a97be557\") " Nov 25 10:47:48 crc kubenswrapper[4821]: I1125 10:47:48.354438 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f289352e-7670-4503-b6bf-ccdaa45a7c36-dns-svc\") pod \"f289352e-7670-4503-b6bf-ccdaa45a7c36\" (UID: \"f289352e-7670-4503-b6bf-ccdaa45a7c36\") " Nov 25 10:47:48 crc kubenswrapper[4821]: I1125 10:47:48.354777 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m4hz8\" (UniqueName: \"kubernetes.io/projected/802c0ca8-b806-4218-86a0-e88b3325a762-kube-api-access-m4hz8\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:48 crc kubenswrapper[4821]: I1125 10:47:48.354796 4821 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/802c0ca8-b806-4218-86a0-e88b3325a762-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:48 crc kubenswrapper[4821]: I1125 10:47:48.354917 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f10473c7-e8cd-4534-8232-7cb6a97be557-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f10473c7-e8cd-4534-8232-7cb6a97be557" (UID: "f10473c7-e8cd-4534-8232-7cb6a97be557"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:47:48 crc kubenswrapper[4821]: I1125 10:47:48.361691 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f289352e-7670-4503-b6bf-ccdaa45a7c36-kube-api-access-qgs6z" (OuterVolumeSpecName: "kube-api-access-qgs6z") pod "f289352e-7670-4503-b6bf-ccdaa45a7c36" (UID: "f289352e-7670-4503-b6bf-ccdaa45a7c36"). InnerVolumeSpecName "kube-api-access-qgs6z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:47:48 crc kubenswrapper[4821]: I1125 10:47:48.377612 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f10473c7-e8cd-4534-8232-7cb6a97be557-kube-api-access-tg9cs" (OuterVolumeSpecName: "kube-api-access-tg9cs") pod "f10473c7-e8cd-4534-8232-7cb6a97be557" (UID: "f10473c7-e8cd-4534-8232-7cb6a97be557"). InnerVolumeSpecName "kube-api-access-tg9cs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:47:48 crc kubenswrapper[4821]: I1125 10:47:48.405693 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f289352e-7670-4503-b6bf-ccdaa45a7c36-config" (OuterVolumeSpecName: "config") pod "f289352e-7670-4503-b6bf-ccdaa45a7c36" (UID: "f289352e-7670-4503-b6bf-ccdaa45a7c36"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:47:48 crc kubenswrapper[4821]: I1125 10:47:48.437617 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f289352e-7670-4503-b6bf-ccdaa45a7c36-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f289352e-7670-4503-b6bf-ccdaa45a7c36" (UID: "f289352e-7670-4503-b6bf-ccdaa45a7c36"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:47:48 crc kubenswrapper[4821]: I1125 10:47:48.456967 4821 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f289352e-7670-4503-b6bf-ccdaa45a7c36-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:48 crc kubenswrapper[4821]: I1125 10:47:48.457015 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tg9cs\" (UniqueName: \"kubernetes.io/projected/f10473c7-e8cd-4534-8232-7cb6a97be557-kube-api-access-tg9cs\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:48 crc kubenswrapper[4821]: I1125 10:47:48.457055 4821 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f289352e-7670-4503-b6bf-ccdaa45a7c36-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:48 crc kubenswrapper[4821]: I1125 10:47:48.457072 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qgs6z\" (UniqueName: \"kubernetes.io/projected/f289352e-7670-4503-b6bf-ccdaa45a7c36-kube-api-access-qgs6z\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:48 crc kubenswrapper[4821]: I1125 10:47:48.457084 4821 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f10473c7-e8cd-4534-8232-7cb6a97be557-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:48 crc kubenswrapper[4821]: I1125 10:47:48.616030 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-gmrpw" event={"ID":"1aa8233d-1fe4-44ec-acc6-0d5548c70d59","Type":"ContainerStarted","Data":"ae1c0c61500eed63e791b99b32ff19038072001ce79a8b7b8ac11dacb158e692"} Nov 25 10:47:48 crc kubenswrapper[4821]: I1125 10:47:48.619611 4821 generic.go:334] "Generic (PLEG): container finished" podID="0685db28-2cf1-4e13-b6d8-d37983c03d42" containerID="8e6ed8fd87660ebcfb7743265992ee1ef1b5540b8654a15cbd85c05a88d48d39" exitCode=0 Nov 25 10:47:48 crc kubenswrapper[4821]: I1125 10:47:48.619711 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-5ef2-account-create-mmmzg" event={"ID":"0685db28-2cf1-4e13-b6d8-d37983c03d42","Type":"ContainerDied","Data":"8e6ed8fd87660ebcfb7743265992ee1ef1b5540b8654a15cbd85c05a88d48d39"} Nov 25 10:47:48 crc kubenswrapper[4821]: I1125 10:47:48.619742 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-5ef2-account-create-mmmzg" event={"ID":"0685db28-2cf1-4e13-b6d8-d37983c03d42","Type":"ContainerStarted","Data":"9e543f846fe1ee4d33ac5f0d4a25b8fe50b6807fb7ac7a971cc4e4e81568ee40"} Nov 25 10:47:48 crc kubenswrapper[4821]: I1125 10:47:48.623100 4821 generic.go:334] "Generic (PLEG): container finished" podID="52803d63-d5ab-42c6-a17e-fcef8e92ea44" containerID="f939ace9603f0998d01d1c273b0f949305cc4aac5343cf21417718d0ccc3c1b0" exitCode=0 Nov 25 10:47:48 crc kubenswrapper[4821]: I1125 10:47:48.623186 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-9c2a-account-create-r76sx" event={"ID":"52803d63-d5ab-42c6-a17e-fcef8e92ea44","Type":"ContainerDied","Data":"f939ace9603f0998d01d1c273b0f949305cc4aac5343cf21417718d0ccc3c1b0"} Nov 25 10:47:48 crc kubenswrapper[4821]: I1125 10:47:48.623210 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-9c2a-account-create-r76sx" event={"ID":"52803d63-d5ab-42c6-a17e-fcef8e92ea44","Type":"ContainerStarted","Data":"c918a933bcd2b403c0212a2607cf002dfd4855ad921d999fede80208b4a4af4c"} Nov 25 10:47:48 crc kubenswrapper[4821]: I1125 10:47:48.625367 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-kt7h7" Nov 25 10:47:48 crc kubenswrapper[4821]: I1125 10:47:48.625390 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-kt7h7" event={"ID":"f10473c7-e8cd-4534-8232-7cb6a97be557","Type":"ContainerDied","Data":"73a980f8874116d8a525d99fbae3ad2bef612e79148ec91802a5f48d975b5a06"} Nov 25 10:47:48 crc kubenswrapper[4821]: I1125 10:47:48.625499 4821 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="73a980f8874116d8a525d99fbae3ad2bef612e79148ec91802a5f48d975b5a06" Nov 25 10:47:48 crc kubenswrapper[4821]: I1125 10:47:48.627511 4821 generic.go:334] "Generic (PLEG): container finished" podID="7ed5a7f1-1e7b-438a-b083-f7f45b28443a" containerID="b4c4327cc6c0a7cd77d3806069faf04d3f5c4c2d18f990f2d67b0909e777e4e8" exitCode=0 Nov 25 10:47:48 crc kubenswrapper[4821]: I1125 10:47:48.627587 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-dzlxn" event={"ID":"7ed5a7f1-1e7b-438a-b083-f7f45b28443a","Type":"ContainerDied","Data":"b4c4327cc6c0a7cd77d3806069faf04d3f5c4c2d18f990f2d67b0909e777e4e8"} Nov 25 10:47:48 crc kubenswrapper[4821]: I1125 10:47:48.627609 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-dzlxn" event={"ID":"7ed5a7f1-1e7b-438a-b083-f7f45b28443a","Type":"ContainerStarted","Data":"42e0771652147b006d205f16c2e21b614a2f51d482ecf246a1998946a407932e"} Nov 25 10:47:48 crc kubenswrapper[4821]: I1125 10:47:48.629845 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7774-account-create-xq572" event={"ID":"802c0ca8-b806-4218-86a0-e88b3325a762","Type":"ContainerDied","Data":"e79fd27579ed80f38657046cc5061948106ccee8e9d685d87947695aed89cc80"} Nov 25 10:47:48 crc kubenswrapper[4821]: I1125 10:47:48.629873 4821 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e79fd27579ed80f38657046cc5061948106ccee8e9d685d87947695aed89cc80" Nov 25 10:47:48 crc kubenswrapper[4821]: I1125 10:47:48.629997 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7774-account-create-xq572" Nov 25 10:47:48 crc kubenswrapper[4821]: I1125 10:47:48.631574 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-h5qkz" Nov 25 10:47:48 crc kubenswrapper[4821]: I1125 10:47:48.631606 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-h5qkz" event={"ID":"f289352e-7670-4503-b6bf-ccdaa45a7c36","Type":"ContainerDied","Data":"a08708b9197edf8c74a04f9d6c70e52120c11bdcb6e9d4c4e246a60c04eb6ddd"} Nov 25 10:47:48 crc kubenswrapper[4821]: I1125 10:47:48.631692 4821 scope.go:117] "RemoveContainer" containerID="238bc5cdcac84072257aa4c9c41a4c1f1dbcc245eb1007ce0b9326f4683bab66" Nov 25 10:47:48 crc kubenswrapper[4821]: I1125 10:47:48.633245 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-kxgp9" event={"ID":"a9ee739b-0ed7-43aa-b6f1-a779cb718758","Type":"ContainerDied","Data":"a8f86e90ffd1260d19e65e3fff5245b9f89c7d4adee1df176322748222d104e5"} Nov 25 10:47:48 crc kubenswrapper[4821]: I1125 10:47:48.633264 4821 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a8f86e90ffd1260d19e65e3fff5245b9f89c7d4adee1df176322748222d104e5" Nov 25 10:47:48 crc kubenswrapper[4821]: I1125 10:47:48.633311 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-kxgp9" Nov 25 10:47:48 crc kubenswrapper[4821]: I1125 10:47:48.677275 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-gmrpw" podStartSLOduration=2.069614312 podStartE2EDuration="10.677253005s" podCreationTimestamp="2025-11-25 10:47:38 +0000 UTC" firstStartedPulling="2025-11-25 10:47:39.239106774 +0000 UTC m=+929.775426621" lastFinishedPulling="2025-11-25 10:47:47.846745467 +0000 UTC m=+938.383065314" observedRunningTime="2025-11-25 10:47:48.640964045 +0000 UTC m=+939.177283902" watchObservedRunningTime="2025-11-25 10:47:48.677253005 +0000 UTC m=+939.213572862" Nov 25 10:47:48 crc kubenswrapper[4821]: I1125 10:47:48.729724 4821 scope.go:117] "RemoveContainer" containerID="d0b34e2120b350fb56b94467176b871e29f1dd028a34ba33b0362146aef3e828" Nov 25 10:47:48 crc kubenswrapper[4821]: I1125 10:47:48.741375 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-h5qkz"] Nov 25 10:47:48 crc kubenswrapper[4821]: I1125 10:47:48.747485 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-h5qkz"] Nov 25 10:47:49 crc kubenswrapper[4821]: I1125 10:47:49.129648 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Nov 25 10:47:50 crc kubenswrapper[4821]: I1125 10:47:50.123293 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-9c2a-account-create-r76sx" Nov 25 10:47:50 crc kubenswrapper[4821]: I1125 10:47:50.162290 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f289352e-7670-4503-b6bf-ccdaa45a7c36" path="/var/lib/kubelet/pods/f289352e-7670-4503-b6bf-ccdaa45a7c36/volumes" Nov 25 10:47:50 crc kubenswrapper[4821]: I1125 10:47:50.179935 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-dzlxn" Nov 25 10:47:50 crc kubenswrapper[4821]: I1125 10:47:50.181951 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-5ef2-account-create-mmmzg" Nov 25 10:47:50 crc kubenswrapper[4821]: I1125 10:47:50.288732 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ed5a7f1-1e7b-438a-b083-f7f45b28443a-operator-scripts\") pod \"7ed5a7f1-1e7b-438a-b083-f7f45b28443a\" (UID: \"7ed5a7f1-1e7b-438a-b083-f7f45b28443a\") " Nov 25 10:47:50 crc kubenswrapper[4821]: I1125 10:47:50.288984 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9528m\" (UniqueName: \"kubernetes.io/projected/7ed5a7f1-1e7b-438a-b083-f7f45b28443a-kube-api-access-9528m\") pod \"7ed5a7f1-1e7b-438a-b083-f7f45b28443a\" (UID: \"7ed5a7f1-1e7b-438a-b083-f7f45b28443a\") " Nov 25 10:47:50 crc kubenswrapper[4821]: I1125 10:47:50.289061 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-drtsw\" (UniqueName: \"kubernetes.io/projected/0685db28-2cf1-4e13-b6d8-d37983c03d42-kube-api-access-drtsw\") pod \"0685db28-2cf1-4e13-b6d8-d37983c03d42\" (UID: \"0685db28-2cf1-4e13-b6d8-d37983c03d42\") " Nov 25 10:47:50 crc kubenswrapper[4821]: I1125 10:47:50.289124 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/52803d63-d5ab-42c6-a17e-fcef8e92ea44-operator-scripts\") pod \"52803d63-d5ab-42c6-a17e-fcef8e92ea44\" (UID: \"52803d63-d5ab-42c6-a17e-fcef8e92ea44\") " Nov 25 10:47:50 crc kubenswrapper[4821]: I1125 10:47:50.289197 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0685db28-2cf1-4e13-b6d8-d37983c03d42-operator-scripts\") pod \"0685db28-2cf1-4e13-b6d8-d37983c03d42\" (UID: \"0685db28-2cf1-4e13-b6d8-d37983c03d42\") " Nov 25 10:47:50 crc kubenswrapper[4821]: I1125 10:47:50.289281 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j72cd\" (UniqueName: \"kubernetes.io/projected/52803d63-d5ab-42c6-a17e-fcef8e92ea44-kube-api-access-j72cd\") pod \"52803d63-d5ab-42c6-a17e-fcef8e92ea44\" (UID: \"52803d63-d5ab-42c6-a17e-fcef8e92ea44\") " Nov 25 10:47:50 crc kubenswrapper[4821]: I1125 10:47:50.289928 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0685db28-2cf1-4e13-b6d8-d37983c03d42-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0685db28-2cf1-4e13-b6d8-d37983c03d42" (UID: "0685db28-2cf1-4e13-b6d8-d37983c03d42"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:47:50 crc kubenswrapper[4821]: I1125 10:47:50.291087 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ed5a7f1-1e7b-438a-b083-f7f45b28443a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7ed5a7f1-1e7b-438a-b083-f7f45b28443a" (UID: "7ed5a7f1-1e7b-438a-b083-f7f45b28443a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:47:50 crc kubenswrapper[4821]: I1125 10:47:50.291236 4821 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0685db28-2cf1-4e13-b6d8-d37983c03d42-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:50 crc kubenswrapper[4821]: I1125 10:47:50.293673 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/52803d63-d5ab-42c6-a17e-fcef8e92ea44-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "52803d63-d5ab-42c6-a17e-fcef8e92ea44" (UID: "52803d63-d5ab-42c6-a17e-fcef8e92ea44"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:47:50 crc kubenswrapper[4821]: I1125 10:47:50.300673 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ed5a7f1-1e7b-438a-b083-f7f45b28443a-kube-api-access-9528m" (OuterVolumeSpecName: "kube-api-access-9528m") pod "7ed5a7f1-1e7b-438a-b083-f7f45b28443a" (UID: "7ed5a7f1-1e7b-438a-b083-f7f45b28443a"). InnerVolumeSpecName "kube-api-access-9528m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:47:50 crc kubenswrapper[4821]: I1125 10:47:50.300711 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0685db28-2cf1-4e13-b6d8-d37983c03d42-kube-api-access-drtsw" (OuterVolumeSpecName: "kube-api-access-drtsw") pod "0685db28-2cf1-4e13-b6d8-d37983c03d42" (UID: "0685db28-2cf1-4e13-b6d8-d37983c03d42"). InnerVolumeSpecName "kube-api-access-drtsw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:47:50 crc kubenswrapper[4821]: I1125 10:47:50.300882 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52803d63-d5ab-42c6-a17e-fcef8e92ea44-kube-api-access-j72cd" (OuterVolumeSpecName: "kube-api-access-j72cd") pod "52803d63-d5ab-42c6-a17e-fcef8e92ea44" (UID: "52803d63-d5ab-42c6-a17e-fcef8e92ea44"). InnerVolumeSpecName "kube-api-access-j72cd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:47:50 crc kubenswrapper[4821]: I1125 10:47:50.392578 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9528m\" (UniqueName: \"kubernetes.io/projected/7ed5a7f1-1e7b-438a-b083-f7f45b28443a-kube-api-access-9528m\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:50 crc kubenswrapper[4821]: I1125 10:47:50.392625 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-drtsw\" (UniqueName: \"kubernetes.io/projected/0685db28-2cf1-4e13-b6d8-d37983c03d42-kube-api-access-drtsw\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:50 crc kubenswrapper[4821]: I1125 10:47:50.392638 4821 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/52803d63-d5ab-42c6-a17e-fcef8e92ea44-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:50 crc kubenswrapper[4821]: I1125 10:47:50.392651 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j72cd\" (UniqueName: \"kubernetes.io/projected/52803d63-d5ab-42c6-a17e-fcef8e92ea44-kube-api-access-j72cd\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:50 crc kubenswrapper[4821]: I1125 10:47:50.392664 4821 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ed5a7f1-1e7b-438a-b083-f7f45b28443a-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:50 crc kubenswrapper[4821]: I1125 10:47:50.658700 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-5ef2-account-create-mmmzg" event={"ID":"0685db28-2cf1-4e13-b6d8-d37983c03d42","Type":"ContainerDied","Data":"9e543f846fe1ee4d33ac5f0d4a25b8fe50b6807fb7ac7a971cc4e4e81568ee40"} Nov 25 10:47:50 crc kubenswrapper[4821]: I1125 10:47:50.658965 4821 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9e543f846fe1ee4d33ac5f0d4a25b8fe50b6807fb7ac7a971cc4e4e81568ee40" Nov 25 10:47:50 crc kubenswrapper[4821]: I1125 10:47:50.659119 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-5ef2-account-create-mmmzg" Nov 25 10:47:50 crc kubenswrapper[4821]: I1125 10:47:50.692830 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-9c2a-account-create-r76sx" event={"ID":"52803d63-d5ab-42c6-a17e-fcef8e92ea44","Type":"ContainerDied","Data":"c918a933bcd2b403c0212a2607cf002dfd4855ad921d999fede80208b4a4af4c"} Nov 25 10:47:50 crc kubenswrapper[4821]: I1125 10:47:50.692882 4821 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c918a933bcd2b403c0212a2607cf002dfd4855ad921d999fede80208b4a4af4c" Nov 25 10:47:50 crc kubenswrapper[4821]: I1125 10:47:50.692935 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-9c2a-account-create-r76sx" Nov 25 10:47:50 crc kubenswrapper[4821]: I1125 10:47:50.706706 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-dzlxn" event={"ID":"7ed5a7f1-1e7b-438a-b083-f7f45b28443a","Type":"ContainerDied","Data":"42e0771652147b006d205f16c2e21b614a2f51d482ecf246a1998946a407932e"} Nov 25 10:47:50 crc kubenswrapper[4821]: I1125 10:47:50.706753 4821 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="42e0771652147b006d205f16c2e21b614a2f51d482ecf246a1998946a407932e" Nov 25 10:47:50 crc kubenswrapper[4821]: I1125 10:47:50.706819 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-dzlxn" Nov 25 10:47:53 crc kubenswrapper[4821]: I1125 10:47:53.861191 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0ffb0da2-affb-4a95-87f3-eac442dc754d-etc-swift\") pod \"swift-storage-0\" (UID: \"0ffb0da2-affb-4a95-87f3-eac442dc754d\") " pod="openstack/swift-storage-0" Nov 25 10:47:53 crc kubenswrapper[4821]: E1125 10:47:53.861381 4821 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 25 10:47:53 crc kubenswrapper[4821]: E1125 10:47:53.861577 4821 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 25 10:47:53 crc kubenswrapper[4821]: E1125 10:47:53.861635 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/0ffb0da2-affb-4a95-87f3-eac442dc754d-etc-swift podName:0ffb0da2-affb-4a95-87f3-eac442dc754d nodeName:}" failed. No retries permitted until 2025-11-25 10:48:09.861616828 +0000 UTC m=+960.397936675 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/0ffb0da2-affb-4a95-87f3-eac442dc754d-etc-swift") pod "swift-storage-0" (UID: "0ffb0da2-affb-4a95-87f3-eac442dc754d") : configmap "swift-ring-files" not found Nov 25 10:47:55 crc kubenswrapper[4821]: I1125 10:47:55.062129 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-2nhpq"] Nov 25 10:47:55 crc kubenswrapper[4821]: E1125 10:47:55.062479 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f289352e-7670-4503-b6bf-ccdaa45a7c36" containerName="init" Nov 25 10:47:55 crc kubenswrapper[4821]: I1125 10:47:55.062492 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="f289352e-7670-4503-b6bf-ccdaa45a7c36" containerName="init" Nov 25 10:47:55 crc kubenswrapper[4821]: E1125 10:47:55.062504 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f10473c7-e8cd-4534-8232-7cb6a97be557" containerName="mariadb-database-create" Nov 25 10:47:55 crc kubenswrapper[4821]: I1125 10:47:55.062509 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="f10473c7-e8cd-4534-8232-7cb6a97be557" containerName="mariadb-database-create" Nov 25 10:47:55 crc kubenswrapper[4821]: E1125 10:47:55.062527 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9ee739b-0ed7-43aa-b6f1-a779cb718758" containerName="mariadb-database-create" Nov 25 10:47:55 crc kubenswrapper[4821]: I1125 10:47:55.062534 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9ee739b-0ed7-43aa-b6f1-a779cb718758" containerName="mariadb-database-create" Nov 25 10:47:55 crc kubenswrapper[4821]: E1125 10:47:55.062543 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f289352e-7670-4503-b6bf-ccdaa45a7c36" containerName="dnsmasq-dns" Nov 25 10:47:55 crc kubenswrapper[4821]: I1125 10:47:55.062548 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="f289352e-7670-4503-b6bf-ccdaa45a7c36" containerName="dnsmasq-dns" Nov 25 10:47:55 crc kubenswrapper[4821]: E1125 10:47:55.062557 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="802c0ca8-b806-4218-86a0-e88b3325a762" containerName="mariadb-account-create" Nov 25 10:47:55 crc kubenswrapper[4821]: I1125 10:47:55.062563 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="802c0ca8-b806-4218-86a0-e88b3325a762" containerName="mariadb-account-create" Nov 25 10:47:55 crc kubenswrapper[4821]: E1125 10:47:55.062577 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0685db28-2cf1-4e13-b6d8-d37983c03d42" containerName="mariadb-account-create" Nov 25 10:47:55 crc kubenswrapper[4821]: I1125 10:47:55.062583 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="0685db28-2cf1-4e13-b6d8-d37983c03d42" containerName="mariadb-account-create" Nov 25 10:47:55 crc kubenswrapper[4821]: E1125 10:47:55.062597 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ed5a7f1-1e7b-438a-b083-f7f45b28443a" containerName="mariadb-database-create" Nov 25 10:47:55 crc kubenswrapper[4821]: I1125 10:47:55.062602 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ed5a7f1-1e7b-438a-b083-f7f45b28443a" containerName="mariadb-database-create" Nov 25 10:47:55 crc kubenswrapper[4821]: E1125 10:47:55.062614 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52803d63-d5ab-42c6-a17e-fcef8e92ea44" containerName="mariadb-account-create" Nov 25 10:47:55 crc kubenswrapper[4821]: I1125 10:47:55.062619 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="52803d63-d5ab-42c6-a17e-fcef8e92ea44" containerName="mariadb-account-create" Nov 25 10:47:55 crc kubenswrapper[4821]: I1125 10:47:55.062762 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ed5a7f1-1e7b-438a-b083-f7f45b28443a" containerName="mariadb-database-create" Nov 25 10:47:55 crc kubenswrapper[4821]: I1125 10:47:55.062772 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="802c0ca8-b806-4218-86a0-e88b3325a762" containerName="mariadb-account-create" Nov 25 10:47:55 crc kubenswrapper[4821]: I1125 10:47:55.062783 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="f289352e-7670-4503-b6bf-ccdaa45a7c36" containerName="dnsmasq-dns" Nov 25 10:47:55 crc kubenswrapper[4821]: I1125 10:47:55.062794 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="f10473c7-e8cd-4534-8232-7cb6a97be557" containerName="mariadb-database-create" Nov 25 10:47:55 crc kubenswrapper[4821]: I1125 10:47:55.062803 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9ee739b-0ed7-43aa-b6f1-a779cb718758" containerName="mariadb-database-create" Nov 25 10:47:55 crc kubenswrapper[4821]: I1125 10:47:55.062819 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="0685db28-2cf1-4e13-b6d8-d37983c03d42" containerName="mariadb-account-create" Nov 25 10:47:55 crc kubenswrapper[4821]: I1125 10:47:55.062828 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="52803d63-d5ab-42c6-a17e-fcef8e92ea44" containerName="mariadb-account-create" Nov 25 10:47:55 crc kubenswrapper[4821]: I1125 10:47:55.063560 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-2nhpq" Nov 25 10:47:55 crc kubenswrapper[4821]: I1125 10:47:55.065509 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Nov 25 10:47:55 crc kubenswrapper[4821]: I1125 10:47:55.066316 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-lqdn5" Nov 25 10:47:55 crc kubenswrapper[4821]: I1125 10:47:55.073451 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-2nhpq"] Nov 25 10:47:55 crc kubenswrapper[4821]: I1125 10:47:55.183691 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8123055-1953-42b3-9c04-716061499d03-config-data\") pod \"glance-db-sync-2nhpq\" (UID: \"a8123055-1953-42b3-9c04-716061499d03\") " pod="openstack/glance-db-sync-2nhpq" Nov 25 10:47:55 crc kubenswrapper[4821]: I1125 10:47:55.183749 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a8123055-1953-42b3-9c04-716061499d03-db-sync-config-data\") pod \"glance-db-sync-2nhpq\" (UID: \"a8123055-1953-42b3-9c04-716061499d03\") " pod="openstack/glance-db-sync-2nhpq" Nov 25 10:47:55 crc kubenswrapper[4821]: I1125 10:47:55.184687 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cgxhq\" (UniqueName: \"kubernetes.io/projected/a8123055-1953-42b3-9c04-716061499d03-kube-api-access-cgxhq\") pod \"glance-db-sync-2nhpq\" (UID: \"a8123055-1953-42b3-9c04-716061499d03\") " pod="openstack/glance-db-sync-2nhpq" Nov 25 10:47:55 crc kubenswrapper[4821]: I1125 10:47:55.184824 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8123055-1953-42b3-9c04-716061499d03-combined-ca-bundle\") pod \"glance-db-sync-2nhpq\" (UID: \"a8123055-1953-42b3-9c04-716061499d03\") " pod="openstack/glance-db-sync-2nhpq" Nov 25 10:47:55 crc kubenswrapper[4821]: I1125 10:47:55.286736 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cgxhq\" (UniqueName: \"kubernetes.io/projected/a8123055-1953-42b3-9c04-716061499d03-kube-api-access-cgxhq\") pod \"glance-db-sync-2nhpq\" (UID: \"a8123055-1953-42b3-9c04-716061499d03\") " pod="openstack/glance-db-sync-2nhpq" Nov 25 10:47:55 crc kubenswrapper[4821]: I1125 10:47:55.286812 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8123055-1953-42b3-9c04-716061499d03-combined-ca-bundle\") pod \"glance-db-sync-2nhpq\" (UID: \"a8123055-1953-42b3-9c04-716061499d03\") " pod="openstack/glance-db-sync-2nhpq" Nov 25 10:47:55 crc kubenswrapper[4821]: I1125 10:47:55.286907 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8123055-1953-42b3-9c04-716061499d03-config-data\") pod \"glance-db-sync-2nhpq\" (UID: \"a8123055-1953-42b3-9c04-716061499d03\") " pod="openstack/glance-db-sync-2nhpq" Nov 25 10:47:55 crc kubenswrapper[4821]: I1125 10:47:55.286953 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a8123055-1953-42b3-9c04-716061499d03-db-sync-config-data\") pod \"glance-db-sync-2nhpq\" (UID: \"a8123055-1953-42b3-9c04-716061499d03\") " pod="openstack/glance-db-sync-2nhpq" Nov 25 10:47:55 crc kubenswrapper[4821]: I1125 10:47:55.293275 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a8123055-1953-42b3-9c04-716061499d03-db-sync-config-data\") pod \"glance-db-sync-2nhpq\" (UID: \"a8123055-1953-42b3-9c04-716061499d03\") " pod="openstack/glance-db-sync-2nhpq" Nov 25 10:47:55 crc kubenswrapper[4821]: I1125 10:47:55.294219 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8123055-1953-42b3-9c04-716061499d03-config-data\") pod \"glance-db-sync-2nhpq\" (UID: \"a8123055-1953-42b3-9c04-716061499d03\") " pod="openstack/glance-db-sync-2nhpq" Nov 25 10:47:55 crc kubenswrapper[4821]: I1125 10:47:55.301747 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8123055-1953-42b3-9c04-716061499d03-combined-ca-bundle\") pod \"glance-db-sync-2nhpq\" (UID: \"a8123055-1953-42b3-9c04-716061499d03\") " pod="openstack/glance-db-sync-2nhpq" Nov 25 10:47:55 crc kubenswrapper[4821]: I1125 10:47:55.303260 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cgxhq\" (UniqueName: \"kubernetes.io/projected/a8123055-1953-42b3-9c04-716061499d03-kube-api-access-cgxhq\") pod \"glance-db-sync-2nhpq\" (UID: \"a8123055-1953-42b3-9c04-716061499d03\") " pod="openstack/glance-db-sync-2nhpq" Nov 25 10:47:55 crc kubenswrapper[4821]: I1125 10:47:55.416501 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-2nhpq" Nov 25 10:47:55 crc kubenswrapper[4821]: I1125 10:47:55.752087 4821 generic.go:334] "Generic (PLEG): container finished" podID="1aa8233d-1fe4-44ec-acc6-0d5548c70d59" containerID="ae1c0c61500eed63e791b99b32ff19038072001ce79a8b7b8ac11dacb158e692" exitCode=0 Nov 25 10:47:55 crc kubenswrapper[4821]: I1125 10:47:55.752415 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-gmrpw" event={"ID":"1aa8233d-1fe4-44ec-acc6-0d5548c70d59","Type":"ContainerDied","Data":"ae1c0c61500eed63e791b99b32ff19038072001ce79a8b7b8ac11dacb158e692"} Nov 25 10:47:55 crc kubenswrapper[4821]: W1125 10:47:55.926989 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda8123055_1953_42b3_9c04_716061499d03.slice/crio-7bb6c92ed8e23690d9e6b3c4ba5501456d37cfed8991b3e2f91c3a6509bf8b58 WatchSource:0}: Error finding container 7bb6c92ed8e23690d9e6b3c4ba5501456d37cfed8991b3e2f91c3a6509bf8b58: Status 404 returned error can't find the container with id 7bb6c92ed8e23690d9e6b3c4ba5501456d37cfed8991b3e2f91c3a6509bf8b58 Nov 25 10:47:55 crc kubenswrapper[4821]: I1125 10:47:55.930876 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-2nhpq"] Nov 25 10:47:56 crc kubenswrapper[4821]: I1125 10:47:56.761839 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-2nhpq" event={"ID":"a8123055-1953-42b3-9c04-716061499d03","Type":"ContainerStarted","Data":"7bb6c92ed8e23690d9e6b3c4ba5501456d37cfed8991b3e2f91c3a6509bf8b58"} Nov 25 10:47:57 crc kubenswrapper[4821]: I1125 10:47:57.108938 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-gmrpw" Nov 25 10:47:57 crc kubenswrapper[4821]: I1125 10:47:57.215690 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1aa8233d-1fe4-44ec-acc6-0d5548c70d59-etc-swift\") pod \"1aa8233d-1fe4-44ec-acc6-0d5548c70d59\" (UID: \"1aa8233d-1fe4-44ec-acc6-0d5548c70d59\") " Nov 25 10:47:57 crc kubenswrapper[4821]: I1125 10:47:57.215809 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-php28\" (UniqueName: \"kubernetes.io/projected/1aa8233d-1fe4-44ec-acc6-0d5548c70d59-kube-api-access-php28\") pod \"1aa8233d-1fe4-44ec-acc6-0d5548c70d59\" (UID: \"1aa8233d-1fe4-44ec-acc6-0d5548c70d59\") " Nov 25 10:47:57 crc kubenswrapper[4821]: I1125 10:47:57.215834 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1aa8233d-1fe4-44ec-acc6-0d5548c70d59-swiftconf\") pod \"1aa8233d-1fe4-44ec-acc6-0d5548c70d59\" (UID: \"1aa8233d-1fe4-44ec-acc6-0d5548c70d59\") " Nov 25 10:47:57 crc kubenswrapper[4821]: I1125 10:47:57.215881 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1aa8233d-1fe4-44ec-acc6-0d5548c70d59-scripts\") pod \"1aa8233d-1fe4-44ec-acc6-0d5548c70d59\" (UID: \"1aa8233d-1fe4-44ec-acc6-0d5548c70d59\") " Nov 25 10:47:57 crc kubenswrapper[4821]: I1125 10:47:57.215922 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1aa8233d-1fe4-44ec-acc6-0d5548c70d59-combined-ca-bundle\") pod \"1aa8233d-1fe4-44ec-acc6-0d5548c70d59\" (UID: \"1aa8233d-1fe4-44ec-acc6-0d5548c70d59\") " Nov 25 10:47:57 crc kubenswrapper[4821]: I1125 10:47:57.215946 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1aa8233d-1fe4-44ec-acc6-0d5548c70d59-ring-data-devices\") pod \"1aa8233d-1fe4-44ec-acc6-0d5548c70d59\" (UID: \"1aa8233d-1fe4-44ec-acc6-0d5548c70d59\") " Nov 25 10:47:57 crc kubenswrapper[4821]: I1125 10:47:57.215989 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1aa8233d-1fe4-44ec-acc6-0d5548c70d59-dispersionconf\") pod \"1aa8233d-1fe4-44ec-acc6-0d5548c70d59\" (UID: \"1aa8233d-1fe4-44ec-acc6-0d5548c70d59\") " Nov 25 10:47:57 crc kubenswrapper[4821]: I1125 10:47:57.218423 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1aa8233d-1fe4-44ec-acc6-0d5548c70d59-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "1aa8233d-1fe4-44ec-acc6-0d5548c70d59" (UID: "1aa8233d-1fe4-44ec-acc6-0d5548c70d59"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:47:57 crc kubenswrapper[4821]: I1125 10:47:57.224591 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1aa8233d-1fe4-44ec-acc6-0d5548c70d59-kube-api-access-php28" (OuterVolumeSpecName: "kube-api-access-php28") pod "1aa8233d-1fe4-44ec-acc6-0d5548c70d59" (UID: "1aa8233d-1fe4-44ec-acc6-0d5548c70d59"). InnerVolumeSpecName "kube-api-access-php28". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:47:57 crc kubenswrapper[4821]: I1125 10:47:57.228453 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1aa8233d-1fe4-44ec-acc6-0d5548c70d59-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "1aa8233d-1fe4-44ec-acc6-0d5548c70d59" (UID: "1aa8233d-1fe4-44ec-acc6-0d5548c70d59"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:47:57 crc kubenswrapper[4821]: I1125 10:47:57.282376 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1aa8233d-1fe4-44ec-acc6-0d5548c70d59-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "1aa8233d-1fe4-44ec-acc6-0d5548c70d59" (UID: "1aa8233d-1fe4-44ec-acc6-0d5548c70d59"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:47:57 crc kubenswrapper[4821]: I1125 10:47:57.296378 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1aa8233d-1fe4-44ec-acc6-0d5548c70d59-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "1aa8233d-1fe4-44ec-acc6-0d5548c70d59" (UID: "1aa8233d-1fe4-44ec-acc6-0d5548c70d59"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:47:57 crc kubenswrapper[4821]: I1125 10:47:57.304348 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1aa8233d-1fe4-44ec-acc6-0d5548c70d59-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1aa8233d-1fe4-44ec-acc6-0d5548c70d59" (UID: "1aa8233d-1fe4-44ec-acc6-0d5548c70d59"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:47:57 crc kubenswrapper[4821]: I1125 10:47:57.307718 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1aa8233d-1fe4-44ec-acc6-0d5548c70d59-scripts" (OuterVolumeSpecName: "scripts") pod "1aa8233d-1fe4-44ec-acc6-0d5548c70d59" (UID: "1aa8233d-1fe4-44ec-acc6-0d5548c70d59"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:47:57 crc kubenswrapper[4821]: I1125 10:47:57.324699 4821 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1aa8233d-1fe4-44ec-acc6-0d5548c70d59-etc-swift\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:57 crc kubenswrapper[4821]: I1125 10:47:57.324729 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-php28\" (UniqueName: \"kubernetes.io/projected/1aa8233d-1fe4-44ec-acc6-0d5548c70d59-kube-api-access-php28\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:57 crc kubenswrapper[4821]: I1125 10:47:57.324743 4821 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1aa8233d-1fe4-44ec-acc6-0d5548c70d59-swiftconf\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:57 crc kubenswrapper[4821]: I1125 10:47:57.324752 4821 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1aa8233d-1fe4-44ec-acc6-0d5548c70d59-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:57 crc kubenswrapper[4821]: I1125 10:47:57.324762 4821 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1aa8233d-1fe4-44ec-acc6-0d5548c70d59-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:57 crc kubenswrapper[4821]: I1125 10:47:57.324773 4821 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1aa8233d-1fe4-44ec-acc6-0d5548c70d59-ring-data-devices\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:57 crc kubenswrapper[4821]: I1125 10:47:57.324783 4821 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1aa8233d-1fe4-44ec-acc6-0d5548c70d59-dispersionconf\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:57 crc kubenswrapper[4821]: I1125 10:47:57.771144 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-gmrpw" event={"ID":"1aa8233d-1fe4-44ec-acc6-0d5548c70d59","Type":"ContainerDied","Data":"d4053685b0a60e6d2adcf8455c3d95f804252d698b630071dae91597fb96ff6f"} Nov 25 10:47:57 crc kubenswrapper[4821]: I1125 10:47:57.771195 4821 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d4053685b0a60e6d2adcf8455c3d95f804252d698b630071dae91597fb96ff6f" Nov 25 10:47:57 crc kubenswrapper[4821]: I1125 10:47:57.771251 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-gmrpw" Nov 25 10:47:59 crc kubenswrapper[4821]: I1125 10:47:59.788908 4821 generic.go:334] "Generic (PLEG): container finished" podID="2bbb754a-fff3-4ba2-ab75-68e48b626ab0" containerID="ef45f64d359b5165b93856160cad982dc3eb70d24d3b0aed64df4b49a699da2b" exitCode=0 Nov 25 10:47:59 crc kubenswrapper[4821]: I1125 10:47:59.789062 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"2bbb754a-fff3-4ba2-ab75-68e48b626ab0","Type":"ContainerDied","Data":"ef45f64d359b5165b93856160cad982dc3eb70d24d3b0aed64df4b49a699da2b"} Nov 25 10:47:59 crc kubenswrapper[4821]: I1125 10:47:59.791846 4821 generic.go:334] "Generic (PLEG): container finished" podID="330a479e-3d76-4de3-bc84-1d7c2d65d265" containerID="658c4ee0f18233ace6320ae015dedf80011737bf96f761266c9995e11d4f5dc7" exitCode=0 Nov 25 10:47:59 crc kubenswrapper[4821]: I1125 10:47:59.791865 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"330a479e-3d76-4de3-bc84-1d7c2d65d265","Type":"ContainerDied","Data":"658c4ee0f18233ace6320ae015dedf80011737bf96f761266c9995e11d4f5dc7"} Nov 25 10:48:00 crc kubenswrapper[4821]: I1125 10:48:00.405379 4821 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-ltxc6" podUID="47eaf58d-190b-4e75-9aac-11d7e51df259" containerName="ovn-controller" probeResult="failure" output=< Nov 25 10:48:00 crc kubenswrapper[4821]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 25 10:48:00 crc kubenswrapper[4821]: > Nov 25 10:48:00 crc kubenswrapper[4821]: I1125 10:48:00.446259 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-6mbr2" Nov 25 10:48:00 crc kubenswrapper[4821]: I1125 10:48:00.452764 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-6mbr2" Nov 25 10:48:00 crc kubenswrapper[4821]: I1125 10:48:00.685464 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ltxc6-config-dmm7r"] Nov 25 10:48:00 crc kubenswrapper[4821]: E1125 10:48:00.685847 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1aa8233d-1fe4-44ec-acc6-0d5548c70d59" containerName="swift-ring-rebalance" Nov 25 10:48:00 crc kubenswrapper[4821]: I1125 10:48:00.685869 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="1aa8233d-1fe4-44ec-acc6-0d5548c70d59" containerName="swift-ring-rebalance" Nov 25 10:48:00 crc kubenswrapper[4821]: I1125 10:48:00.686134 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="1aa8233d-1fe4-44ec-acc6-0d5548c70d59" containerName="swift-ring-rebalance" Nov 25 10:48:00 crc kubenswrapper[4821]: I1125 10:48:00.686798 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ltxc6-config-dmm7r" Nov 25 10:48:00 crc kubenswrapper[4821]: I1125 10:48:00.690789 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Nov 25 10:48:00 crc kubenswrapper[4821]: I1125 10:48:00.707103 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ltxc6-config-dmm7r"] Nov 25 10:48:00 crc kubenswrapper[4821]: I1125 10:48:00.787261 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d65mj\" (UniqueName: \"kubernetes.io/projected/e90b3eff-e1ad-4e9e-9a16-cfbdff709384-kube-api-access-d65mj\") pod \"ovn-controller-ltxc6-config-dmm7r\" (UID: \"e90b3eff-e1ad-4e9e-9a16-cfbdff709384\") " pod="openstack/ovn-controller-ltxc6-config-dmm7r" Nov 25 10:48:00 crc kubenswrapper[4821]: I1125 10:48:00.787324 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e90b3eff-e1ad-4e9e-9a16-cfbdff709384-var-log-ovn\") pod \"ovn-controller-ltxc6-config-dmm7r\" (UID: \"e90b3eff-e1ad-4e9e-9a16-cfbdff709384\") " pod="openstack/ovn-controller-ltxc6-config-dmm7r" Nov 25 10:48:00 crc kubenswrapper[4821]: I1125 10:48:00.787356 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e90b3eff-e1ad-4e9e-9a16-cfbdff709384-var-run\") pod \"ovn-controller-ltxc6-config-dmm7r\" (UID: \"e90b3eff-e1ad-4e9e-9a16-cfbdff709384\") " pod="openstack/ovn-controller-ltxc6-config-dmm7r" Nov 25 10:48:00 crc kubenswrapper[4821]: I1125 10:48:00.787573 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e90b3eff-e1ad-4e9e-9a16-cfbdff709384-scripts\") pod \"ovn-controller-ltxc6-config-dmm7r\" (UID: \"e90b3eff-e1ad-4e9e-9a16-cfbdff709384\") " pod="openstack/ovn-controller-ltxc6-config-dmm7r" Nov 25 10:48:00 crc kubenswrapper[4821]: I1125 10:48:00.787693 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e90b3eff-e1ad-4e9e-9a16-cfbdff709384-var-run-ovn\") pod \"ovn-controller-ltxc6-config-dmm7r\" (UID: \"e90b3eff-e1ad-4e9e-9a16-cfbdff709384\") " pod="openstack/ovn-controller-ltxc6-config-dmm7r" Nov 25 10:48:00 crc kubenswrapper[4821]: I1125 10:48:00.787873 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e90b3eff-e1ad-4e9e-9a16-cfbdff709384-additional-scripts\") pod \"ovn-controller-ltxc6-config-dmm7r\" (UID: \"e90b3eff-e1ad-4e9e-9a16-cfbdff709384\") " pod="openstack/ovn-controller-ltxc6-config-dmm7r" Nov 25 10:48:00 crc kubenswrapper[4821]: I1125 10:48:00.819378 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"330a479e-3d76-4de3-bc84-1d7c2d65d265","Type":"ContainerStarted","Data":"7848d9adca99a2cef746a81eaafad2bfec909b4451ebcf92763c11f7448b1d15"} Nov 25 10:48:00 crc kubenswrapper[4821]: I1125 10:48:00.819712 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:48:00 crc kubenswrapper[4821]: I1125 10:48:00.830126 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"2bbb754a-fff3-4ba2-ab75-68e48b626ab0","Type":"ContainerStarted","Data":"60497933ff5037a6512cb5d8a0394d48cdda3bb66db9431a0698858ee9f8eb43"} Nov 25 10:48:00 crc kubenswrapper[4821]: I1125 10:48:00.830723 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 25 10:48:00 crc kubenswrapper[4821]: I1125 10:48:00.862797 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=53.054509321 podStartE2EDuration="1m1.862758189s" podCreationTimestamp="2025-11-25 10:46:59 +0000 UTC" firstStartedPulling="2025-11-25 10:47:13.301850139 +0000 UTC m=+903.838169986" lastFinishedPulling="2025-11-25 10:47:22.110099007 +0000 UTC m=+912.646418854" observedRunningTime="2025-11-25 10:48:00.847470491 +0000 UTC m=+951.383790338" watchObservedRunningTime="2025-11-25 10:48:00.862758189 +0000 UTC m=+951.399078036" Nov 25 10:48:00 crc kubenswrapper[4821]: I1125 10:48:00.889907 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e90b3eff-e1ad-4e9e-9a16-cfbdff709384-var-run-ovn\") pod \"ovn-controller-ltxc6-config-dmm7r\" (UID: \"e90b3eff-e1ad-4e9e-9a16-cfbdff709384\") " pod="openstack/ovn-controller-ltxc6-config-dmm7r" Nov 25 10:48:00 crc kubenswrapper[4821]: I1125 10:48:00.889992 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e90b3eff-e1ad-4e9e-9a16-cfbdff709384-additional-scripts\") pod \"ovn-controller-ltxc6-config-dmm7r\" (UID: \"e90b3eff-e1ad-4e9e-9a16-cfbdff709384\") " pod="openstack/ovn-controller-ltxc6-config-dmm7r" Nov 25 10:48:00 crc kubenswrapper[4821]: I1125 10:48:00.890076 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d65mj\" (UniqueName: \"kubernetes.io/projected/e90b3eff-e1ad-4e9e-9a16-cfbdff709384-kube-api-access-d65mj\") pod \"ovn-controller-ltxc6-config-dmm7r\" (UID: \"e90b3eff-e1ad-4e9e-9a16-cfbdff709384\") " pod="openstack/ovn-controller-ltxc6-config-dmm7r" Nov 25 10:48:00 crc kubenswrapper[4821]: I1125 10:48:00.890106 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e90b3eff-e1ad-4e9e-9a16-cfbdff709384-var-log-ovn\") pod \"ovn-controller-ltxc6-config-dmm7r\" (UID: \"e90b3eff-e1ad-4e9e-9a16-cfbdff709384\") " pod="openstack/ovn-controller-ltxc6-config-dmm7r" Nov 25 10:48:00 crc kubenswrapper[4821]: I1125 10:48:00.890133 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e90b3eff-e1ad-4e9e-9a16-cfbdff709384-var-run\") pod \"ovn-controller-ltxc6-config-dmm7r\" (UID: \"e90b3eff-e1ad-4e9e-9a16-cfbdff709384\") " pod="openstack/ovn-controller-ltxc6-config-dmm7r" Nov 25 10:48:00 crc kubenswrapper[4821]: I1125 10:48:00.890224 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e90b3eff-e1ad-4e9e-9a16-cfbdff709384-scripts\") pod \"ovn-controller-ltxc6-config-dmm7r\" (UID: \"e90b3eff-e1ad-4e9e-9a16-cfbdff709384\") " pod="openstack/ovn-controller-ltxc6-config-dmm7r" Nov 25 10:48:00 crc kubenswrapper[4821]: I1125 10:48:00.890286 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e90b3eff-e1ad-4e9e-9a16-cfbdff709384-var-run-ovn\") pod \"ovn-controller-ltxc6-config-dmm7r\" (UID: \"e90b3eff-e1ad-4e9e-9a16-cfbdff709384\") " pod="openstack/ovn-controller-ltxc6-config-dmm7r" Nov 25 10:48:00 crc kubenswrapper[4821]: I1125 10:48:00.890382 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e90b3eff-e1ad-4e9e-9a16-cfbdff709384-var-log-ovn\") pod \"ovn-controller-ltxc6-config-dmm7r\" (UID: \"e90b3eff-e1ad-4e9e-9a16-cfbdff709384\") " pod="openstack/ovn-controller-ltxc6-config-dmm7r" Nov 25 10:48:00 crc kubenswrapper[4821]: I1125 10:48:00.890416 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e90b3eff-e1ad-4e9e-9a16-cfbdff709384-var-run\") pod \"ovn-controller-ltxc6-config-dmm7r\" (UID: \"e90b3eff-e1ad-4e9e-9a16-cfbdff709384\") " pod="openstack/ovn-controller-ltxc6-config-dmm7r" Nov 25 10:48:00 crc kubenswrapper[4821]: I1125 10:48:00.892770 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e90b3eff-e1ad-4e9e-9a16-cfbdff709384-scripts\") pod \"ovn-controller-ltxc6-config-dmm7r\" (UID: \"e90b3eff-e1ad-4e9e-9a16-cfbdff709384\") " pod="openstack/ovn-controller-ltxc6-config-dmm7r" Nov 25 10:48:00 crc kubenswrapper[4821]: I1125 10:48:00.893855 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e90b3eff-e1ad-4e9e-9a16-cfbdff709384-additional-scripts\") pod \"ovn-controller-ltxc6-config-dmm7r\" (UID: \"e90b3eff-e1ad-4e9e-9a16-cfbdff709384\") " pod="openstack/ovn-controller-ltxc6-config-dmm7r" Nov 25 10:48:00 crc kubenswrapper[4821]: I1125 10:48:00.916902 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=55.536213939 podStartE2EDuration="1m0.916885631s" podCreationTimestamp="2025-11-25 10:47:00 +0000 UTC" firstStartedPulling="2025-11-25 10:47:16.728346047 +0000 UTC m=+907.264665884" lastFinishedPulling="2025-11-25 10:47:22.109017729 +0000 UTC m=+912.645337576" observedRunningTime="2025-11-25 10:48:00.915096626 +0000 UTC m=+951.451416473" watchObservedRunningTime="2025-11-25 10:48:00.916885631 +0000 UTC m=+951.453205478" Nov 25 10:48:00 crc kubenswrapper[4821]: I1125 10:48:00.928917 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d65mj\" (UniqueName: \"kubernetes.io/projected/e90b3eff-e1ad-4e9e-9a16-cfbdff709384-kube-api-access-d65mj\") pod \"ovn-controller-ltxc6-config-dmm7r\" (UID: \"e90b3eff-e1ad-4e9e-9a16-cfbdff709384\") " pod="openstack/ovn-controller-ltxc6-config-dmm7r" Nov 25 10:48:01 crc kubenswrapper[4821]: I1125 10:48:01.017347 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ltxc6-config-dmm7r" Nov 25 10:48:01 crc kubenswrapper[4821]: I1125 10:48:01.487281 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ltxc6-config-dmm7r"] Nov 25 10:48:01 crc kubenswrapper[4821]: I1125 10:48:01.840609 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ltxc6-config-dmm7r" event={"ID":"e90b3eff-e1ad-4e9e-9a16-cfbdff709384","Type":"ContainerStarted","Data":"6950178aad59d19cd80bb0a741c400699828134cafedb603197165656121141c"} Nov 25 10:48:01 crc kubenswrapper[4821]: I1125 10:48:01.840980 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ltxc6-config-dmm7r" event={"ID":"e90b3eff-e1ad-4e9e-9a16-cfbdff709384","Type":"ContainerStarted","Data":"3c673287fbd6297d8e16730d4945891bfe493922642bca4cc9f130e76e62dae6"} Nov 25 10:48:02 crc kubenswrapper[4821]: I1125 10:48:02.849776 4821 generic.go:334] "Generic (PLEG): container finished" podID="e90b3eff-e1ad-4e9e-9a16-cfbdff709384" containerID="6950178aad59d19cd80bb0a741c400699828134cafedb603197165656121141c" exitCode=0 Nov 25 10:48:02 crc kubenswrapper[4821]: I1125 10:48:02.849823 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ltxc6-config-dmm7r" event={"ID":"e90b3eff-e1ad-4e9e-9a16-cfbdff709384","Type":"ContainerDied","Data":"6950178aad59d19cd80bb0a741c400699828134cafedb603197165656121141c"} Nov 25 10:48:04 crc kubenswrapper[4821]: I1125 10:48:04.715076 4821 patch_prober.go:28] interesting pod/machine-config-daemon-2krbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:48:04 crc kubenswrapper[4821]: I1125 10:48:04.715460 4821 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:48:05 crc kubenswrapper[4821]: I1125 10:48:05.403646 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ltxc6" Nov 25 10:48:08 crc kubenswrapper[4821]: I1125 10:48:08.777719 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ltxc6-config-dmm7r" Nov 25 10:48:08 crc kubenswrapper[4821]: I1125 10:48:08.907436 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ltxc6-config-dmm7r" event={"ID":"e90b3eff-e1ad-4e9e-9a16-cfbdff709384","Type":"ContainerDied","Data":"3c673287fbd6297d8e16730d4945891bfe493922642bca4cc9f130e76e62dae6"} Nov 25 10:48:08 crc kubenswrapper[4821]: I1125 10:48:08.907477 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ltxc6-config-dmm7r" Nov 25 10:48:08 crc kubenswrapper[4821]: I1125 10:48:08.907485 4821 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3c673287fbd6297d8e16730d4945891bfe493922642bca4cc9f130e76e62dae6" Nov 25 10:48:08 crc kubenswrapper[4821]: I1125 10:48:08.954842 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e90b3eff-e1ad-4e9e-9a16-cfbdff709384-scripts\") pod \"e90b3eff-e1ad-4e9e-9a16-cfbdff709384\" (UID: \"e90b3eff-e1ad-4e9e-9a16-cfbdff709384\") " Nov 25 10:48:08 crc kubenswrapper[4821]: I1125 10:48:08.954979 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e90b3eff-e1ad-4e9e-9a16-cfbdff709384-var-run-ovn\") pod \"e90b3eff-e1ad-4e9e-9a16-cfbdff709384\" (UID: \"e90b3eff-e1ad-4e9e-9a16-cfbdff709384\") " Nov 25 10:48:08 crc kubenswrapper[4821]: I1125 10:48:08.955059 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d65mj\" (UniqueName: \"kubernetes.io/projected/e90b3eff-e1ad-4e9e-9a16-cfbdff709384-kube-api-access-d65mj\") pod \"e90b3eff-e1ad-4e9e-9a16-cfbdff709384\" (UID: \"e90b3eff-e1ad-4e9e-9a16-cfbdff709384\") " Nov 25 10:48:08 crc kubenswrapper[4821]: I1125 10:48:08.955089 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e90b3eff-e1ad-4e9e-9a16-cfbdff709384-var-log-ovn\") pod \"e90b3eff-e1ad-4e9e-9a16-cfbdff709384\" (UID: \"e90b3eff-e1ad-4e9e-9a16-cfbdff709384\") " Nov 25 10:48:08 crc kubenswrapper[4821]: I1125 10:48:08.955108 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e90b3eff-e1ad-4e9e-9a16-cfbdff709384-var-run\") pod \"e90b3eff-e1ad-4e9e-9a16-cfbdff709384\" (UID: \"e90b3eff-e1ad-4e9e-9a16-cfbdff709384\") " Nov 25 10:48:08 crc kubenswrapper[4821]: I1125 10:48:08.955178 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e90b3eff-e1ad-4e9e-9a16-cfbdff709384-additional-scripts\") pod \"e90b3eff-e1ad-4e9e-9a16-cfbdff709384\" (UID: \"e90b3eff-e1ad-4e9e-9a16-cfbdff709384\") " Nov 25 10:48:08 crc kubenswrapper[4821]: I1125 10:48:08.955371 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e90b3eff-e1ad-4e9e-9a16-cfbdff709384-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "e90b3eff-e1ad-4e9e-9a16-cfbdff709384" (UID: "e90b3eff-e1ad-4e9e-9a16-cfbdff709384"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:48:08 crc kubenswrapper[4821]: I1125 10:48:08.955370 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e90b3eff-e1ad-4e9e-9a16-cfbdff709384-var-run" (OuterVolumeSpecName: "var-run") pod "e90b3eff-e1ad-4e9e-9a16-cfbdff709384" (UID: "e90b3eff-e1ad-4e9e-9a16-cfbdff709384"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:48:08 crc kubenswrapper[4821]: I1125 10:48:08.955869 4821 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e90b3eff-e1ad-4e9e-9a16-cfbdff709384-var-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 25 10:48:08 crc kubenswrapper[4821]: I1125 10:48:08.955883 4821 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e90b3eff-e1ad-4e9e-9a16-cfbdff709384-var-run\") on node \"crc\" DevicePath \"\"" Nov 25 10:48:08 crc kubenswrapper[4821]: I1125 10:48:08.955322 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e90b3eff-e1ad-4e9e-9a16-cfbdff709384-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "e90b3eff-e1ad-4e9e-9a16-cfbdff709384" (UID: "e90b3eff-e1ad-4e9e-9a16-cfbdff709384"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:48:08 crc kubenswrapper[4821]: I1125 10:48:08.956020 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e90b3eff-e1ad-4e9e-9a16-cfbdff709384-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "e90b3eff-e1ad-4e9e-9a16-cfbdff709384" (UID: "e90b3eff-e1ad-4e9e-9a16-cfbdff709384"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:48:08 crc kubenswrapper[4821]: I1125 10:48:08.956323 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e90b3eff-e1ad-4e9e-9a16-cfbdff709384-scripts" (OuterVolumeSpecName: "scripts") pod "e90b3eff-e1ad-4e9e-9a16-cfbdff709384" (UID: "e90b3eff-e1ad-4e9e-9a16-cfbdff709384"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:48:08 crc kubenswrapper[4821]: I1125 10:48:08.959578 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e90b3eff-e1ad-4e9e-9a16-cfbdff709384-kube-api-access-d65mj" (OuterVolumeSpecName: "kube-api-access-d65mj") pod "e90b3eff-e1ad-4e9e-9a16-cfbdff709384" (UID: "e90b3eff-e1ad-4e9e-9a16-cfbdff709384"). InnerVolumeSpecName "kube-api-access-d65mj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:48:09 crc kubenswrapper[4821]: I1125 10:48:09.058034 4821 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e90b3eff-e1ad-4e9e-9a16-cfbdff709384-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:48:09 crc kubenswrapper[4821]: I1125 10:48:09.058075 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d65mj\" (UniqueName: \"kubernetes.io/projected/e90b3eff-e1ad-4e9e-9a16-cfbdff709384-kube-api-access-d65mj\") on node \"crc\" DevicePath \"\"" Nov 25 10:48:09 crc kubenswrapper[4821]: I1125 10:48:09.058091 4821 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e90b3eff-e1ad-4e9e-9a16-cfbdff709384-var-log-ovn\") on node \"crc\" DevicePath \"\"" Nov 25 10:48:09 crc kubenswrapper[4821]: I1125 10:48:09.058105 4821 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e90b3eff-e1ad-4e9e-9a16-cfbdff709384-additional-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:48:09 crc kubenswrapper[4821]: I1125 10:48:09.869527 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0ffb0da2-affb-4a95-87f3-eac442dc754d-etc-swift\") pod \"swift-storage-0\" (UID: \"0ffb0da2-affb-4a95-87f3-eac442dc754d\") " pod="openstack/swift-storage-0" Nov 25 10:48:09 crc kubenswrapper[4821]: I1125 10:48:09.875688 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0ffb0da2-affb-4a95-87f3-eac442dc754d-etc-swift\") pod \"swift-storage-0\" (UID: \"0ffb0da2-affb-4a95-87f3-eac442dc754d\") " pod="openstack/swift-storage-0" Nov 25 10:48:09 crc kubenswrapper[4821]: I1125 10:48:09.902732 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ltxc6-config-dmm7r"] Nov 25 10:48:09 crc kubenswrapper[4821]: I1125 10:48:09.910327 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-ltxc6-config-dmm7r"] Nov 25 10:48:09 crc kubenswrapper[4821]: I1125 10:48:09.917951 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-2nhpq" event={"ID":"a8123055-1953-42b3-9c04-716061499d03","Type":"ContainerStarted","Data":"c8ccf9a6a3df8286261caebc19be09c2fc61f8c79473aded6a8b5d775e5bfc6a"} Nov 25 10:48:09 crc kubenswrapper[4821]: I1125 10:48:09.938709 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-2nhpq" podStartSLOduration=1.918920124 podStartE2EDuration="14.938689921s" podCreationTimestamp="2025-11-25 10:47:55 +0000 UTC" firstStartedPulling="2025-11-25 10:47:55.929762321 +0000 UTC m=+946.466082168" lastFinishedPulling="2025-11-25 10:48:08.949532118 +0000 UTC m=+959.485851965" observedRunningTime="2025-11-25 10:48:09.933170061 +0000 UTC m=+960.469489898" watchObservedRunningTime="2025-11-25 10:48:09.938689921 +0000 UTC m=+960.475009768" Nov 25 10:48:10 crc kubenswrapper[4821]: I1125 10:48:10.014300 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 25 10:48:10 crc kubenswrapper[4821]: I1125 10:48:10.144585 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e90b3eff-e1ad-4e9e-9a16-cfbdff709384" path="/var/lib/kubelet/pods/e90b3eff-e1ad-4e9e-9a16-cfbdff709384/volumes" Nov 25 10:48:10 crc kubenswrapper[4821]: I1125 10:48:10.597430 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Nov 25 10:48:10 crc kubenswrapper[4821]: I1125 10:48:10.929649 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0ffb0da2-affb-4a95-87f3-eac442dc754d","Type":"ContainerStarted","Data":"5450ab3346bf050b82ffcddba4ec09c3b743068f345a5671cc4258a428848b73"} Nov 25 10:48:11 crc kubenswrapper[4821]: I1125 10:48:11.392394 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:48:11 crc kubenswrapper[4821]: I1125 10:48:11.485345 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.067346 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-xpctj"] Nov 25 10:48:13 crc kubenswrapper[4821]: E1125 10:48:13.068071 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e90b3eff-e1ad-4e9e-9a16-cfbdff709384" containerName="ovn-config" Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.068092 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="e90b3eff-e1ad-4e9e-9a16-cfbdff709384" containerName="ovn-config" Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.068341 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="e90b3eff-e1ad-4e9e-9a16-cfbdff709384" containerName="ovn-config" Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.068992 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-xpctj" Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.090563 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-xpctj"] Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.157257 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-55n47"] Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.158270 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-55n47" Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.171937 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-55n47"] Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.178939 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-a9df-account-create-xlkf6"] Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.180294 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-a9df-account-create-xlkf6" Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.183422 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.197077 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-a9df-account-create-xlkf6"] Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.230064 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gnrqm\" (UniqueName: \"kubernetes.io/projected/2d5f3fd0-3830-4079-bbf2-0968a8a454f1-kube-api-access-gnrqm\") pod \"cinder-db-create-xpctj\" (UID: \"2d5f3fd0-3830-4079-bbf2-0968a8a454f1\") " pod="openstack/cinder-db-create-xpctj" Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.230307 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d5f3fd0-3830-4079-bbf2-0968a8a454f1-operator-scripts\") pod \"cinder-db-create-xpctj\" (UID: \"2d5f3fd0-3830-4079-bbf2-0968a8a454f1\") " pod="openstack/cinder-db-create-xpctj" Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.289884 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-ff23-account-create-n68t7"] Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.292117 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-ff23-account-create-n68t7" Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.296508 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.313318 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-ff23-account-create-n68t7"] Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.332390 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/65d5f171-33f0-4c2d-9cfb-6847f2b98c49-operator-scripts\") pod \"cinder-a9df-account-create-xlkf6\" (UID: \"65d5f171-33f0-4c2d-9cfb-6847f2b98c49\") " pod="openstack/cinder-a9df-account-create-xlkf6" Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.332496 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cec42c9b-c925-4248-b73c-a076d1d0a8c4-operator-scripts\") pod \"barbican-db-create-55n47\" (UID: \"cec42c9b-c925-4248-b73c-a076d1d0a8c4\") " pod="openstack/barbican-db-create-55n47" Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.332547 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d5f3fd0-3830-4079-bbf2-0968a8a454f1-operator-scripts\") pod \"cinder-db-create-xpctj\" (UID: \"2d5f3fd0-3830-4079-bbf2-0968a8a454f1\") " pod="openstack/cinder-db-create-xpctj" Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.332582 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9k9ww\" (UniqueName: \"kubernetes.io/projected/65d5f171-33f0-4c2d-9cfb-6847f2b98c49-kube-api-access-9k9ww\") pod \"cinder-a9df-account-create-xlkf6\" (UID: \"65d5f171-33f0-4c2d-9cfb-6847f2b98c49\") " pod="openstack/cinder-a9df-account-create-xlkf6" Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.332621 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gvd62\" (UniqueName: \"kubernetes.io/projected/cec42c9b-c925-4248-b73c-a076d1d0a8c4-kube-api-access-gvd62\") pod \"barbican-db-create-55n47\" (UID: \"cec42c9b-c925-4248-b73c-a076d1d0a8c4\") " pod="openstack/barbican-db-create-55n47" Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.332640 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gnrqm\" (UniqueName: \"kubernetes.io/projected/2d5f3fd0-3830-4079-bbf2-0968a8a454f1-kube-api-access-gnrqm\") pod \"cinder-db-create-xpctj\" (UID: \"2d5f3fd0-3830-4079-bbf2-0968a8a454f1\") " pod="openstack/cinder-db-create-xpctj" Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.333527 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d5f3fd0-3830-4079-bbf2-0968a8a454f1-operator-scripts\") pod \"cinder-db-create-xpctj\" (UID: \"2d5f3fd0-3830-4079-bbf2-0968a8a454f1\") " pod="openstack/cinder-db-create-xpctj" Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.368740 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-sq6xx"] Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.370873 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-sq6xx" Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.371465 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gnrqm\" (UniqueName: \"kubernetes.io/projected/2d5f3fd0-3830-4079-bbf2-0968a8a454f1-kube-api-access-gnrqm\") pod \"cinder-db-create-xpctj\" (UID: \"2d5f3fd0-3830-4079-bbf2-0968a8a454f1\") " pod="openstack/cinder-db-create-xpctj" Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.382363 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-sq6xx"] Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.390099 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-xpctj" Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.434582 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gvd62\" (UniqueName: \"kubernetes.io/projected/cec42c9b-c925-4248-b73c-a076d1d0a8c4-kube-api-access-gvd62\") pod \"barbican-db-create-55n47\" (UID: \"cec42c9b-c925-4248-b73c-a076d1d0a8c4\") " pod="openstack/barbican-db-create-55n47" Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.434651 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/65d5f171-33f0-4c2d-9cfb-6847f2b98c49-operator-scripts\") pod \"cinder-a9df-account-create-xlkf6\" (UID: \"65d5f171-33f0-4c2d-9cfb-6847f2b98c49\") " pod="openstack/cinder-a9df-account-create-xlkf6" Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.434692 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a764b1fe-69e0-4609-814e-ab2d4e8b4438-operator-scripts\") pod \"barbican-ff23-account-create-n68t7\" (UID: \"a764b1fe-69e0-4609-814e-ab2d4e8b4438\") " pod="openstack/barbican-ff23-account-create-n68t7" Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.434795 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cec42c9b-c925-4248-b73c-a076d1d0a8c4-operator-scripts\") pod \"barbican-db-create-55n47\" (UID: \"cec42c9b-c925-4248-b73c-a076d1d0a8c4\") " pod="openstack/barbican-db-create-55n47" Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.434860 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9k9ww\" (UniqueName: \"kubernetes.io/projected/65d5f171-33f0-4c2d-9cfb-6847f2b98c49-kube-api-access-9k9ww\") pod \"cinder-a9df-account-create-xlkf6\" (UID: \"65d5f171-33f0-4c2d-9cfb-6847f2b98c49\") " pod="openstack/cinder-a9df-account-create-xlkf6" Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.434908 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85mx6\" (UniqueName: \"kubernetes.io/projected/a764b1fe-69e0-4609-814e-ab2d4e8b4438-kube-api-access-85mx6\") pod \"barbican-ff23-account-create-n68t7\" (UID: \"a764b1fe-69e0-4609-814e-ab2d4e8b4438\") " pod="openstack/barbican-ff23-account-create-n68t7" Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.435961 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/65d5f171-33f0-4c2d-9cfb-6847f2b98c49-operator-scripts\") pod \"cinder-a9df-account-create-xlkf6\" (UID: \"65d5f171-33f0-4c2d-9cfb-6847f2b98c49\") " pod="openstack/cinder-a9df-account-create-xlkf6" Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.435964 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cec42c9b-c925-4248-b73c-a076d1d0a8c4-operator-scripts\") pod \"barbican-db-create-55n47\" (UID: \"cec42c9b-c925-4248-b73c-a076d1d0a8c4\") " pod="openstack/barbican-db-create-55n47" Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.439527 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-bhnq8"] Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.440507 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-bhnq8" Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.446913 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.447268 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-rzxc4" Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.447526 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.447583 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.453370 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-bhnq8"] Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.468792 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9k9ww\" (UniqueName: \"kubernetes.io/projected/65d5f171-33f0-4c2d-9cfb-6847f2b98c49-kube-api-access-9k9ww\") pod \"cinder-a9df-account-create-xlkf6\" (UID: \"65d5f171-33f0-4c2d-9cfb-6847f2b98c49\") " pod="openstack/cinder-a9df-account-create-xlkf6" Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.478261 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gvd62\" (UniqueName: \"kubernetes.io/projected/cec42c9b-c925-4248-b73c-a076d1d0a8c4-kube-api-access-gvd62\") pod \"barbican-db-create-55n47\" (UID: \"cec42c9b-c925-4248-b73c-a076d1d0a8c4\") " pod="openstack/barbican-db-create-55n47" Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.478604 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-55n47" Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.493467 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-7da5-account-create-8wqv6"] Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.494657 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7da5-account-create-8wqv6" Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.498394 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-a9df-account-create-xlkf6" Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.499485 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.499707 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7da5-account-create-8wqv6"] Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.543596 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbdf62cf-071e-4e1c-9ef6-22bc39cfa795-config-data\") pod \"keystone-db-sync-bhnq8\" (UID: \"dbdf62cf-071e-4e1c-9ef6-22bc39cfa795\") " pod="openstack/keystone-db-sync-bhnq8" Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.543684 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dffwm\" (UniqueName: \"kubernetes.io/projected/3049cd6a-c38c-4da9-83ce-2195225c3ac9-kube-api-access-dffwm\") pod \"neutron-db-create-sq6xx\" (UID: \"3049cd6a-c38c-4da9-83ce-2195225c3ac9\") " pod="openstack/neutron-db-create-sq6xx" Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.543756 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85mx6\" (UniqueName: \"kubernetes.io/projected/a764b1fe-69e0-4609-814e-ab2d4e8b4438-kube-api-access-85mx6\") pod \"barbican-ff23-account-create-n68t7\" (UID: \"a764b1fe-69e0-4609-814e-ab2d4e8b4438\") " pod="openstack/barbican-ff23-account-create-n68t7" Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.543802 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lltzf\" (UniqueName: \"kubernetes.io/projected/dbdf62cf-071e-4e1c-9ef6-22bc39cfa795-kube-api-access-lltzf\") pod \"keystone-db-sync-bhnq8\" (UID: \"dbdf62cf-071e-4e1c-9ef6-22bc39cfa795\") " pod="openstack/keystone-db-sync-bhnq8" Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.543869 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3049cd6a-c38c-4da9-83ce-2195225c3ac9-operator-scripts\") pod \"neutron-db-create-sq6xx\" (UID: \"3049cd6a-c38c-4da9-83ce-2195225c3ac9\") " pod="openstack/neutron-db-create-sq6xx" Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.543896 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a764b1fe-69e0-4609-814e-ab2d4e8b4438-operator-scripts\") pod \"barbican-ff23-account-create-n68t7\" (UID: \"a764b1fe-69e0-4609-814e-ab2d4e8b4438\") " pod="openstack/barbican-ff23-account-create-n68t7" Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.543941 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbdf62cf-071e-4e1c-9ef6-22bc39cfa795-combined-ca-bundle\") pod \"keystone-db-sync-bhnq8\" (UID: \"dbdf62cf-071e-4e1c-9ef6-22bc39cfa795\") " pod="openstack/keystone-db-sync-bhnq8" Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.545214 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a764b1fe-69e0-4609-814e-ab2d4e8b4438-operator-scripts\") pod \"barbican-ff23-account-create-n68t7\" (UID: \"a764b1fe-69e0-4609-814e-ab2d4e8b4438\") " pod="openstack/barbican-ff23-account-create-n68t7" Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.576350 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85mx6\" (UniqueName: \"kubernetes.io/projected/a764b1fe-69e0-4609-814e-ab2d4e8b4438-kube-api-access-85mx6\") pod \"barbican-ff23-account-create-n68t7\" (UID: \"a764b1fe-69e0-4609-814e-ab2d4e8b4438\") " pod="openstack/barbican-ff23-account-create-n68t7" Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.613789 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-ff23-account-create-n68t7" Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.646183 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lltzf\" (UniqueName: \"kubernetes.io/projected/dbdf62cf-071e-4e1c-9ef6-22bc39cfa795-kube-api-access-lltzf\") pod \"keystone-db-sync-bhnq8\" (UID: \"dbdf62cf-071e-4e1c-9ef6-22bc39cfa795\") " pod="openstack/keystone-db-sync-bhnq8" Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.646268 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3049cd6a-c38c-4da9-83ce-2195225c3ac9-operator-scripts\") pod \"neutron-db-create-sq6xx\" (UID: \"3049cd6a-c38c-4da9-83ce-2195225c3ac9\") " pod="openstack/neutron-db-create-sq6xx" Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.646306 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbdf62cf-071e-4e1c-9ef6-22bc39cfa795-combined-ca-bundle\") pod \"keystone-db-sync-bhnq8\" (UID: \"dbdf62cf-071e-4e1c-9ef6-22bc39cfa795\") " pod="openstack/keystone-db-sync-bhnq8" Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.646347 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8v9gk\" (UniqueName: \"kubernetes.io/projected/ba88f146-f158-4703-b77d-319f4f65d419-kube-api-access-8v9gk\") pod \"neutron-7da5-account-create-8wqv6\" (UID: \"ba88f146-f158-4703-b77d-319f4f65d419\") " pod="openstack/neutron-7da5-account-create-8wqv6" Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.646469 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbdf62cf-071e-4e1c-9ef6-22bc39cfa795-config-data\") pod \"keystone-db-sync-bhnq8\" (UID: \"dbdf62cf-071e-4e1c-9ef6-22bc39cfa795\") " pod="openstack/keystone-db-sync-bhnq8" Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.646516 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ba88f146-f158-4703-b77d-319f4f65d419-operator-scripts\") pod \"neutron-7da5-account-create-8wqv6\" (UID: \"ba88f146-f158-4703-b77d-319f4f65d419\") " pod="openstack/neutron-7da5-account-create-8wqv6" Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.646541 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dffwm\" (UniqueName: \"kubernetes.io/projected/3049cd6a-c38c-4da9-83ce-2195225c3ac9-kube-api-access-dffwm\") pod \"neutron-db-create-sq6xx\" (UID: \"3049cd6a-c38c-4da9-83ce-2195225c3ac9\") " pod="openstack/neutron-db-create-sq6xx" Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.647050 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3049cd6a-c38c-4da9-83ce-2195225c3ac9-operator-scripts\") pod \"neutron-db-create-sq6xx\" (UID: \"3049cd6a-c38c-4da9-83ce-2195225c3ac9\") " pod="openstack/neutron-db-create-sq6xx" Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.650636 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbdf62cf-071e-4e1c-9ef6-22bc39cfa795-config-data\") pod \"keystone-db-sync-bhnq8\" (UID: \"dbdf62cf-071e-4e1c-9ef6-22bc39cfa795\") " pod="openstack/keystone-db-sync-bhnq8" Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.659557 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lltzf\" (UniqueName: \"kubernetes.io/projected/dbdf62cf-071e-4e1c-9ef6-22bc39cfa795-kube-api-access-lltzf\") pod \"keystone-db-sync-bhnq8\" (UID: \"dbdf62cf-071e-4e1c-9ef6-22bc39cfa795\") " pod="openstack/keystone-db-sync-bhnq8" Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.659565 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbdf62cf-071e-4e1c-9ef6-22bc39cfa795-combined-ca-bundle\") pod \"keystone-db-sync-bhnq8\" (UID: \"dbdf62cf-071e-4e1c-9ef6-22bc39cfa795\") " pod="openstack/keystone-db-sync-bhnq8" Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.675554 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dffwm\" (UniqueName: \"kubernetes.io/projected/3049cd6a-c38c-4da9-83ce-2195225c3ac9-kube-api-access-dffwm\") pod \"neutron-db-create-sq6xx\" (UID: \"3049cd6a-c38c-4da9-83ce-2195225c3ac9\") " pod="openstack/neutron-db-create-sq6xx" Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.703662 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-sq6xx" Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.751341 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ba88f146-f158-4703-b77d-319f4f65d419-operator-scripts\") pod \"neutron-7da5-account-create-8wqv6\" (UID: \"ba88f146-f158-4703-b77d-319f4f65d419\") " pod="openstack/neutron-7da5-account-create-8wqv6" Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.751442 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8v9gk\" (UniqueName: \"kubernetes.io/projected/ba88f146-f158-4703-b77d-319f4f65d419-kube-api-access-8v9gk\") pod \"neutron-7da5-account-create-8wqv6\" (UID: \"ba88f146-f158-4703-b77d-319f4f65d419\") " pod="openstack/neutron-7da5-account-create-8wqv6" Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.752367 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ba88f146-f158-4703-b77d-319f4f65d419-operator-scripts\") pod \"neutron-7da5-account-create-8wqv6\" (UID: \"ba88f146-f158-4703-b77d-319f4f65d419\") " pod="openstack/neutron-7da5-account-create-8wqv6" Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.773180 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8v9gk\" (UniqueName: \"kubernetes.io/projected/ba88f146-f158-4703-b77d-319f4f65d419-kube-api-access-8v9gk\") pod \"neutron-7da5-account-create-8wqv6\" (UID: \"ba88f146-f158-4703-b77d-319f4f65d419\") " pod="openstack/neutron-7da5-account-create-8wqv6" Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.846588 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-bhnq8" Nov 25 10:48:13 crc kubenswrapper[4821]: I1125 10:48:13.895634 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7da5-account-create-8wqv6" Nov 25 10:48:14 crc kubenswrapper[4821]: I1125 10:48:14.076409 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-xpctj"] Nov 25 10:48:14 crc kubenswrapper[4821]: W1125 10:48:14.093888 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2d5f3fd0_3830_4079_bbf2_0968a8a454f1.slice/crio-f155827d660d30e178d3aa1fb7620c6e2a75c0e624f00e429b697771ad6c565c WatchSource:0}: Error finding container f155827d660d30e178d3aa1fb7620c6e2a75c0e624f00e429b697771ad6c565c: Status 404 returned error can't find the container with id f155827d660d30e178d3aa1fb7620c6e2a75c0e624f00e429b697771ad6c565c Nov 25 10:48:14 crc kubenswrapper[4821]: I1125 10:48:14.143259 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-55n47"] Nov 25 10:48:14 crc kubenswrapper[4821]: W1125 10:48:14.156869 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod65d5f171_33f0_4c2d_9cfb_6847f2b98c49.slice/crio-30100d28a5b8712886362d24e87df659b47b082870e8ddf7e262f3d6845c037b WatchSource:0}: Error finding container 30100d28a5b8712886362d24e87df659b47b082870e8ddf7e262f3d6845c037b: Status 404 returned error can't find the container with id 30100d28a5b8712886362d24e87df659b47b082870e8ddf7e262f3d6845c037b Nov 25 10:48:14 crc kubenswrapper[4821]: I1125 10:48:14.177340 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-a9df-account-create-xlkf6"] Nov 25 10:48:14 crc kubenswrapper[4821]: I1125 10:48:14.419796 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-ff23-account-create-n68t7"] Nov 25 10:48:14 crc kubenswrapper[4821]: I1125 10:48:14.490412 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-sq6xx"] Nov 25 10:48:14 crc kubenswrapper[4821]: I1125 10:48:14.610719 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-bhnq8"] Nov 25 10:48:14 crc kubenswrapper[4821]: W1125 10:48:14.610809 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddbdf62cf_071e_4e1c_9ef6_22bc39cfa795.slice/crio-884a1ca1174962b85fd4c57a68c054451ce76769a7966564e446be7051cfc247 WatchSource:0}: Error finding container 884a1ca1174962b85fd4c57a68c054451ce76769a7966564e446be7051cfc247: Status 404 returned error can't find the container with id 884a1ca1174962b85fd4c57a68c054451ce76769a7966564e446be7051cfc247 Nov 25 10:48:14 crc kubenswrapper[4821]: I1125 10:48:14.639853 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7da5-account-create-8wqv6"] Nov 25 10:48:14 crc kubenswrapper[4821]: I1125 10:48:14.960150 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7da5-account-create-8wqv6" event={"ID":"ba88f146-f158-4703-b77d-319f4f65d419","Type":"ContainerStarted","Data":"6b2011cd2c7284ba0bfb6c3bc27da8a28ef3cd07fe40398a9b41940ce2706445"} Nov 25 10:48:14 crc kubenswrapper[4821]: I1125 10:48:14.961052 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-bhnq8" event={"ID":"dbdf62cf-071e-4e1c-9ef6-22bc39cfa795","Type":"ContainerStarted","Data":"884a1ca1174962b85fd4c57a68c054451ce76769a7966564e446be7051cfc247"} Nov 25 10:48:14 crc kubenswrapper[4821]: I1125 10:48:14.962465 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-xpctj" event={"ID":"2d5f3fd0-3830-4079-bbf2-0968a8a454f1","Type":"ContainerStarted","Data":"f155827d660d30e178d3aa1fb7620c6e2a75c0e624f00e429b697771ad6c565c"} Nov 25 10:48:14 crc kubenswrapper[4821]: I1125 10:48:14.963989 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-a9df-account-create-xlkf6" event={"ID":"65d5f171-33f0-4c2d-9cfb-6847f2b98c49","Type":"ContainerStarted","Data":"30100d28a5b8712886362d24e87df659b47b082870e8ddf7e262f3d6845c037b"} Nov 25 10:48:14 crc kubenswrapper[4821]: I1125 10:48:14.965274 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-55n47" event={"ID":"cec42c9b-c925-4248-b73c-a076d1d0a8c4","Type":"ContainerStarted","Data":"db23f03f37e137ea5feb795b3e03def85f4c1d1fdef0e9ceb4ede145361364ab"} Nov 25 10:48:14 crc kubenswrapper[4821]: I1125 10:48:14.966140 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-sq6xx" event={"ID":"3049cd6a-c38c-4da9-83ce-2195225c3ac9","Type":"ContainerStarted","Data":"6e2a422a6331b073c6009847649f8d7a5d00e34756d10994b6910f9e5bac9bae"} Nov 25 10:48:14 crc kubenswrapper[4821]: I1125 10:48:14.967050 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-ff23-account-create-n68t7" event={"ID":"a764b1fe-69e0-4609-814e-ab2d4e8b4438","Type":"ContainerStarted","Data":"9c88689df8d8df71b51ff38caf6c0ffb03a0fb6f84bc8fb904ea51bf58904702"} Nov 25 10:48:15 crc kubenswrapper[4821]: I1125 10:48:15.988454 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-sq6xx" event={"ID":"3049cd6a-c38c-4da9-83ce-2195225c3ac9","Type":"ContainerStarted","Data":"71f2dc7f92b4c1969d58583c104e8d6f5aae05207de27f3adf8134f2ab203e20"} Nov 25 10:48:15 crc kubenswrapper[4821]: I1125 10:48:15.989910 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-ff23-account-create-n68t7" event={"ID":"a764b1fe-69e0-4609-814e-ab2d4e8b4438","Type":"ContainerStarted","Data":"93f388ffcf02e4a8292f29e53d8423f497f064c3108c92b59c7e7f8ea4230679"} Nov 25 10:48:15 crc kubenswrapper[4821]: I1125 10:48:15.991581 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7da5-account-create-8wqv6" event={"ID":"ba88f146-f158-4703-b77d-319f4f65d419","Type":"ContainerStarted","Data":"19b20fe65f03e445fec02e7da2a869b70f0effcc221783c3f191c8e4dde87178"} Nov 25 10:48:15 crc kubenswrapper[4821]: I1125 10:48:15.993562 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-xpctj" event={"ID":"2d5f3fd0-3830-4079-bbf2-0968a8a454f1","Type":"ContainerStarted","Data":"eef603c99d063234ffafa903350973a5c52aabccededa1c1088b6e9c03b89934"} Nov 25 10:48:16 crc kubenswrapper[4821]: I1125 10:48:16.009222 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-a9df-account-create-xlkf6" event={"ID":"65d5f171-33f0-4c2d-9cfb-6847f2b98c49","Type":"ContainerStarted","Data":"3c1126a2eb0007e787199e1622a82fcd8e726c76a4cdb353b4ac66d5433a61dd"} Nov 25 10:48:16 crc kubenswrapper[4821]: I1125 10:48:16.014688 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-55n47" event={"ID":"cec42c9b-c925-4248-b73c-a076d1d0a8c4","Type":"ContainerStarted","Data":"2a62662f27a8cfece54ff23a865831648780fa1fb785af45951a406accc7c6a9"} Nov 25 10:48:16 crc kubenswrapper[4821]: I1125 10:48:16.032866 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-create-55n47" podStartSLOduration=3.032845903 podStartE2EDuration="3.032845903s" podCreationTimestamp="2025-11-25 10:48:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:48:16.028198225 +0000 UTC m=+966.564518072" watchObservedRunningTime="2025-11-25 10:48:16.032845903 +0000 UTC m=+966.569165750" Nov 25 10:48:17 crc kubenswrapper[4821]: I1125 10:48:17.039798 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-7da5-account-create-8wqv6" podStartSLOduration=4.039763346 podStartE2EDuration="4.039763346s" podCreationTimestamp="2025-11-25 10:48:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:48:17.037821517 +0000 UTC m=+967.574141384" watchObservedRunningTime="2025-11-25 10:48:17.039763346 +0000 UTC m=+967.576083203" Nov 25 10:48:17 crc kubenswrapper[4821]: I1125 10:48:17.052415 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-ff23-account-create-n68t7" podStartSLOduration=4.052379296 podStartE2EDuration="4.052379296s" podCreationTimestamp="2025-11-25 10:48:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:48:17.050469538 +0000 UTC m=+967.586789385" watchObservedRunningTime="2025-11-25 10:48:17.052379296 +0000 UTC m=+967.588699143" Nov 25 10:48:17 crc kubenswrapper[4821]: I1125 10:48:17.069461 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-create-sq6xx" podStartSLOduration=4.069439099 podStartE2EDuration="4.069439099s" podCreationTimestamp="2025-11-25 10:48:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:48:17.06554058 +0000 UTC m=+967.601860437" watchObservedRunningTime="2025-11-25 10:48:17.069439099 +0000 UTC m=+967.605758946" Nov 25 10:48:17 crc kubenswrapper[4821]: I1125 10:48:17.081308 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-a9df-account-create-xlkf6" podStartSLOduration=4.081286199 podStartE2EDuration="4.081286199s" podCreationTimestamp="2025-11-25 10:48:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:48:17.078749615 +0000 UTC m=+967.615069472" watchObservedRunningTime="2025-11-25 10:48:17.081286199 +0000 UTC m=+967.617606046" Nov 25 10:48:17 crc kubenswrapper[4821]: I1125 10:48:17.098082 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-create-xpctj" podStartSLOduration=4.098063305 podStartE2EDuration="4.098063305s" podCreationTimestamp="2025-11-25 10:48:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:48:17.092583246 +0000 UTC m=+967.628903093" watchObservedRunningTime="2025-11-25 10:48:17.098063305 +0000 UTC m=+967.634383152" Nov 25 10:48:27 crc kubenswrapper[4821]: E1125 10:48:27.804776 4821 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-swift-account:current-podified" Nov 25 10:48:27 crc kubenswrapper[4821]: E1125 10:48:27.805506 4821 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:account-server,Image:quay.io/podified-antelope-centos9/openstack-swift-account:current-podified,Command:[/usr/bin/swift-account-server /etc/swift/account-server.conf.d -v],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:account,HostPort:0,ContainerPort:6202,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5b7h56h9dh94h67bh697h95h55hbh555h556h675h5fdh57dh579h5fbh64fh5c9h687hb6h678h5d4h549h54h98h8ch564h5bh5bch55dhc8hf8q,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:swift,ReadOnly:false,MountPath:/srv/node/pv,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:etc-swift,ReadOnly:false,MountPath:/etc/swift,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cache,ReadOnly:false,MountPath:/var/cache/swift,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:lock,ReadOnly:false,MountPath:/var/lock,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xr9g4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42445,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-storage-0_openstack(0ffb0da2-affb-4a95-87f3-eac442dc754d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 10:48:28 crc kubenswrapper[4821]: I1125 10:48:28.113135 4821 generic.go:334] "Generic (PLEG): container finished" podID="a764b1fe-69e0-4609-814e-ab2d4e8b4438" containerID="93f388ffcf02e4a8292f29e53d8423f497f064c3108c92b59c7e7f8ea4230679" exitCode=0 Nov 25 10:48:28 crc kubenswrapper[4821]: I1125 10:48:28.117571 4821 generic.go:334] "Generic (PLEG): container finished" podID="ba88f146-f158-4703-b77d-319f4f65d419" containerID="19b20fe65f03e445fec02e7da2a869b70f0effcc221783c3f191c8e4dde87178" exitCode=0 Nov 25 10:48:28 crc kubenswrapper[4821]: I1125 10:48:28.119850 4821 generic.go:334] "Generic (PLEG): container finished" podID="2d5f3fd0-3830-4079-bbf2-0968a8a454f1" containerID="eef603c99d063234ffafa903350973a5c52aabccededa1c1088b6e9c03b89934" exitCode=0 Nov 25 10:48:28 crc kubenswrapper[4821]: I1125 10:48:28.123521 4821 generic.go:334] "Generic (PLEG): container finished" podID="65d5f171-33f0-4c2d-9cfb-6847f2b98c49" containerID="3c1126a2eb0007e787199e1622a82fcd8e726c76a4cdb353b4ac66d5433a61dd" exitCode=0 Nov 25 10:48:28 crc kubenswrapper[4821]: I1125 10:48:28.125112 4821 generic.go:334] "Generic (PLEG): container finished" podID="cec42c9b-c925-4248-b73c-a076d1d0a8c4" containerID="2a62662f27a8cfece54ff23a865831648780fa1fb785af45951a406accc7c6a9" exitCode=0 Nov 25 10:48:28 crc kubenswrapper[4821]: I1125 10:48:28.125670 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-ff23-account-create-n68t7" event={"ID":"a764b1fe-69e0-4609-814e-ab2d4e8b4438","Type":"ContainerDied","Data":"93f388ffcf02e4a8292f29e53d8423f497f064c3108c92b59c7e7f8ea4230679"} Nov 25 10:48:28 crc kubenswrapper[4821]: I1125 10:48:28.125709 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7da5-account-create-8wqv6" event={"ID":"ba88f146-f158-4703-b77d-319f4f65d419","Type":"ContainerDied","Data":"19b20fe65f03e445fec02e7da2a869b70f0effcc221783c3f191c8e4dde87178"} Nov 25 10:48:28 crc kubenswrapper[4821]: I1125 10:48:28.125723 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-xpctj" event={"ID":"2d5f3fd0-3830-4079-bbf2-0968a8a454f1","Type":"ContainerDied","Data":"eef603c99d063234ffafa903350973a5c52aabccededa1c1088b6e9c03b89934"} Nov 25 10:48:28 crc kubenswrapper[4821]: I1125 10:48:28.125825 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-a9df-account-create-xlkf6" event={"ID":"65d5f171-33f0-4c2d-9cfb-6847f2b98c49","Type":"ContainerDied","Data":"3c1126a2eb0007e787199e1622a82fcd8e726c76a4cdb353b4ac66d5433a61dd"} Nov 25 10:48:28 crc kubenswrapper[4821]: I1125 10:48:28.125863 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-55n47" event={"ID":"cec42c9b-c925-4248-b73c-a076d1d0a8c4","Type":"ContainerDied","Data":"2a62662f27a8cfece54ff23a865831648780fa1fb785af45951a406accc7c6a9"} Nov 25 10:48:28 crc kubenswrapper[4821]: I1125 10:48:28.128315 4821 generic.go:334] "Generic (PLEG): container finished" podID="3049cd6a-c38c-4da9-83ce-2195225c3ac9" containerID="71f2dc7f92b4c1969d58583c104e8d6f5aae05207de27f3adf8134f2ab203e20" exitCode=0 Nov 25 10:48:28 crc kubenswrapper[4821]: I1125 10:48:28.128361 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-sq6xx" event={"ID":"3049cd6a-c38c-4da9-83ce-2195225c3ac9","Type":"ContainerDied","Data":"71f2dc7f92b4c1969d58583c104e8d6f5aae05207de27f3adf8134f2ab203e20"} Nov 25 10:48:30 crc kubenswrapper[4821]: E1125 10:48:30.431954 4821 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-keystone:current-podified" Nov 25 10:48:30 crc kubenswrapper[4821]: E1125 10:48:30.432610 4821 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:keystone-db-sync,Image:quay.io/podified-antelope-centos9/openstack-keystone:current-podified,Command:[/bin/bash],Args:[-c keystone-manage db_sync],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/keystone/keystone.conf,SubPath:keystone.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lltzf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42425,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42425,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-db-sync-bhnq8_openstack(dbdf62cf-071e-4e1c-9ef6-22bc39cfa795): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 10:48:30 crc kubenswrapper[4821]: E1125 10:48:30.433928 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"keystone-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/keystone-db-sync-bhnq8" podUID="dbdf62cf-071e-4e1c-9ef6-22bc39cfa795" Nov 25 10:48:30 crc kubenswrapper[4821]: I1125 10:48:30.537937 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7da5-account-create-8wqv6" Nov 25 10:48:30 crc kubenswrapper[4821]: I1125 10:48:30.545525 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-a9df-account-create-xlkf6" Nov 25 10:48:30 crc kubenswrapper[4821]: I1125 10:48:30.558538 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-ff23-account-create-n68t7" Nov 25 10:48:30 crc kubenswrapper[4821]: I1125 10:48:30.580640 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-sq6xx" Nov 25 10:48:30 crc kubenswrapper[4821]: I1125 10:48:30.588460 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-xpctj" Nov 25 10:48:30 crc kubenswrapper[4821]: I1125 10:48:30.599473 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-55n47" Nov 25 10:48:30 crc kubenswrapper[4821]: I1125 10:48:30.676725 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9k9ww\" (UniqueName: \"kubernetes.io/projected/65d5f171-33f0-4c2d-9cfb-6847f2b98c49-kube-api-access-9k9ww\") pod \"65d5f171-33f0-4c2d-9cfb-6847f2b98c49\" (UID: \"65d5f171-33f0-4c2d-9cfb-6847f2b98c49\") " Nov 25 10:48:30 crc kubenswrapper[4821]: I1125 10:48:30.676765 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-85mx6\" (UniqueName: \"kubernetes.io/projected/a764b1fe-69e0-4609-814e-ab2d4e8b4438-kube-api-access-85mx6\") pod \"a764b1fe-69e0-4609-814e-ab2d4e8b4438\" (UID: \"a764b1fe-69e0-4609-814e-ab2d4e8b4438\") " Nov 25 10:48:30 crc kubenswrapper[4821]: I1125 10:48:30.676835 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a764b1fe-69e0-4609-814e-ab2d4e8b4438-operator-scripts\") pod \"a764b1fe-69e0-4609-814e-ab2d4e8b4438\" (UID: \"a764b1fe-69e0-4609-814e-ab2d4e8b4438\") " Nov 25 10:48:30 crc kubenswrapper[4821]: I1125 10:48:30.676888 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8v9gk\" (UniqueName: \"kubernetes.io/projected/ba88f146-f158-4703-b77d-319f4f65d419-kube-api-access-8v9gk\") pod \"ba88f146-f158-4703-b77d-319f4f65d419\" (UID: \"ba88f146-f158-4703-b77d-319f4f65d419\") " Nov 25 10:48:30 crc kubenswrapper[4821]: I1125 10:48:30.676929 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d5f3fd0-3830-4079-bbf2-0968a8a454f1-operator-scripts\") pod \"2d5f3fd0-3830-4079-bbf2-0968a8a454f1\" (UID: \"2d5f3fd0-3830-4079-bbf2-0968a8a454f1\") " Nov 25 10:48:30 crc kubenswrapper[4821]: I1125 10:48:30.676959 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dffwm\" (UniqueName: \"kubernetes.io/projected/3049cd6a-c38c-4da9-83ce-2195225c3ac9-kube-api-access-dffwm\") pod \"3049cd6a-c38c-4da9-83ce-2195225c3ac9\" (UID: \"3049cd6a-c38c-4da9-83ce-2195225c3ac9\") " Nov 25 10:48:30 crc kubenswrapper[4821]: I1125 10:48:30.676978 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gnrqm\" (UniqueName: \"kubernetes.io/projected/2d5f3fd0-3830-4079-bbf2-0968a8a454f1-kube-api-access-gnrqm\") pod \"2d5f3fd0-3830-4079-bbf2-0968a8a454f1\" (UID: \"2d5f3fd0-3830-4079-bbf2-0968a8a454f1\") " Nov 25 10:48:30 crc kubenswrapper[4821]: I1125 10:48:30.677021 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/65d5f171-33f0-4c2d-9cfb-6847f2b98c49-operator-scripts\") pod \"65d5f171-33f0-4c2d-9cfb-6847f2b98c49\" (UID: \"65d5f171-33f0-4c2d-9cfb-6847f2b98c49\") " Nov 25 10:48:30 crc kubenswrapper[4821]: I1125 10:48:30.677036 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ba88f146-f158-4703-b77d-319f4f65d419-operator-scripts\") pod \"ba88f146-f158-4703-b77d-319f4f65d419\" (UID: \"ba88f146-f158-4703-b77d-319f4f65d419\") " Nov 25 10:48:30 crc kubenswrapper[4821]: I1125 10:48:30.677054 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3049cd6a-c38c-4da9-83ce-2195225c3ac9-operator-scripts\") pod \"3049cd6a-c38c-4da9-83ce-2195225c3ac9\" (UID: \"3049cd6a-c38c-4da9-83ce-2195225c3ac9\") " Nov 25 10:48:30 crc kubenswrapper[4821]: I1125 10:48:30.678058 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3049cd6a-c38c-4da9-83ce-2195225c3ac9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3049cd6a-c38c-4da9-83ce-2195225c3ac9" (UID: "3049cd6a-c38c-4da9-83ce-2195225c3ac9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:48:30 crc kubenswrapper[4821]: I1125 10:48:30.679107 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a764b1fe-69e0-4609-814e-ab2d4e8b4438-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a764b1fe-69e0-4609-814e-ab2d4e8b4438" (UID: "a764b1fe-69e0-4609-814e-ab2d4e8b4438"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:48:30 crc kubenswrapper[4821]: I1125 10:48:30.679431 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d5f3fd0-3830-4079-bbf2-0968a8a454f1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2d5f3fd0-3830-4079-bbf2-0968a8a454f1" (UID: "2d5f3fd0-3830-4079-bbf2-0968a8a454f1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:48:30 crc kubenswrapper[4821]: I1125 10:48:30.679505 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/65d5f171-33f0-4c2d-9cfb-6847f2b98c49-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "65d5f171-33f0-4c2d-9cfb-6847f2b98c49" (UID: "65d5f171-33f0-4c2d-9cfb-6847f2b98c49"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:48:30 crc kubenswrapper[4821]: I1125 10:48:30.679822 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ba88f146-f158-4703-b77d-319f4f65d419-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ba88f146-f158-4703-b77d-319f4f65d419" (UID: "ba88f146-f158-4703-b77d-319f4f65d419"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:48:30 crc kubenswrapper[4821]: I1125 10:48:30.683611 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3049cd6a-c38c-4da9-83ce-2195225c3ac9-kube-api-access-dffwm" (OuterVolumeSpecName: "kube-api-access-dffwm") pod "3049cd6a-c38c-4da9-83ce-2195225c3ac9" (UID: "3049cd6a-c38c-4da9-83ce-2195225c3ac9"). InnerVolumeSpecName "kube-api-access-dffwm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:48:30 crc kubenswrapper[4821]: I1125 10:48:30.683897 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a764b1fe-69e0-4609-814e-ab2d4e8b4438-kube-api-access-85mx6" (OuterVolumeSpecName: "kube-api-access-85mx6") pod "a764b1fe-69e0-4609-814e-ab2d4e8b4438" (UID: "a764b1fe-69e0-4609-814e-ab2d4e8b4438"). InnerVolumeSpecName "kube-api-access-85mx6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:48:30 crc kubenswrapper[4821]: I1125 10:48:30.684321 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba88f146-f158-4703-b77d-319f4f65d419-kube-api-access-8v9gk" (OuterVolumeSpecName: "kube-api-access-8v9gk") pod "ba88f146-f158-4703-b77d-319f4f65d419" (UID: "ba88f146-f158-4703-b77d-319f4f65d419"). InnerVolumeSpecName "kube-api-access-8v9gk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:48:30 crc kubenswrapper[4821]: I1125 10:48:30.684438 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65d5f171-33f0-4c2d-9cfb-6847f2b98c49-kube-api-access-9k9ww" (OuterVolumeSpecName: "kube-api-access-9k9ww") pod "65d5f171-33f0-4c2d-9cfb-6847f2b98c49" (UID: "65d5f171-33f0-4c2d-9cfb-6847f2b98c49"). InnerVolumeSpecName "kube-api-access-9k9ww". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:48:30 crc kubenswrapper[4821]: I1125 10:48:30.687233 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d5f3fd0-3830-4079-bbf2-0968a8a454f1-kube-api-access-gnrqm" (OuterVolumeSpecName: "kube-api-access-gnrqm") pod "2d5f3fd0-3830-4079-bbf2-0968a8a454f1" (UID: "2d5f3fd0-3830-4079-bbf2-0968a8a454f1"). InnerVolumeSpecName "kube-api-access-gnrqm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:48:30 crc kubenswrapper[4821]: I1125 10:48:30.778549 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cec42c9b-c925-4248-b73c-a076d1d0a8c4-operator-scripts\") pod \"cec42c9b-c925-4248-b73c-a076d1d0a8c4\" (UID: \"cec42c9b-c925-4248-b73c-a076d1d0a8c4\") " Nov 25 10:48:30 crc kubenswrapper[4821]: I1125 10:48:30.778919 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gvd62\" (UniqueName: \"kubernetes.io/projected/cec42c9b-c925-4248-b73c-a076d1d0a8c4-kube-api-access-gvd62\") pod \"cec42c9b-c925-4248-b73c-a076d1d0a8c4\" (UID: \"cec42c9b-c925-4248-b73c-a076d1d0a8c4\") " Nov 25 10:48:30 crc kubenswrapper[4821]: I1125 10:48:30.779305 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9k9ww\" (UniqueName: \"kubernetes.io/projected/65d5f171-33f0-4c2d-9cfb-6847f2b98c49-kube-api-access-9k9ww\") on node \"crc\" DevicePath \"\"" Nov 25 10:48:30 crc kubenswrapper[4821]: I1125 10:48:30.779324 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-85mx6\" (UniqueName: \"kubernetes.io/projected/a764b1fe-69e0-4609-814e-ab2d4e8b4438-kube-api-access-85mx6\") on node \"crc\" DevicePath \"\"" Nov 25 10:48:30 crc kubenswrapper[4821]: I1125 10:48:30.779336 4821 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a764b1fe-69e0-4609-814e-ab2d4e8b4438-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:48:30 crc kubenswrapper[4821]: I1125 10:48:30.779350 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8v9gk\" (UniqueName: \"kubernetes.io/projected/ba88f146-f158-4703-b77d-319f4f65d419-kube-api-access-8v9gk\") on node \"crc\" DevicePath \"\"" Nov 25 10:48:30 crc kubenswrapper[4821]: I1125 10:48:30.779360 4821 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d5f3fd0-3830-4079-bbf2-0968a8a454f1-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:48:30 crc kubenswrapper[4821]: I1125 10:48:30.779369 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dffwm\" (UniqueName: \"kubernetes.io/projected/3049cd6a-c38c-4da9-83ce-2195225c3ac9-kube-api-access-dffwm\") on node \"crc\" DevicePath \"\"" Nov 25 10:48:30 crc kubenswrapper[4821]: I1125 10:48:30.779378 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gnrqm\" (UniqueName: \"kubernetes.io/projected/2d5f3fd0-3830-4079-bbf2-0968a8a454f1-kube-api-access-gnrqm\") on node \"crc\" DevicePath \"\"" Nov 25 10:48:30 crc kubenswrapper[4821]: I1125 10:48:30.779414 4821 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/65d5f171-33f0-4c2d-9cfb-6847f2b98c49-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:48:30 crc kubenswrapper[4821]: I1125 10:48:30.779422 4821 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ba88f146-f158-4703-b77d-319f4f65d419-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:48:30 crc kubenswrapper[4821]: I1125 10:48:30.779431 4821 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3049cd6a-c38c-4da9-83ce-2195225c3ac9-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:48:30 crc kubenswrapper[4821]: I1125 10:48:30.780633 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cec42c9b-c925-4248-b73c-a076d1d0a8c4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "cec42c9b-c925-4248-b73c-a076d1d0a8c4" (UID: "cec42c9b-c925-4248-b73c-a076d1d0a8c4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:48:30 crc kubenswrapper[4821]: I1125 10:48:30.782612 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cec42c9b-c925-4248-b73c-a076d1d0a8c4-kube-api-access-gvd62" (OuterVolumeSpecName: "kube-api-access-gvd62") pod "cec42c9b-c925-4248-b73c-a076d1d0a8c4" (UID: "cec42c9b-c925-4248-b73c-a076d1d0a8c4"). InnerVolumeSpecName "kube-api-access-gvd62". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:48:30 crc kubenswrapper[4821]: I1125 10:48:30.881600 4821 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cec42c9b-c925-4248-b73c-a076d1d0a8c4-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:48:30 crc kubenswrapper[4821]: I1125 10:48:30.881630 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gvd62\" (UniqueName: \"kubernetes.io/projected/cec42c9b-c925-4248-b73c-a076d1d0a8c4-kube-api-access-gvd62\") on node \"crc\" DevicePath \"\"" Nov 25 10:48:31 crc kubenswrapper[4821]: I1125 10:48:31.171766 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-ff23-account-create-n68t7" event={"ID":"a764b1fe-69e0-4609-814e-ab2d4e8b4438","Type":"ContainerDied","Data":"9c88689df8d8df71b51ff38caf6c0ffb03a0fb6f84bc8fb904ea51bf58904702"} Nov 25 10:48:31 crc kubenswrapper[4821]: I1125 10:48:31.171814 4821 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9c88689df8d8df71b51ff38caf6c0ffb03a0fb6f84bc8fb904ea51bf58904702" Nov 25 10:48:31 crc kubenswrapper[4821]: I1125 10:48:31.171884 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-ff23-account-create-n68t7" Nov 25 10:48:31 crc kubenswrapper[4821]: I1125 10:48:31.178677 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7da5-account-create-8wqv6" event={"ID":"ba88f146-f158-4703-b77d-319f4f65d419","Type":"ContainerDied","Data":"6b2011cd2c7284ba0bfb6c3bc27da8a28ef3cd07fe40398a9b41940ce2706445"} Nov 25 10:48:31 crc kubenswrapper[4821]: I1125 10:48:31.178721 4821 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6b2011cd2c7284ba0bfb6c3bc27da8a28ef3cd07fe40398a9b41940ce2706445" Nov 25 10:48:31 crc kubenswrapper[4821]: I1125 10:48:31.178724 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7da5-account-create-8wqv6" Nov 25 10:48:31 crc kubenswrapper[4821]: I1125 10:48:31.181227 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-xpctj" event={"ID":"2d5f3fd0-3830-4079-bbf2-0968a8a454f1","Type":"ContainerDied","Data":"f155827d660d30e178d3aa1fb7620c6e2a75c0e624f00e429b697771ad6c565c"} Nov 25 10:48:31 crc kubenswrapper[4821]: I1125 10:48:31.181244 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-xpctj" Nov 25 10:48:31 crc kubenswrapper[4821]: I1125 10:48:31.181259 4821 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f155827d660d30e178d3aa1fb7620c6e2a75c0e624f00e429b697771ad6c565c" Nov 25 10:48:31 crc kubenswrapper[4821]: I1125 10:48:31.190119 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-a9df-account-create-xlkf6" event={"ID":"65d5f171-33f0-4c2d-9cfb-6847f2b98c49","Type":"ContainerDied","Data":"30100d28a5b8712886362d24e87df659b47b082870e8ddf7e262f3d6845c037b"} Nov 25 10:48:31 crc kubenswrapper[4821]: I1125 10:48:31.190304 4821 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="30100d28a5b8712886362d24e87df659b47b082870e8ddf7e262f3d6845c037b" Nov 25 10:48:31 crc kubenswrapper[4821]: I1125 10:48:31.190371 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-a9df-account-create-xlkf6" Nov 25 10:48:31 crc kubenswrapper[4821]: I1125 10:48:31.193071 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-55n47" event={"ID":"cec42c9b-c925-4248-b73c-a076d1d0a8c4","Type":"ContainerDied","Data":"db23f03f37e137ea5feb795b3e03def85f4c1d1fdef0e9ceb4ede145361364ab"} Nov 25 10:48:31 crc kubenswrapper[4821]: I1125 10:48:31.193181 4821 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="db23f03f37e137ea5feb795b3e03def85f4c1d1fdef0e9ceb4ede145361364ab" Nov 25 10:48:31 crc kubenswrapper[4821]: I1125 10:48:31.193274 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-55n47" Nov 25 10:48:31 crc kubenswrapper[4821]: I1125 10:48:31.195696 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-sq6xx" event={"ID":"3049cd6a-c38c-4da9-83ce-2195225c3ac9","Type":"ContainerDied","Data":"6e2a422a6331b073c6009847649f8d7a5d00e34756d10994b6910f9e5bac9bae"} Nov 25 10:48:31 crc kubenswrapper[4821]: I1125 10:48:31.195921 4821 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6e2a422a6331b073c6009847649f8d7a5d00e34756d10994b6910f9e5bac9bae" Nov 25 10:48:31 crc kubenswrapper[4821]: I1125 10:48:31.195899 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-sq6xx" Nov 25 10:48:31 crc kubenswrapper[4821]: I1125 10:48:31.198973 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0ffb0da2-affb-4a95-87f3-eac442dc754d","Type":"ContainerStarted","Data":"50f411e1010752e15ef0fb8aa73554f4f219463757ec41baddd22893c02e48e4"} Nov 25 10:48:31 crc kubenswrapper[4821]: E1125 10:48:31.202584 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"keystone-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-keystone:current-podified\\\"\"" pod="openstack/keystone-db-sync-bhnq8" podUID="dbdf62cf-071e-4e1c-9ef6-22bc39cfa795" Nov 25 10:48:32 crc kubenswrapper[4821]: I1125 10:48:32.208295 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0ffb0da2-affb-4a95-87f3-eac442dc754d","Type":"ContainerStarted","Data":"4aa760c9a621945e813e4f7cd00c2bad542eee3e9ef18dbed1cfd12d516200ae"} Nov 25 10:48:32 crc kubenswrapper[4821]: I1125 10:48:32.208658 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0ffb0da2-affb-4a95-87f3-eac442dc754d","Type":"ContainerStarted","Data":"781fad2d9b8026772e271f180df3add961aec07e6c3b07a22d13f57278fc460a"} Nov 25 10:48:32 crc kubenswrapper[4821]: I1125 10:48:32.208672 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0ffb0da2-affb-4a95-87f3-eac442dc754d","Type":"ContainerStarted","Data":"be7f54977b9bec4f51649d4a6105f09dd67e53d0ac9ca2db55a7ff7852331fb5"} Nov 25 10:48:33 crc kubenswrapper[4821]: I1125 10:48:33.221061 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0ffb0da2-affb-4a95-87f3-eac442dc754d","Type":"ContainerStarted","Data":"2ea52bc959cc493a389039d576fd36c367a15a98f9d211e38fade9f9a14f1d8a"} Nov 25 10:48:33 crc kubenswrapper[4821]: I1125 10:48:33.221357 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0ffb0da2-affb-4a95-87f3-eac442dc754d","Type":"ContainerStarted","Data":"0d71318f8277d57333abd8283f50bd0508517f8267ecfc7f688010b4328a57b0"} Nov 25 10:48:33 crc kubenswrapper[4821]: I1125 10:48:33.221368 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0ffb0da2-affb-4a95-87f3-eac442dc754d","Type":"ContainerStarted","Data":"b7af6c350e31a04a1156ce96cd783788f8117cfa8a5768d781118fa6e0659642"} Nov 25 10:48:33 crc kubenswrapper[4821]: E1125 10:48:33.934922 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"account-server\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"account-replicator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-swift-account:current-podified\\\"\", failed to \"StartContainer\" for \"account-auditor\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-swift-account:current-podified\\\"\", failed to \"StartContainer\" for \"account-reaper\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-swift-account:current-podified\\\"\"]" pod="openstack/swift-storage-0" podUID="0ffb0da2-affb-4a95-87f3-eac442dc754d" Nov 25 10:48:34 crc kubenswrapper[4821]: I1125 10:48:34.251892 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0ffb0da2-affb-4a95-87f3-eac442dc754d","Type":"ContainerStarted","Data":"533a691fd141bf25fe6c3eee52e57d4d328b7ad58ffee3bb9a69c7bdf344c68b"} Nov 25 10:48:34 crc kubenswrapper[4821]: I1125 10:48:34.251947 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0ffb0da2-affb-4a95-87f3-eac442dc754d","Type":"ContainerStarted","Data":"57d932f6491e0b0f6efc76d11a6d1038a643f210a72a9b10f831579dc35b5e04"} Nov 25 10:48:34 crc kubenswrapper[4821]: I1125 10:48:34.251959 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0ffb0da2-affb-4a95-87f3-eac442dc754d","Type":"ContainerStarted","Data":"e15e141b27080e846121fc4ff3d2bb67e9e3790b5181ff626b4d8412bb054ca6"} Nov 25 10:48:34 crc kubenswrapper[4821]: I1125 10:48:34.251971 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0ffb0da2-affb-4a95-87f3-eac442dc754d","Type":"ContainerStarted","Data":"8a07936656b6906ea31227288dc401fcc50f15a38d01ad5a1acf73cbf1b57213"} Nov 25 10:48:34 crc kubenswrapper[4821]: E1125 10:48:34.259374 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"account-server\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-swift-account:current-podified\\\"\", failed to \"StartContainer\" for \"account-replicator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-swift-account:current-podified\\\"\", failed to \"StartContainer\" for \"account-auditor\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-swift-account:current-podified\\\"\", failed to \"StartContainer\" for \"account-reaper\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-swift-account:current-podified\\\"\"]" pod="openstack/swift-storage-0" podUID="0ffb0da2-affb-4a95-87f3-eac442dc754d" Nov 25 10:48:34 crc kubenswrapper[4821]: I1125 10:48:34.715479 4821 patch_prober.go:28] interesting pod/machine-config-daemon-2krbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:48:34 crc kubenswrapper[4821]: I1125 10:48:34.716133 4821 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:48:34 crc kubenswrapper[4821]: I1125 10:48:34.716292 4821 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" Nov 25 10:48:34 crc kubenswrapper[4821]: I1125 10:48:34.717129 4821 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c1742b8b51dd841b8f5beefd1352b72fe8669e2095b0356503a696291b279fa5"} pod="openshift-machine-config-operator/machine-config-daemon-2krbf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 10:48:34 crc kubenswrapper[4821]: I1125 10:48:34.717319 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerName="machine-config-daemon" containerID="cri-o://c1742b8b51dd841b8f5beefd1352b72fe8669e2095b0356503a696291b279fa5" gracePeriod=600 Nov 25 10:48:35 crc kubenswrapper[4821]: I1125 10:48:35.260794 4821 generic.go:334] "Generic (PLEG): container finished" podID="a8123055-1953-42b3-9c04-716061499d03" containerID="c8ccf9a6a3df8286261caebc19be09c2fc61f8c79473aded6a8b5d775e5bfc6a" exitCode=0 Nov 25 10:48:35 crc kubenswrapper[4821]: I1125 10:48:35.260897 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-2nhpq" event={"ID":"a8123055-1953-42b3-9c04-716061499d03","Type":"ContainerDied","Data":"c8ccf9a6a3df8286261caebc19be09c2fc61f8c79473aded6a8b5d775e5bfc6a"} Nov 25 10:48:35 crc kubenswrapper[4821]: I1125 10:48:35.266080 4821 generic.go:334] "Generic (PLEG): container finished" podID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerID="c1742b8b51dd841b8f5beefd1352b72fe8669e2095b0356503a696291b279fa5" exitCode=0 Nov 25 10:48:35 crc kubenswrapper[4821]: I1125 10:48:35.266178 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" event={"ID":"5f948b87-ac86-4de6-ad64-c2ef947f84d4","Type":"ContainerDied","Data":"c1742b8b51dd841b8f5beefd1352b72fe8669e2095b0356503a696291b279fa5"} Nov 25 10:48:35 crc kubenswrapper[4821]: I1125 10:48:35.266223 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" event={"ID":"5f948b87-ac86-4de6-ad64-c2ef947f84d4","Type":"ContainerStarted","Data":"8e91d42cd72af932de083af71bf0c37ad5242c32b692e67d8b88f724fd1c26a8"} Nov 25 10:48:35 crc kubenswrapper[4821]: I1125 10:48:35.266244 4821 scope.go:117] "RemoveContainer" containerID="d231cd7c0910d85e280090ca76750286ce837c9e610ff6ebf83f2f5b9e078c96" Nov 25 10:48:35 crc kubenswrapper[4821]: E1125 10:48:35.273685 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"account-server\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-swift-account:current-podified\\\"\", failed to \"StartContainer\" for \"account-replicator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-swift-account:current-podified\\\"\", failed to \"StartContainer\" for \"account-auditor\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-swift-account:current-podified\\\"\", failed to \"StartContainer\" for \"account-reaper\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-swift-account:current-podified\\\"\"]" pod="openstack/swift-storage-0" podUID="0ffb0da2-affb-4a95-87f3-eac442dc754d" Nov 25 10:48:36 crc kubenswrapper[4821]: I1125 10:48:36.637924 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-2nhpq" Nov 25 10:48:36 crc kubenswrapper[4821]: I1125 10:48:36.775752 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8123055-1953-42b3-9c04-716061499d03-combined-ca-bundle\") pod \"a8123055-1953-42b3-9c04-716061499d03\" (UID: \"a8123055-1953-42b3-9c04-716061499d03\") " Nov 25 10:48:36 crc kubenswrapper[4821]: I1125 10:48:36.775862 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cgxhq\" (UniqueName: \"kubernetes.io/projected/a8123055-1953-42b3-9c04-716061499d03-kube-api-access-cgxhq\") pod \"a8123055-1953-42b3-9c04-716061499d03\" (UID: \"a8123055-1953-42b3-9c04-716061499d03\") " Nov 25 10:48:36 crc kubenswrapper[4821]: I1125 10:48:36.775905 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8123055-1953-42b3-9c04-716061499d03-config-data\") pod \"a8123055-1953-42b3-9c04-716061499d03\" (UID: \"a8123055-1953-42b3-9c04-716061499d03\") " Nov 25 10:48:36 crc kubenswrapper[4821]: I1125 10:48:36.775976 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a8123055-1953-42b3-9c04-716061499d03-db-sync-config-data\") pod \"a8123055-1953-42b3-9c04-716061499d03\" (UID: \"a8123055-1953-42b3-9c04-716061499d03\") " Nov 25 10:48:36 crc kubenswrapper[4821]: I1125 10:48:36.781927 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8123055-1953-42b3-9c04-716061499d03-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "a8123055-1953-42b3-9c04-716061499d03" (UID: "a8123055-1953-42b3-9c04-716061499d03"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:48:36 crc kubenswrapper[4821]: I1125 10:48:36.782566 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8123055-1953-42b3-9c04-716061499d03-kube-api-access-cgxhq" (OuterVolumeSpecName: "kube-api-access-cgxhq") pod "a8123055-1953-42b3-9c04-716061499d03" (UID: "a8123055-1953-42b3-9c04-716061499d03"). InnerVolumeSpecName "kube-api-access-cgxhq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:48:36 crc kubenswrapper[4821]: I1125 10:48:36.798899 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8123055-1953-42b3-9c04-716061499d03-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a8123055-1953-42b3-9c04-716061499d03" (UID: "a8123055-1953-42b3-9c04-716061499d03"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:48:36 crc kubenswrapper[4821]: I1125 10:48:36.819886 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8123055-1953-42b3-9c04-716061499d03-config-data" (OuterVolumeSpecName: "config-data") pod "a8123055-1953-42b3-9c04-716061499d03" (UID: "a8123055-1953-42b3-9c04-716061499d03"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:48:36 crc kubenswrapper[4821]: I1125 10:48:36.878975 4821 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8123055-1953-42b3-9c04-716061499d03-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:48:36 crc kubenswrapper[4821]: I1125 10:48:36.879017 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cgxhq\" (UniqueName: \"kubernetes.io/projected/a8123055-1953-42b3-9c04-716061499d03-kube-api-access-cgxhq\") on node \"crc\" DevicePath \"\"" Nov 25 10:48:36 crc kubenswrapper[4821]: I1125 10:48:36.879028 4821 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8123055-1953-42b3-9c04-716061499d03-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:48:36 crc kubenswrapper[4821]: I1125 10:48:36.879038 4821 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a8123055-1953-42b3-9c04-716061499d03-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:48:37 crc kubenswrapper[4821]: I1125 10:48:37.285802 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-2nhpq" event={"ID":"a8123055-1953-42b3-9c04-716061499d03","Type":"ContainerDied","Data":"7bb6c92ed8e23690d9e6b3c4ba5501456d37cfed8991b3e2f91c3a6509bf8b58"} Nov 25 10:48:37 crc kubenswrapper[4821]: I1125 10:48:37.286471 4821 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7bb6c92ed8e23690d9e6b3c4ba5501456d37cfed8991b3e2f91c3a6509bf8b58" Nov 25 10:48:37 crc kubenswrapper[4821]: I1125 10:48:37.286440 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-2nhpq" Nov 25 10:48:37 crc kubenswrapper[4821]: I1125 10:48:37.636637 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-74dc88fc-bqbsk"] Nov 25 10:48:37 crc kubenswrapper[4821]: E1125 10:48:37.636980 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a764b1fe-69e0-4609-814e-ab2d4e8b4438" containerName="mariadb-account-create" Nov 25 10:48:37 crc kubenswrapper[4821]: I1125 10:48:37.637003 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="a764b1fe-69e0-4609-814e-ab2d4e8b4438" containerName="mariadb-account-create" Nov 25 10:48:37 crc kubenswrapper[4821]: E1125 10:48:37.637017 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3049cd6a-c38c-4da9-83ce-2195225c3ac9" containerName="mariadb-database-create" Nov 25 10:48:37 crc kubenswrapper[4821]: I1125 10:48:37.637023 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="3049cd6a-c38c-4da9-83ce-2195225c3ac9" containerName="mariadb-database-create" Nov 25 10:48:37 crc kubenswrapper[4821]: E1125 10:48:37.637037 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8123055-1953-42b3-9c04-716061499d03" containerName="glance-db-sync" Nov 25 10:48:37 crc kubenswrapper[4821]: I1125 10:48:37.637045 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8123055-1953-42b3-9c04-716061499d03" containerName="glance-db-sync" Nov 25 10:48:37 crc kubenswrapper[4821]: E1125 10:48:37.637057 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d5f3fd0-3830-4079-bbf2-0968a8a454f1" containerName="mariadb-database-create" Nov 25 10:48:37 crc kubenswrapper[4821]: I1125 10:48:37.637062 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d5f3fd0-3830-4079-bbf2-0968a8a454f1" containerName="mariadb-database-create" Nov 25 10:48:37 crc kubenswrapper[4821]: E1125 10:48:37.637076 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cec42c9b-c925-4248-b73c-a076d1d0a8c4" containerName="mariadb-database-create" Nov 25 10:48:37 crc kubenswrapper[4821]: I1125 10:48:37.637083 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="cec42c9b-c925-4248-b73c-a076d1d0a8c4" containerName="mariadb-database-create" Nov 25 10:48:37 crc kubenswrapper[4821]: E1125 10:48:37.637096 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba88f146-f158-4703-b77d-319f4f65d419" containerName="mariadb-account-create" Nov 25 10:48:37 crc kubenswrapper[4821]: I1125 10:48:37.637101 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba88f146-f158-4703-b77d-319f4f65d419" containerName="mariadb-account-create" Nov 25 10:48:37 crc kubenswrapper[4821]: E1125 10:48:37.637120 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65d5f171-33f0-4c2d-9cfb-6847f2b98c49" containerName="mariadb-account-create" Nov 25 10:48:37 crc kubenswrapper[4821]: I1125 10:48:37.637127 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="65d5f171-33f0-4c2d-9cfb-6847f2b98c49" containerName="mariadb-account-create" Nov 25 10:48:37 crc kubenswrapper[4821]: I1125 10:48:37.637315 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="cec42c9b-c925-4248-b73c-a076d1d0a8c4" containerName="mariadb-database-create" Nov 25 10:48:37 crc kubenswrapper[4821]: I1125 10:48:37.637334 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="65d5f171-33f0-4c2d-9cfb-6847f2b98c49" containerName="mariadb-account-create" Nov 25 10:48:37 crc kubenswrapper[4821]: I1125 10:48:37.637351 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d5f3fd0-3830-4079-bbf2-0968a8a454f1" containerName="mariadb-database-create" Nov 25 10:48:37 crc kubenswrapper[4821]: I1125 10:48:37.637360 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="a764b1fe-69e0-4609-814e-ab2d4e8b4438" containerName="mariadb-account-create" Nov 25 10:48:37 crc kubenswrapper[4821]: I1125 10:48:37.637376 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="3049cd6a-c38c-4da9-83ce-2195225c3ac9" containerName="mariadb-database-create" Nov 25 10:48:37 crc kubenswrapper[4821]: I1125 10:48:37.637391 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8123055-1953-42b3-9c04-716061499d03" containerName="glance-db-sync" Nov 25 10:48:37 crc kubenswrapper[4821]: I1125 10:48:37.637402 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba88f146-f158-4703-b77d-319f4f65d419" containerName="mariadb-account-create" Nov 25 10:48:37 crc kubenswrapper[4821]: I1125 10:48:37.638397 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74dc88fc-bqbsk" Nov 25 10:48:37 crc kubenswrapper[4821]: I1125 10:48:37.658640 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74dc88fc-bqbsk"] Nov 25 10:48:37 crc kubenswrapper[4821]: I1125 10:48:37.794098 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mwzch\" (UniqueName: \"kubernetes.io/projected/ccd85fd6-b0a0-4184-8082-054d2de0adec-kube-api-access-mwzch\") pod \"dnsmasq-dns-74dc88fc-bqbsk\" (UID: \"ccd85fd6-b0a0-4184-8082-054d2de0adec\") " pod="openstack/dnsmasq-dns-74dc88fc-bqbsk" Nov 25 10:48:37 crc kubenswrapper[4821]: I1125 10:48:37.794266 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ccd85fd6-b0a0-4184-8082-054d2de0adec-ovsdbserver-sb\") pod \"dnsmasq-dns-74dc88fc-bqbsk\" (UID: \"ccd85fd6-b0a0-4184-8082-054d2de0adec\") " pod="openstack/dnsmasq-dns-74dc88fc-bqbsk" Nov 25 10:48:37 crc kubenswrapper[4821]: I1125 10:48:37.794309 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ccd85fd6-b0a0-4184-8082-054d2de0adec-ovsdbserver-nb\") pod \"dnsmasq-dns-74dc88fc-bqbsk\" (UID: \"ccd85fd6-b0a0-4184-8082-054d2de0adec\") " pod="openstack/dnsmasq-dns-74dc88fc-bqbsk" Nov 25 10:48:37 crc kubenswrapper[4821]: I1125 10:48:37.794344 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ccd85fd6-b0a0-4184-8082-054d2de0adec-dns-svc\") pod \"dnsmasq-dns-74dc88fc-bqbsk\" (UID: \"ccd85fd6-b0a0-4184-8082-054d2de0adec\") " pod="openstack/dnsmasq-dns-74dc88fc-bqbsk" Nov 25 10:48:37 crc kubenswrapper[4821]: I1125 10:48:37.794429 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ccd85fd6-b0a0-4184-8082-054d2de0adec-config\") pod \"dnsmasq-dns-74dc88fc-bqbsk\" (UID: \"ccd85fd6-b0a0-4184-8082-054d2de0adec\") " pod="openstack/dnsmasq-dns-74dc88fc-bqbsk" Nov 25 10:48:37 crc kubenswrapper[4821]: I1125 10:48:37.895884 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mwzch\" (UniqueName: \"kubernetes.io/projected/ccd85fd6-b0a0-4184-8082-054d2de0adec-kube-api-access-mwzch\") pod \"dnsmasq-dns-74dc88fc-bqbsk\" (UID: \"ccd85fd6-b0a0-4184-8082-054d2de0adec\") " pod="openstack/dnsmasq-dns-74dc88fc-bqbsk" Nov 25 10:48:37 crc kubenswrapper[4821]: I1125 10:48:37.895948 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ccd85fd6-b0a0-4184-8082-054d2de0adec-ovsdbserver-sb\") pod \"dnsmasq-dns-74dc88fc-bqbsk\" (UID: \"ccd85fd6-b0a0-4184-8082-054d2de0adec\") " pod="openstack/dnsmasq-dns-74dc88fc-bqbsk" Nov 25 10:48:37 crc kubenswrapper[4821]: I1125 10:48:37.895974 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ccd85fd6-b0a0-4184-8082-054d2de0adec-ovsdbserver-nb\") pod \"dnsmasq-dns-74dc88fc-bqbsk\" (UID: \"ccd85fd6-b0a0-4184-8082-054d2de0adec\") " pod="openstack/dnsmasq-dns-74dc88fc-bqbsk" Nov 25 10:48:37 crc kubenswrapper[4821]: I1125 10:48:37.896001 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ccd85fd6-b0a0-4184-8082-054d2de0adec-dns-svc\") pod \"dnsmasq-dns-74dc88fc-bqbsk\" (UID: \"ccd85fd6-b0a0-4184-8082-054d2de0adec\") " pod="openstack/dnsmasq-dns-74dc88fc-bqbsk" Nov 25 10:48:37 crc kubenswrapper[4821]: I1125 10:48:37.896045 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ccd85fd6-b0a0-4184-8082-054d2de0adec-config\") pod \"dnsmasq-dns-74dc88fc-bqbsk\" (UID: \"ccd85fd6-b0a0-4184-8082-054d2de0adec\") " pod="openstack/dnsmasq-dns-74dc88fc-bqbsk" Nov 25 10:48:37 crc kubenswrapper[4821]: I1125 10:48:37.897030 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ccd85fd6-b0a0-4184-8082-054d2de0adec-ovsdbserver-nb\") pod \"dnsmasq-dns-74dc88fc-bqbsk\" (UID: \"ccd85fd6-b0a0-4184-8082-054d2de0adec\") " pod="openstack/dnsmasq-dns-74dc88fc-bqbsk" Nov 25 10:48:37 crc kubenswrapper[4821]: I1125 10:48:37.897186 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ccd85fd6-b0a0-4184-8082-054d2de0adec-dns-svc\") pod \"dnsmasq-dns-74dc88fc-bqbsk\" (UID: \"ccd85fd6-b0a0-4184-8082-054d2de0adec\") " pod="openstack/dnsmasq-dns-74dc88fc-bqbsk" Nov 25 10:48:37 crc kubenswrapper[4821]: I1125 10:48:37.897263 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ccd85fd6-b0a0-4184-8082-054d2de0adec-ovsdbserver-sb\") pod \"dnsmasq-dns-74dc88fc-bqbsk\" (UID: \"ccd85fd6-b0a0-4184-8082-054d2de0adec\") " pod="openstack/dnsmasq-dns-74dc88fc-bqbsk" Nov 25 10:48:37 crc kubenswrapper[4821]: I1125 10:48:37.897298 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ccd85fd6-b0a0-4184-8082-054d2de0adec-config\") pod \"dnsmasq-dns-74dc88fc-bqbsk\" (UID: \"ccd85fd6-b0a0-4184-8082-054d2de0adec\") " pod="openstack/dnsmasq-dns-74dc88fc-bqbsk" Nov 25 10:48:37 crc kubenswrapper[4821]: I1125 10:48:37.918896 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mwzch\" (UniqueName: \"kubernetes.io/projected/ccd85fd6-b0a0-4184-8082-054d2de0adec-kube-api-access-mwzch\") pod \"dnsmasq-dns-74dc88fc-bqbsk\" (UID: \"ccd85fd6-b0a0-4184-8082-054d2de0adec\") " pod="openstack/dnsmasq-dns-74dc88fc-bqbsk" Nov 25 10:48:37 crc kubenswrapper[4821]: I1125 10:48:37.963938 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74dc88fc-bqbsk" Nov 25 10:48:38 crc kubenswrapper[4821]: I1125 10:48:38.412242 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74dc88fc-bqbsk"] Nov 25 10:48:38 crc kubenswrapper[4821]: W1125 10:48:38.419416 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podccd85fd6_b0a0_4184_8082_054d2de0adec.slice/crio-6531123427f2e88df3a3269538b8d53b4d6f734bd55cebaf7a49d8b5f65a3169 WatchSource:0}: Error finding container 6531123427f2e88df3a3269538b8d53b4d6f734bd55cebaf7a49d8b5f65a3169: Status 404 returned error can't find the container with id 6531123427f2e88df3a3269538b8d53b4d6f734bd55cebaf7a49d8b5f65a3169 Nov 25 10:48:39 crc kubenswrapper[4821]: I1125 10:48:39.306692 4821 generic.go:334] "Generic (PLEG): container finished" podID="ccd85fd6-b0a0-4184-8082-054d2de0adec" containerID="9be898024013d7da65251768072c3f8f4bbbaa8d591cd6cc751e41d5ee6da326" exitCode=0 Nov 25 10:48:39 crc kubenswrapper[4821]: I1125 10:48:39.306762 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74dc88fc-bqbsk" event={"ID":"ccd85fd6-b0a0-4184-8082-054d2de0adec","Type":"ContainerDied","Data":"9be898024013d7da65251768072c3f8f4bbbaa8d591cd6cc751e41d5ee6da326"} Nov 25 10:48:39 crc kubenswrapper[4821]: I1125 10:48:39.307072 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74dc88fc-bqbsk" event={"ID":"ccd85fd6-b0a0-4184-8082-054d2de0adec","Type":"ContainerStarted","Data":"6531123427f2e88df3a3269538b8d53b4d6f734bd55cebaf7a49d8b5f65a3169"} Nov 25 10:48:40 crc kubenswrapper[4821]: I1125 10:48:40.317225 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74dc88fc-bqbsk" event={"ID":"ccd85fd6-b0a0-4184-8082-054d2de0adec","Type":"ContainerStarted","Data":"67a95be9b3aa84d207a73ce02dd8c799210d3c4070ac21ef95de932bfab350ab"} Nov 25 10:48:40 crc kubenswrapper[4821]: I1125 10:48:40.317524 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-74dc88fc-bqbsk" Nov 25 10:48:40 crc kubenswrapper[4821]: I1125 10:48:40.336880 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-74dc88fc-bqbsk" podStartSLOduration=3.336858619 podStartE2EDuration="3.336858619s" podCreationTimestamp="2025-11-25 10:48:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:48:40.332750376 +0000 UTC m=+990.869070223" watchObservedRunningTime="2025-11-25 10:48:40.336858619 +0000 UTC m=+990.873178476" Nov 25 10:48:45 crc kubenswrapper[4821]: I1125 10:48:45.351730 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-bhnq8" event={"ID":"dbdf62cf-071e-4e1c-9ef6-22bc39cfa795","Type":"ContainerStarted","Data":"4e9bd31c0491a3c78124341f952f334b25e755f3eac5df640fcc5a9961e6359f"} Nov 25 10:48:45 crc kubenswrapper[4821]: I1125 10:48:45.401046 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-bhnq8" podStartSLOduration=2.468558508 podStartE2EDuration="32.401025343s" podCreationTimestamp="2025-11-25 10:48:13 +0000 UTC" firstStartedPulling="2025-11-25 10:48:14.614012501 +0000 UTC m=+965.150332348" lastFinishedPulling="2025-11-25 10:48:44.546479336 +0000 UTC m=+995.082799183" observedRunningTime="2025-11-25 10:48:45.398802737 +0000 UTC m=+995.935122584" watchObservedRunningTime="2025-11-25 10:48:45.401025343 +0000 UTC m=+995.937345190" Nov 25 10:48:47 crc kubenswrapper[4821]: I1125 10:48:47.966410 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-74dc88fc-bqbsk" Nov 25 10:48:48 crc kubenswrapper[4821]: I1125 10:48:48.047771 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-wctvh"] Nov 25 10:48:48 crc kubenswrapper[4821]: I1125 10:48:48.048456 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-b8fbc5445-wctvh" podUID="0333edd1-89f3-4ff0-9da3-544fac7620bb" containerName="dnsmasq-dns" containerID="cri-o://b07e78295a4d2230639263bc614ce41918d2a2e30f14bafc5a14bae2081a653b" gracePeriod=10 Nov 25 10:48:48 crc kubenswrapper[4821]: I1125 10:48:48.391597 4821 generic.go:334] "Generic (PLEG): container finished" podID="dbdf62cf-071e-4e1c-9ef6-22bc39cfa795" containerID="4e9bd31c0491a3c78124341f952f334b25e755f3eac5df640fcc5a9961e6359f" exitCode=0 Nov 25 10:48:48 crc kubenswrapper[4821]: I1125 10:48:48.391667 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-bhnq8" event={"ID":"dbdf62cf-071e-4e1c-9ef6-22bc39cfa795","Type":"ContainerDied","Data":"4e9bd31c0491a3c78124341f952f334b25e755f3eac5df640fcc5a9961e6359f"} Nov 25 10:48:48 crc kubenswrapper[4821]: I1125 10:48:48.396918 4821 generic.go:334] "Generic (PLEG): container finished" podID="0333edd1-89f3-4ff0-9da3-544fac7620bb" containerID="b07e78295a4d2230639263bc614ce41918d2a2e30f14bafc5a14bae2081a653b" exitCode=0 Nov 25 10:48:48 crc kubenswrapper[4821]: I1125 10:48:48.397010 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-wctvh" event={"ID":"0333edd1-89f3-4ff0-9da3-544fac7620bb","Type":"ContainerDied","Data":"b07e78295a4d2230639263bc614ce41918d2a2e30f14bafc5a14bae2081a653b"} Nov 25 10:48:48 crc kubenswrapper[4821]: I1125 10:48:48.402110 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0ffb0da2-affb-4a95-87f3-eac442dc754d","Type":"ContainerStarted","Data":"c75978097f6c552bae29d393330fe912433c778656c55da9ca540c04fad99c86"} Nov 25 10:48:48 crc kubenswrapper[4821]: I1125 10:48:48.402142 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0ffb0da2-affb-4a95-87f3-eac442dc754d","Type":"ContainerStarted","Data":"11fb0113e2c478cd79cd1eb0ce56f63a713f38d436c7788e032b78f5dd2c4fa7"} Nov 25 10:48:48 crc kubenswrapper[4821]: I1125 10:48:48.402153 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0ffb0da2-affb-4a95-87f3-eac442dc754d","Type":"ContainerStarted","Data":"7b531045660469822dba718e00a3d42a3ff828684052ce407a0ef69f445d2e27"} Nov 25 10:48:48 crc kubenswrapper[4821]: I1125 10:48:48.402184 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"0ffb0da2-affb-4a95-87f3-eac442dc754d","Type":"ContainerStarted","Data":"6b9a9d9eabbe73f4d03218544c6f27fffae419ac7c16df98852c7e48c7097ff4"} Nov 25 10:48:48 crc kubenswrapper[4821]: I1125 10:48:48.436892 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-wctvh" Nov 25 10:48:48 crc kubenswrapper[4821]: I1125 10:48:48.574250 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-759mz\" (UniqueName: \"kubernetes.io/projected/0333edd1-89f3-4ff0-9da3-544fac7620bb-kube-api-access-759mz\") pod \"0333edd1-89f3-4ff0-9da3-544fac7620bb\" (UID: \"0333edd1-89f3-4ff0-9da3-544fac7620bb\") " Nov 25 10:48:48 crc kubenswrapper[4821]: I1125 10:48:48.574739 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0333edd1-89f3-4ff0-9da3-544fac7620bb-ovsdbserver-sb\") pod \"0333edd1-89f3-4ff0-9da3-544fac7620bb\" (UID: \"0333edd1-89f3-4ff0-9da3-544fac7620bb\") " Nov 25 10:48:48 crc kubenswrapper[4821]: I1125 10:48:48.574932 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0333edd1-89f3-4ff0-9da3-544fac7620bb-dns-svc\") pod \"0333edd1-89f3-4ff0-9da3-544fac7620bb\" (UID: \"0333edd1-89f3-4ff0-9da3-544fac7620bb\") " Nov 25 10:48:48 crc kubenswrapper[4821]: I1125 10:48:48.575041 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0333edd1-89f3-4ff0-9da3-544fac7620bb-config\") pod \"0333edd1-89f3-4ff0-9da3-544fac7620bb\" (UID: \"0333edd1-89f3-4ff0-9da3-544fac7620bb\") " Nov 25 10:48:48 crc kubenswrapper[4821]: I1125 10:48:48.575228 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0333edd1-89f3-4ff0-9da3-544fac7620bb-ovsdbserver-nb\") pod \"0333edd1-89f3-4ff0-9da3-544fac7620bb\" (UID: \"0333edd1-89f3-4ff0-9da3-544fac7620bb\") " Nov 25 10:48:48 crc kubenswrapper[4821]: I1125 10:48:48.579759 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0333edd1-89f3-4ff0-9da3-544fac7620bb-kube-api-access-759mz" (OuterVolumeSpecName: "kube-api-access-759mz") pod "0333edd1-89f3-4ff0-9da3-544fac7620bb" (UID: "0333edd1-89f3-4ff0-9da3-544fac7620bb"). InnerVolumeSpecName "kube-api-access-759mz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:48:48 crc kubenswrapper[4821]: I1125 10:48:48.614923 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0333edd1-89f3-4ff0-9da3-544fac7620bb-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0333edd1-89f3-4ff0-9da3-544fac7620bb" (UID: "0333edd1-89f3-4ff0-9da3-544fac7620bb"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:48:48 crc kubenswrapper[4821]: I1125 10:48:48.616292 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0333edd1-89f3-4ff0-9da3-544fac7620bb-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0333edd1-89f3-4ff0-9da3-544fac7620bb" (UID: "0333edd1-89f3-4ff0-9da3-544fac7620bb"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:48:48 crc kubenswrapper[4821]: I1125 10:48:48.616811 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0333edd1-89f3-4ff0-9da3-544fac7620bb-config" (OuterVolumeSpecName: "config") pod "0333edd1-89f3-4ff0-9da3-544fac7620bb" (UID: "0333edd1-89f3-4ff0-9da3-544fac7620bb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:48:48 crc kubenswrapper[4821]: I1125 10:48:48.618606 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0333edd1-89f3-4ff0-9da3-544fac7620bb-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0333edd1-89f3-4ff0-9da3-544fac7620bb" (UID: "0333edd1-89f3-4ff0-9da3-544fac7620bb"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:48:48 crc kubenswrapper[4821]: I1125 10:48:48.677109 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-759mz\" (UniqueName: \"kubernetes.io/projected/0333edd1-89f3-4ff0-9da3-544fac7620bb-kube-api-access-759mz\") on node \"crc\" DevicePath \"\"" Nov 25 10:48:48 crc kubenswrapper[4821]: I1125 10:48:48.677176 4821 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0333edd1-89f3-4ff0-9da3-544fac7620bb-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 10:48:48 crc kubenswrapper[4821]: I1125 10:48:48.677193 4821 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0333edd1-89f3-4ff0-9da3-544fac7620bb-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 10:48:48 crc kubenswrapper[4821]: I1125 10:48:48.677205 4821 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0333edd1-89f3-4ff0-9da3-544fac7620bb-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:48:48 crc kubenswrapper[4821]: I1125 10:48:48.677216 4821 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0333edd1-89f3-4ff0-9da3-544fac7620bb-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 10:48:49 crc kubenswrapper[4821]: I1125 10:48:49.411348 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-wctvh" event={"ID":"0333edd1-89f3-4ff0-9da3-544fac7620bb","Type":"ContainerDied","Data":"7bfb195a271d2a394c92d33cf0882fc4bde496bec9c182da2ebc130f98daf5fb"} Nov 25 10:48:49 crc kubenswrapper[4821]: I1125 10:48:49.411416 4821 scope.go:117] "RemoveContainer" containerID="b07e78295a4d2230639263bc614ce41918d2a2e30f14bafc5a14bae2081a653b" Nov 25 10:48:49 crc kubenswrapper[4821]: I1125 10:48:49.411424 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-wctvh" Nov 25 10:48:49 crc kubenswrapper[4821]: I1125 10:48:49.439729 4821 scope.go:117] "RemoveContainer" containerID="1b15650f1234d447a4a820c0d5f0a318e138ce1c4873e82e6ba0d4fb074ad408" Nov 25 10:48:49 crc kubenswrapper[4821]: I1125 10:48:49.465048 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=36.480761047 podStartE2EDuration="1m13.465027917s" podCreationTimestamp="2025-11-25 10:47:36 +0000 UTC" firstStartedPulling="2025-11-25 10:48:10.618381563 +0000 UTC m=+961.154701400" lastFinishedPulling="2025-11-25 10:48:47.602648423 +0000 UTC m=+998.138968270" observedRunningTime="2025-11-25 10:48:49.458287855 +0000 UTC m=+999.994607712" watchObservedRunningTime="2025-11-25 10:48:49.465027917 +0000 UTC m=+1000.001347764" Nov 25 10:48:49 crc kubenswrapper[4821]: I1125 10:48:49.502028 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-wctvh"] Nov 25 10:48:49 crc kubenswrapper[4821]: I1125 10:48:49.523697 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-wctvh"] Nov 25 10:48:49 crc kubenswrapper[4821]: I1125 10:48:49.732262 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-9qt9t"] Nov 25 10:48:49 crc kubenswrapper[4821]: E1125 10:48:49.732616 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0333edd1-89f3-4ff0-9da3-544fac7620bb" containerName="dnsmasq-dns" Nov 25 10:48:49 crc kubenswrapper[4821]: I1125 10:48:49.732634 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="0333edd1-89f3-4ff0-9da3-544fac7620bb" containerName="dnsmasq-dns" Nov 25 10:48:49 crc kubenswrapper[4821]: E1125 10:48:49.732647 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0333edd1-89f3-4ff0-9da3-544fac7620bb" containerName="init" Nov 25 10:48:49 crc kubenswrapper[4821]: I1125 10:48:49.732654 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="0333edd1-89f3-4ff0-9da3-544fac7620bb" containerName="init" Nov 25 10:48:49 crc kubenswrapper[4821]: I1125 10:48:49.732812 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="0333edd1-89f3-4ff0-9da3-544fac7620bb" containerName="dnsmasq-dns" Nov 25 10:48:49 crc kubenswrapper[4821]: I1125 10:48:49.734524 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f59b8f679-9qt9t" Nov 25 10:48:49 crc kubenswrapper[4821]: I1125 10:48:49.736001 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Nov 25 10:48:49 crc kubenswrapper[4821]: I1125 10:48:49.751978 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-9qt9t"] Nov 25 10:48:49 crc kubenswrapper[4821]: I1125 10:48:49.759869 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-bhnq8" Nov 25 10:48:49 crc kubenswrapper[4821]: I1125 10:48:49.795503 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2351168e-5815-4d63-b87d-e65011824547-ovsdbserver-sb\") pod \"dnsmasq-dns-5f59b8f679-9qt9t\" (UID: \"2351168e-5815-4d63-b87d-e65011824547\") " pod="openstack/dnsmasq-dns-5f59b8f679-9qt9t" Nov 25 10:48:49 crc kubenswrapper[4821]: I1125 10:48:49.795590 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2351168e-5815-4d63-b87d-e65011824547-dns-swift-storage-0\") pod \"dnsmasq-dns-5f59b8f679-9qt9t\" (UID: \"2351168e-5815-4d63-b87d-e65011824547\") " pod="openstack/dnsmasq-dns-5f59b8f679-9qt9t" Nov 25 10:48:49 crc kubenswrapper[4821]: I1125 10:48:49.795617 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2351168e-5815-4d63-b87d-e65011824547-config\") pod \"dnsmasq-dns-5f59b8f679-9qt9t\" (UID: \"2351168e-5815-4d63-b87d-e65011824547\") " pod="openstack/dnsmasq-dns-5f59b8f679-9qt9t" Nov 25 10:48:49 crc kubenswrapper[4821]: I1125 10:48:49.795725 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2351168e-5815-4d63-b87d-e65011824547-dns-svc\") pod \"dnsmasq-dns-5f59b8f679-9qt9t\" (UID: \"2351168e-5815-4d63-b87d-e65011824547\") " pod="openstack/dnsmasq-dns-5f59b8f679-9qt9t" Nov 25 10:48:49 crc kubenswrapper[4821]: I1125 10:48:49.795750 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2351168e-5815-4d63-b87d-e65011824547-ovsdbserver-nb\") pod \"dnsmasq-dns-5f59b8f679-9qt9t\" (UID: \"2351168e-5815-4d63-b87d-e65011824547\") " pod="openstack/dnsmasq-dns-5f59b8f679-9qt9t" Nov 25 10:48:49 crc kubenswrapper[4821]: I1125 10:48:49.795767 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6slbb\" (UniqueName: \"kubernetes.io/projected/2351168e-5815-4d63-b87d-e65011824547-kube-api-access-6slbb\") pod \"dnsmasq-dns-5f59b8f679-9qt9t\" (UID: \"2351168e-5815-4d63-b87d-e65011824547\") " pod="openstack/dnsmasq-dns-5f59b8f679-9qt9t" Nov 25 10:48:49 crc kubenswrapper[4821]: I1125 10:48:49.897150 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbdf62cf-071e-4e1c-9ef6-22bc39cfa795-combined-ca-bundle\") pod \"dbdf62cf-071e-4e1c-9ef6-22bc39cfa795\" (UID: \"dbdf62cf-071e-4e1c-9ef6-22bc39cfa795\") " Nov 25 10:48:49 crc kubenswrapper[4821]: I1125 10:48:49.897347 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbdf62cf-071e-4e1c-9ef6-22bc39cfa795-config-data\") pod \"dbdf62cf-071e-4e1c-9ef6-22bc39cfa795\" (UID: \"dbdf62cf-071e-4e1c-9ef6-22bc39cfa795\") " Nov 25 10:48:49 crc kubenswrapper[4821]: I1125 10:48:49.897406 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lltzf\" (UniqueName: \"kubernetes.io/projected/dbdf62cf-071e-4e1c-9ef6-22bc39cfa795-kube-api-access-lltzf\") pod \"dbdf62cf-071e-4e1c-9ef6-22bc39cfa795\" (UID: \"dbdf62cf-071e-4e1c-9ef6-22bc39cfa795\") " Nov 25 10:48:49 crc kubenswrapper[4821]: I1125 10:48:49.897651 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2351168e-5815-4d63-b87d-e65011824547-dns-svc\") pod \"dnsmasq-dns-5f59b8f679-9qt9t\" (UID: \"2351168e-5815-4d63-b87d-e65011824547\") " pod="openstack/dnsmasq-dns-5f59b8f679-9qt9t" Nov 25 10:48:49 crc kubenswrapper[4821]: I1125 10:48:49.897694 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2351168e-5815-4d63-b87d-e65011824547-ovsdbserver-nb\") pod \"dnsmasq-dns-5f59b8f679-9qt9t\" (UID: \"2351168e-5815-4d63-b87d-e65011824547\") " pod="openstack/dnsmasq-dns-5f59b8f679-9qt9t" Nov 25 10:48:49 crc kubenswrapper[4821]: I1125 10:48:49.897720 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6slbb\" (UniqueName: \"kubernetes.io/projected/2351168e-5815-4d63-b87d-e65011824547-kube-api-access-6slbb\") pod \"dnsmasq-dns-5f59b8f679-9qt9t\" (UID: \"2351168e-5815-4d63-b87d-e65011824547\") " pod="openstack/dnsmasq-dns-5f59b8f679-9qt9t" Nov 25 10:48:49 crc kubenswrapper[4821]: I1125 10:48:49.897764 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2351168e-5815-4d63-b87d-e65011824547-ovsdbserver-sb\") pod \"dnsmasq-dns-5f59b8f679-9qt9t\" (UID: \"2351168e-5815-4d63-b87d-e65011824547\") " pod="openstack/dnsmasq-dns-5f59b8f679-9qt9t" Nov 25 10:48:49 crc kubenswrapper[4821]: I1125 10:48:49.897810 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2351168e-5815-4d63-b87d-e65011824547-dns-swift-storage-0\") pod \"dnsmasq-dns-5f59b8f679-9qt9t\" (UID: \"2351168e-5815-4d63-b87d-e65011824547\") " pod="openstack/dnsmasq-dns-5f59b8f679-9qt9t" Nov 25 10:48:49 crc kubenswrapper[4821]: I1125 10:48:49.897835 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2351168e-5815-4d63-b87d-e65011824547-config\") pod \"dnsmasq-dns-5f59b8f679-9qt9t\" (UID: \"2351168e-5815-4d63-b87d-e65011824547\") " pod="openstack/dnsmasq-dns-5f59b8f679-9qt9t" Nov 25 10:48:49 crc kubenswrapper[4821]: I1125 10:48:49.898724 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2351168e-5815-4d63-b87d-e65011824547-ovsdbserver-nb\") pod \"dnsmasq-dns-5f59b8f679-9qt9t\" (UID: \"2351168e-5815-4d63-b87d-e65011824547\") " pod="openstack/dnsmasq-dns-5f59b8f679-9qt9t" Nov 25 10:48:49 crc kubenswrapper[4821]: I1125 10:48:49.898790 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2351168e-5815-4d63-b87d-e65011824547-dns-svc\") pod \"dnsmasq-dns-5f59b8f679-9qt9t\" (UID: \"2351168e-5815-4d63-b87d-e65011824547\") " pod="openstack/dnsmasq-dns-5f59b8f679-9qt9t" Nov 25 10:48:49 crc kubenswrapper[4821]: I1125 10:48:49.898932 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2351168e-5815-4d63-b87d-e65011824547-ovsdbserver-sb\") pod \"dnsmasq-dns-5f59b8f679-9qt9t\" (UID: \"2351168e-5815-4d63-b87d-e65011824547\") " pod="openstack/dnsmasq-dns-5f59b8f679-9qt9t" Nov 25 10:48:49 crc kubenswrapper[4821]: I1125 10:48:49.899145 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2351168e-5815-4d63-b87d-e65011824547-dns-swift-storage-0\") pod \"dnsmasq-dns-5f59b8f679-9qt9t\" (UID: \"2351168e-5815-4d63-b87d-e65011824547\") " pod="openstack/dnsmasq-dns-5f59b8f679-9qt9t" Nov 25 10:48:49 crc kubenswrapper[4821]: I1125 10:48:49.899877 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2351168e-5815-4d63-b87d-e65011824547-config\") pod \"dnsmasq-dns-5f59b8f679-9qt9t\" (UID: \"2351168e-5815-4d63-b87d-e65011824547\") " pod="openstack/dnsmasq-dns-5f59b8f679-9qt9t" Nov 25 10:48:49 crc kubenswrapper[4821]: I1125 10:48:49.903002 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dbdf62cf-071e-4e1c-9ef6-22bc39cfa795-kube-api-access-lltzf" (OuterVolumeSpecName: "kube-api-access-lltzf") pod "dbdf62cf-071e-4e1c-9ef6-22bc39cfa795" (UID: "dbdf62cf-071e-4e1c-9ef6-22bc39cfa795"). InnerVolumeSpecName "kube-api-access-lltzf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:48:49 crc kubenswrapper[4821]: I1125 10:48:49.921577 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6slbb\" (UniqueName: \"kubernetes.io/projected/2351168e-5815-4d63-b87d-e65011824547-kube-api-access-6slbb\") pod \"dnsmasq-dns-5f59b8f679-9qt9t\" (UID: \"2351168e-5815-4d63-b87d-e65011824547\") " pod="openstack/dnsmasq-dns-5f59b8f679-9qt9t" Nov 25 10:48:49 crc kubenswrapper[4821]: I1125 10:48:49.928521 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbdf62cf-071e-4e1c-9ef6-22bc39cfa795-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dbdf62cf-071e-4e1c-9ef6-22bc39cfa795" (UID: "dbdf62cf-071e-4e1c-9ef6-22bc39cfa795"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:48:49 crc kubenswrapper[4821]: I1125 10:48:49.947438 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbdf62cf-071e-4e1c-9ef6-22bc39cfa795-config-data" (OuterVolumeSpecName: "config-data") pod "dbdf62cf-071e-4e1c-9ef6-22bc39cfa795" (UID: "dbdf62cf-071e-4e1c-9ef6-22bc39cfa795"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:48:50 crc kubenswrapper[4821]: I1125 10:48:50.003956 4821 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbdf62cf-071e-4e1c-9ef6-22bc39cfa795-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:48:50 crc kubenswrapper[4821]: I1125 10:48:50.003984 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lltzf\" (UniqueName: \"kubernetes.io/projected/dbdf62cf-071e-4e1c-9ef6-22bc39cfa795-kube-api-access-lltzf\") on node \"crc\" DevicePath \"\"" Nov 25 10:48:50 crc kubenswrapper[4821]: I1125 10:48:50.003994 4821 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbdf62cf-071e-4e1c-9ef6-22bc39cfa795-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:48:50 crc kubenswrapper[4821]: I1125 10:48:50.098288 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f59b8f679-9qt9t" Nov 25 10:48:50 crc kubenswrapper[4821]: I1125 10:48:50.124320 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0333edd1-89f3-4ff0-9da3-544fac7620bb" path="/var/lib/kubelet/pods/0333edd1-89f3-4ff0-9da3-544fac7620bb/volumes" Nov 25 10:48:50 crc kubenswrapper[4821]: I1125 10:48:50.420783 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-bhnq8" event={"ID":"dbdf62cf-071e-4e1c-9ef6-22bc39cfa795","Type":"ContainerDied","Data":"884a1ca1174962b85fd4c57a68c054451ce76769a7966564e446be7051cfc247"} Nov 25 10:48:50 crc kubenswrapper[4821]: I1125 10:48:50.420824 4821 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="884a1ca1174962b85fd4c57a68c054451ce76769a7966564e446be7051cfc247" Nov 25 10:48:50 crc kubenswrapper[4821]: I1125 10:48:50.420899 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-bhnq8" Nov 25 10:48:50 crc kubenswrapper[4821]: I1125 10:48:50.512672 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-9qt9t"] Nov 25 10:48:50 crc kubenswrapper[4821]: W1125 10:48:50.519986 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2351168e_5815_4d63_b87d_e65011824547.slice/crio-5528b12573989025e5ba048441f3a409fd07232de5daa53360df0b789a08cb12 WatchSource:0}: Error finding container 5528b12573989025e5ba048441f3a409fd07232de5daa53360df0b789a08cb12: Status 404 returned error can't find the container with id 5528b12573989025e5ba048441f3a409fd07232de5daa53360df0b789a08cb12 Nov 25 10:48:50 crc kubenswrapper[4821]: I1125 10:48:50.643328 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-9qt9t"] Nov 25 10:48:50 crc kubenswrapper[4821]: I1125 10:48:50.701224 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-b5qsb"] Nov 25 10:48:50 crc kubenswrapper[4821]: E1125 10:48:50.701612 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbdf62cf-071e-4e1c-9ef6-22bc39cfa795" containerName="keystone-db-sync" Nov 25 10:48:50 crc kubenswrapper[4821]: I1125 10:48:50.701623 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbdf62cf-071e-4e1c-9ef6-22bc39cfa795" containerName="keystone-db-sync" Nov 25 10:48:50 crc kubenswrapper[4821]: I1125 10:48:50.701792 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="dbdf62cf-071e-4e1c-9ef6-22bc39cfa795" containerName="keystone-db-sync" Nov 25 10:48:50 crc kubenswrapper[4821]: I1125 10:48:50.702685 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bbf5cc879-b5qsb" Nov 25 10:48:50 crc kubenswrapper[4821]: I1125 10:48:50.720296 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-b5qsb"] Nov 25 10:48:50 crc kubenswrapper[4821]: I1125 10:48:50.732721 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-2mjz9"] Nov 25 10:48:50 crc kubenswrapper[4821]: I1125 10:48:50.755301 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-2mjz9"] Nov 25 10:48:50 crc kubenswrapper[4821]: I1125 10:48:50.755397 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2mjz9" Nov 25 10:48:50 crc kubenswrapper[4821]: I1125 10:48:50.760546 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 25 10:48:50 crc kubenswrapper[4821]: I1125 10:48:50.760841 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 25 10:48:50 crc kubenswrapper[4821]: I1125 10:48:50.760944 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-rzxc4" Nov 25 10:48:50 crc kubenswrapper[4821]: I1125 10:48:50.761447 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 25 10:48:50 crc kubenswrapper[4821]: I1125 10:48:50.761581 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 25 10:48:50 crc kubenswrapper[4821]: I1125 10:48:50.821813 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/099443e1-44f8-4fa0-a567-2a8675d1d740-ovsdbserver-sb\") pod \"dnsmasq-dns-bbf5cc879-b5qsb\" (UID: \"099443e1-44f8-4fa0-a567-2a8675d1d740\") " pod="openstack/dnsmasq-dns-bbf5cc879-b5qsb" Nov 25 10:48:50 crc kubenswrapper[4821]: I1125 10:48:50.821893 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a8462fa1-9270-4fa6-ae89-ea51e6246764-credential-keys\") pod \"keystone-bootstrap-2mjz9\" (UID: \"a8462fa1-9270-4fa6-ae89-ea51e6246764\") " pod="openstack/keystone-bootstrap-2mjz9" Nov 25 10:48:50 crc kubenswrapper[4821]: I1125 10:48:50.821953 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/099443e1-44f8-4fa0-a567-2a8675d1d740-config\") pod \"dnsmasq-dns-bbf5cc879-b5qsb\" (UID: \"099443e1-44f8-4fa0-a567-2a8675d1d740\") " pod="openstack/dnsmasq-dns-bbf5cc879-b5qsb" Nov 25 10:48:50 crc kubenswrapper[4821]: I1125 10:48:50.821979 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/099443e1-44f8-4fa0-a567-2a8675d1d740-dns-swift-storage-0\") pod \"dnsmasq-dns-bbf5cc879-b5qsb\" (UID: \"099443e1-44f8-4fa0-a567-2a8675d1d740\") " pod="openstack/dnsmasq-dns-bbf5cc879-b5qsb" Nov 25 10:48:50 crc kubenswrapper[4821]: I1125 10:48:50.822005 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8462fa1-9270-4fa6-ae89-ea51e6246764-combined-ca-bundle\") pod \"keystone-bootstrap-2mjz9\" (UID: \"a8462fa1-9270-4fa6-ae89-ea51e6246764\") " pod="openstack/keystone-bootstrap-2mjz9" Nov 25 10:48:50 crc kubenswrapper[4821]: I1125 10:48:50.822034 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8462fa1-9270-4fa6-ae89-ea51e6246764-scripts\") pod \"keystone-bootstrap-2mjz9\" (UID: \"a8462fa1-9270-4fa6-ae89-ea51e6246764\") " pod="openstack/keystone-bootstrap-2mjz9" Nov 25 10:48:50 crc kubenswrapper[4821]: I1125 10:48:50.822052 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8462fa1-9270-4fa6-ae89-ea51e6246764-config-data\") pod \"keystone-bootstrap-2mjz9\" (UID: \"a8462fa1-9270-4fa6-ae89-ea51e6246764\") " pod="openstack/keystone-bootstrap-2mjz9" Nov 25 10:48:50 crc kubenswrapper[4821]: I1125 10:48:50.822085 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a8462fa1-9270-4fa6-ae89-ea51e6246764-fernet-keys\") pod \"keystone-bootstrap-2mjz9\" (UID: \"a8462fa1-9270-4fa6-ae89-ea51e6246764\") " pod="openstack/keystone-bootstrap-2mjz9" Nov 25 10:48:50 crc kubenswrapper[4821]: I1125 10:48:50.822109 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/099443e1-44f8-4fa0-a567-2a8675d1d740-ovsdbserver-nb\") pod \"dnsmasq-dns-bbf5cc879-b5qsb\" (UID: \"099443e1-44f8-4fa0-a567-2a8675d1d740\") " pod="openstack/dnsmasq-dns-bbf5cc879-b5qsb" Nov 25 10:48:50 crc kubenswrapper[4821]: I1125 10:48:50.822192 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxmf5\" (UniqueName: \"kubernetes.io/projected/a8462fa1-9270-4fa6-ae89-ea51e6246764-kube-api-access-hxmf5\") pod \"keystone-bootstrap-2mjz9\" (UID: \"a8462fa1-9270-4fa6-ae89-ea51e6246764\") " pod="openstack/keystone-bootstrap-2mjz9" Nov 25 10:48:50 crc kubenswrapper[4821]: I1125 10:48:50.822227 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/099443e1-44f8-4fa0-a567-2a8675d1d740-dns-svc\") pod \"dnsmasq-dns-bbf5cc879-b5qsb\" (UID: \"099443e1-44f8-4fa0-a567-2a8675d1d740\") " pod="openstack/dnsmasq-dns-bbf5cc879-b5qsb" Nov 25 10:48:50 crc kubenswrapper[4821]: I1125 10:48:50.822254 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-smwmq\" (UniqueName: \"kubernetes.io/projected/099443e1-44f8-4fa0-a567-2a8675d1d740-kube-api-access-smwmq\") pod \"dnsmasq-dns-bbf5cc879-b5qsb\" (UID: \"099443e1-44f8-4fa0-a567-2a8675d1d740\") " pod="openstack/dnsmasq-dns-bbf5cc879-b5qsb" Nov 25 10:48:50 crc kubenswrapper[4821]: I1125 10:48:50.904251 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-8bp9n"] Nov 25 10:48:50 crc kubenswrapper[4821]: I1125 10:48:50.906118 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-8bp9n" Nov 25 10:48:50 crc kubenswrapper[4821]: I1125 10:48:50.909869 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 25 10:48:50 crc kubenswrapper[4821]: I1125 10:48:50.911984 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-zf45b" Nov 25 10:48:50 crc kubenswrapper[4821]: I1125 10:48:50.912389 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 25 10:48:50 crc kubenswrapper[4821]: I1125 10:48:50.923463 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-8bp9n"] Nov 25 10:48:50 crc kubenswrapper[4821]: I1125 10:48:50.927516 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8462fa1-9270-4fa6-ae89-ea51e6246764-scripts\") pod \"keystone-bootstrap-2mjz9\" (UID: \"a8462fa1-9270-4fa6-ae89-ea51e6246764\") " pod="openstack/keystone-bootstrap-2mjz9" Nov 25 10:48:50 crc kubenswrapper[4821]: I1125 10:48:50.927772 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8462fa1-9270-4fa6-ae89-ea51e6246764-config-data\") pod \"keystone-bootstrap-2mjz9\" (UID: \"a8462fa1-9270-4fa6-ae89-ea51e6246764\") " pod="openstack/keystone-bootstrap-2mjz9" Nov 25 10:48:50 crc kubenswrapper[4821]: I1125 10:48:50.927906 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a8462fa1-9270-4fa6-ae89-ea51e6246764-fernet-keys\") pod \"keystone-bootstrap-2mjz9\" (UID: \"a8462fa1-9270-4fa6-ae89-ea51e6246764\") " pod="openstack/keystone-bootstrap-2mjz9" Nov 25 10:48:50 crc kubenswrapper[4821]: I1125 10:48:50.928029 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/099443e1-44f8-4fa0-a567-2a8675d1d740-ovsdbserver-nb\") pod \"dnsmasq-dns-bbf5cc879-b5qsb\" (UID: \"099443e1-44f8-4fa0-a567-2a8675d1d740\") " pod="openstack/dnsmasq-dns-bbf5cc879-b5qsb" Nov 25 10:48:50 crc kubenswrapper[4821]: I1125 10:48:50.928308 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxmf5\" (UniqueName: \"kubernetes.io/projected/a8462fa1-9270-4fa6-ae89-ea51e6246764-kube-api-access-hxmf5\") pod \"keystone-bootstrap-2mjz9\" (UID: \"a8462fa1-9270-4fa6-ae89-ea51e6246764\") " pod="openstack/keystone-bootstrap-2mjz9" Nov 25 10:48:50 crc kubenswrapper[4821]: I1125 10:48:50.929075 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/099443e1-44f8-4fa0-a567-2a8675d1d740-dns-svc\") pod \"dnsmasq-dns-bbf5cc879-b5qsb\" (UID: \"099443e1-44f8-4fa0-a567-2a8675d1d740\") " pod="openstack/dnsmasq-dns-bbf5cc879-b5qsb" Nov 25 10:48:50 crc kubenswrapper[4821]: I1125 10:48:50.929251 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-smwmq\" (UniqueName: \"kubernetes.io/projected/099443e1-44f8-4fa0-a567-2a8675d1d740-kube-api-access-smwmq\") pod \"dnsmasq-dns-bbf5cc879-b5qsb\" (UID: \"099443e1-44f8-4fa0-a567-2a8675d1d740\") " pod="openstack/dnsmasq-dns-bbf5cc879-b5qsb" Nov 25 10:48:50 crc kubenswrapper[4821]: I1125 10:48:50.929390 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/099443e1-44f8-4fa0-a567-2a8675d1d740-ovsdbserver-sb\") pod \"dnsmasq-dns-bbf5cc879-b5qsb\" (UID: \"099443e1-44f8-4fa0-a567-2a8675d1d740\") " pod="openstack/dnsmasq-dns-bbf5cc879-b5qsb" Nov 25 10:48:50 crc kubenswrapper[4821]: I1125 10:48:50.929550 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a8462fa1-9270-4fa6-ae89-ea51e6246764-credential-keys\") pod \"keystone-bootstrap-2mjz9\" (UID: \"a8462fa1-9270-4fa6-ae89-ea51e6246764\") " pod="openstack/keystone-bootstrap-2mjz9" Nov 25 10:48:50 crc kubenswrapper[4821]: I1125 10:48:50.929740 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/099443e1-44f8-4fa0-a567-2a8675d1d740-config\") pod \"dnsmasq-dns-bbf5cc879-b5qsb\" (UID: \"099443e1-44f8-4fa0-a567-2a8675d1d740\") " pod="openstack/dnsmasq-dns-bbf5cc879-b5qsb" Nov 25 10:48:50 crc kubenswrapper[4821]: I1125 10:48:50.929849 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/099443e1-44f8-4fa0-a567-2a8675d1d740-dns-swift-storage-0\") pod \"dnsmasq-dns-bbf5cc879-b5qsb\" (UID: \"099443e1-44f8-4fa0-a567-2a8675d1d740\") " pod="openstack/dnsmasq-dns-bbf5cc879-b5qsb" Nov 25 10:48:50 crc kubenswrapper[4821]: I1125 10:48:50.929965 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8462fa1-9270-4fa6-ae89-ea51e6246764-combined-ca-bundle\") pod \"keystone-bootstrap-2mjz9\" (UID: \"a8462fa1-9270-4fa6-ae89-ea51e6246764\") " pod="openstack/keystone-bootstrap-2mjz9" Nov 25 10:48:50 crc kubenswrapper[4821]: I1125 10:48:50.931054 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/099443e1-44f8-4fa0-a567-2a8675d1d740-dns-svc\") pod \"dnsmasq-dns-bbf5cc879-b5qsb\" (UID: \"099443e1-44f8-4fa0-a567-2a8675d1d740\") " pod="openstack/dnsmasq-dns-bbf5cc879-b5qsb" Nov 25 10:48:50 crc kubenswrapper[4821]: I1125 10:48:50.934332 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/099443e1-44f8-4fa0-a567-2a8675d1d740-ovsdbserver-nb\") pod \"dnsmasq-dns-bbf5cc879-b5qsb\" (UID: \"099443e1-44f8-4fa0-a567-2a8675d1d740\") " pod="openstack/dnsmasq-dns-bbf5cc879-b5qsb" Nov 25 10:48:50 crc kubenswrapper[4821]: I1125 10:48:50.934781 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/099443e1-44f8-4fa0-a567-2a8675d1d740-config\") pod \"dnsmasq-dns-bbf5cc879-b5qsb\" (UID: \"099443e1-44f8-4fa0-a567-2a8675d1d740\") " pod="openstack/dnsmasq-dns-bbf5cc879-b5qsb" Nov 25 10:48:50 crc kubenswrapper[4821]: I1125 10:48:50.935444 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/099443e1-44f8-4fa0-a567-2a8675d1d740-dns-swift-storage-0\") pod \"dnsmasq-dns-bbf5cc879-b5qsb\" (UID: \"099443e1-44f8-4fa0-a567-2a8675d1d740\") " pod="openstack/dnsmasq-dns-bbf5cc879-b5qsb" Nov 25 10:48:50 crc kubenswrapper[4821]: I1125 10:48:50.935869 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/099443e1-44f8-4fa0-a567-2a8675d1d740-ovsdbserver-sb\") pod \"dnsmasq-dns-bbf5cc879-b5qsb\" (UID: \"099443e1-44f8-4fa0-a567-2a8675d1d740\") " pod="openstack/dnsmasq-dns-bbf5cc879-b5qsb" Nov 25 10:48:50 crc kubenswrapper[4821]: I1125 10:48:50.942734 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8462fa1-9270-4fa6-ae89-ea51e6246764-scripts\") pod \"keystone-bootstrap-2mjz9\" (UID: \"a8462fa1-9270-4fa6-ae89-ea51e6246764\") " pod="openstack/keystone-bootstrap-2mjz9" Nov 25 10:48:50 crc kubenswrapper[4821]: I1125 10:48:50.943566 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8462fa1-9270-4fa6-ae89-ea51e6246764-config-data\") pod \"keystone-bootstrap-2mjz9\" (UID: \"a8462fa1-9270-4fa6-ae89-ea51e6246764\") " pod="openstack/keystone-bootstrap-2mjz9" Nov 25 10:48:50 crc kubenswrapper[4821]: I1125 10:48:50.956619 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8462fa1-9270-4fa6-ae89-ea51e6246764-combined-ca-bundle\") pod \"keystone-bootstrap-2mjz9\" (UID: \"a8462fa1-9270-4fa6-ae89-ea51e6246764\") " pod="openstack/keystone-bootstrap-2mjz9" Nov 25 10:48:50 crc kubenswrapper[4821]: I1125 10:48:50.963714 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a8462fa1-9270-4fa6-ae89-ea51e6246764-credential-keys\") pod \"keystone-bootstrap-2mjz9\" (UID: \"a8462fa1-9270-4fa6-ae89-ea51e6246764\") " pod="openstack/keystone-bootstrap-2mjz9" Nov 25 10:48:50 crc kubenswrapper[4821]: I1125 10:48:50.976060 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a8462fa1-9270-4fa6-ae89-ea51e6246764-fernet-keys\") pod \"keystone-bootstrap-2mjz9\" (UID: \"a8462fa1-9270-4fa6-ae89-ea51e6246764\") " pod="openstack/keystone-bootstrap-2mjz9" Nov 25 10:48:50 crc kubenswrapper[4821]: I1125 10:48:50.981598 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-574c6969fc-sjxzg"] Nov 25 10:48:50 crc kubenswrapper[4821]: I1125 10:48:50.982286 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxmf5\" (UniqueName: \"kubernetes.io/projected/a8462fa1-9270-4fa6-ae89-ea51e6246764-kube-api-access-hxmf5\") pod \"keystone-bootstrap-2mjz9\" (UID: \"a8462fa1-9270-4fa6-ae89-ea51e6246764\") " pod="openstack/keystone-bootstrap-2mjz9" Nov 25 10:48:50 crc kubenswrapper[4821]: I1125 10:48:50.992312 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-574c6969fc-sjxzg" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.014933 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-574c6969fc-sjxzg"] Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.016933 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-smwmq\" (UniqueName: \"kubernetes.io/projected/099443e1-44f8-4fa0-a567-2a8675d1d740-kube-api-access-smwmq\") pod \"dnsmasq-dns-bbf5cc879-b5qsb\" (UID: \"099443e1-44f8-4fa0-a567-2a8675d1d740\") " pod="openstack/dnsmasq-dns-bbf5cc879-b5qsb" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.017203 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.017417 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.017658 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.018181 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-kpfkn" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.038415 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rfcbt\" (UniqueName: \"kubernetes.io/projected/eb6c9acf-d69c-4788-a85d-b72135de1ef3-kube-api-access-rfcbt\") pod \"cinder-db-sync-8bp9n\" (UID: \"eb6c9acf-d69c-4788-a85d-b72135de1ef3\") " pod="openstack/cinder-db-sync-8bp9n" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.038768 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/eb6c9acf-d69c-4788-a85d-b72135de1ef3-etc-machine-id\") pod \"cinder-db-sync-8bp9n\" (UID: \"eb6c9acf-d69c-4788-a85d-b72135de1ef3\") " pod="openstack/cinder-db-sync-8bp9n" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.038819 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/eb6c9acf-d69c-4788-a85d-b72135de1ef3-db-sync-config-data\") pod \"cinder-db-sync-8bp9n\" (UID: \"eb6c9acf-d69c-4788-a85d-b72135de1ef3\") " pod="openstack/cinder-db-sync-8bp9n" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.038880 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb6c9acf-d69c-4788-a85d-b72135de1ef3-config-data\") pod \"cinder-db-sync-8bp9n\" (UID: \"eb6c9acf-d69c-4788-a85d-b72135de1ef3\") " pod="openstack/cinder-db-sync-8bp9n" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.038910 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb6c9acf-d69c-4788-a85d-b72135de1ef3-combined-ca-bundle\") pod \"cinder-db-sync-8bp9n\" (UID: \"eb6c9acf-d69c-4788-a85d-b72135de1ef3\") " pod="openstack/cinder-db-sync-8bp9n" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.038952 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb6c9acf-d69c-4788-a85d-b72135de1ef3-scripts\") pod \"cinder-db-sync-8bp9n\" (UID: \"eb6c9acf-d69c-4788-a85d-b72135de1ef3\") " pod="openstack/cinder-db-sync-8bp9n" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.046229 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.048729 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.051008 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.053684 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.061746 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.067939 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-tq9mv"] Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.071309 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-tq9mv" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.081632 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.081874 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-fv5gx" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.081916 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.093086 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bbf5cc879-b5qsb" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.106957 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2mjz9" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.111646 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-b5qsb"] Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.166849 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9cb63eff-3e7e-436e-aea6-ce8254c3db77-log-httpd\") pod \"ceilometer-0\" (UID: \"9cb63eff-3e7e-436e-aea6-ce8254c3db77\") " pod="openstack/ceilometer-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.166912 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/eb6c9acf-d69c-4788-a85d-b72135de1ef3-db-sync-config-data\") pod \"cinder-db-sync-8bp9n\" (UID: \"eb6c9acf-d69c-4788-a85d-b72135de1ef3\") " pod="openstack/cinder-db-sync-8bp9n" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.166939 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/79a12e89-074b-48f2-bf44-b177fe2e84e2-scripts\") pod \"horizon-574c6969fc-sjxzg\" (UID: \"79a12e89-074b-48f2-bf44-b177fe2e84e2\") " pod="openstack/horizon-574c6969fc-sjxzg" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.166975 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/79a12e89-074b-48f2-bf44-b177fe2e84e2-config-data\") pod \"horizon-574c6969fc-sjxzg\" (UID: \"79a12e89-074b-48f2-bf44-b177fe2e84e2\") " pod="openstack/horizon-574c6969fc-sjxzg" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.166997 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cb63eff-3e7e-436e-aea6-ce8254c3db77-config-data\") pod \"ceilometer-0\" (UID: \"9cb63eff-3e7e-436e-aea6-ce8254c3db77\") " pod="openstack/ceilometer-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.167025 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb6c9acf-d69c-4788-a85d-b72135de1ef3-config-data\") pod \"cinder-db-sync-8bp9n\" (UID: \"eb6c9acf-d69c-4788-a85d-b72135de1ef3\") " pod="openstack/cinder-db-sync-8bp9n" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.167070 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb6c9acf-d69c-4788-a85d-b72135de1ef3-combined-ca-bundle\") pod \"cinder-db-sync-8bp9n\" (UID: \"eb6c9acf-d69c-4788-a85d-b72135de1ef3\") " pod="openstack/cinder-db-sync-8bp9n" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.167106 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/79a12e89-074b-48f2-bf44-b177fe2e84e2-logs\") pod \"horizon-574c6969fc-sjxzg\" (UID: \"79a12e89-074b-48f2-bf44-b177fe2e84e2\") " pod="openstack/horizon-574c6969fc-sjxzg" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.167135 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb6c9acf-d69c-4788-a85d-b72135de1ef3-scripts\") pod \"cinder-db-sync-8bp9n\" (UID: \"eb6c9acf-d69c-4788-a85d-b72135de1ef3\") " pod="openstack/cinder-db-sync-8bp9n" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.167192 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5cz7\" (UniqueName: \"kubernetes.io/projected/9cb63eff-3e7e-436e-aea6-ce8254c3db77-kube-api-access-f5cz7\") pod \"ceilometer-0\" (UID: \"9cb63eff-3e7e-436e-aea6-ce8254c3db77\") " pod="openstack/ceilometer-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.167214 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cb63eff-3e7e-436e-aea6-ce8254c3db77-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9cb63eff-3e7e-436e-aea6-ce8254c3db77\") " pod="openstack/ceilometer-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.167232 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/325653d2-264c-4c6c-851f-6e38fe3e61a5-config\") pod \"neutron-db-sync-tq9mv\" (UID: \"325653d2-264c-4c6c-851f-6e38fe3e61a5\") " pod="openstack/neutron-db-sync-tq9mv" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.167254 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9cb63eff-3e7e-436e-aea6-ce8254c3db77-scripts\") pod \"ceilometer-0\" (UID: \"9cb63eff-3e7e-436e-aea6-ce8254c3db77\") " pod="openstack/ceilometer-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.167291 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/79a12e89-074b-48f2-bf44-b177fe2e84e2-horizon-secret-key\") pod \"horizon-574c6969fc-sjxzg\" (UID: \"79a12e89-074b-48f2-bf44-b177fe2e84e2\") " pod="openstack/horizon-574c6969fc-sjxzg" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.167315 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvdv9\" (UniqueName: \"kubernetes.io/projected/325653d2-264c-4c6c-851f-6e38fe3e61a5-kube-api-access-hvdv9\") pod \"neutron-db-sync-tq9mv\" (UID: \"325653d2-264c-4c6c-851f-6e38fe3e61a5\") " pod="openstack/neutron-db-sync-tq9mv" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.167355 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/325653d2-264c-4c6c-851f-6e38fe3e61a5-combined-ca-bundle\") pod \"neutron-db-sync-tq9mv\" (UID: \"325653d2-264c-4c6c-851f-6e38fe3e61a5\") " pod="openstack/neutron-db-sync-tq9mv" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.167403 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rfcbt\" (UniqueName: \"kubernetes.io/projected/eb6c9acf-d69c-4788-a85d-b72135de1ef3-kube-api-access-rfcbt\") pod \"cinder-db-sync-8bp9n\" (UID: \"eb6c9acf-d69c-4788-a85d-b72135de1ef3\") " pod="openstack/cinder-db-sync-8bp9n" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.167430 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-frk9p\" (UniqueName: \"kubernetes.io/projected/79a12e89-074b-48f2-bf44-b177fe2e84e2-kube-api-access-frk9p\") pod \"horizon-574c6969fc-sjxzg\" (UID: \"79a12e89-074b-48f2-bf44-b177fe2e84e2\") " pod="openstack/horizon-574c6969fc-sjxzg" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.167454 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9cb63eff-3e7e-436e-aea6-ce8254c3db77-run-httpd\") pod \"ceilometer-0\" (UID: \"9cb63eff-3e7e-436e-aea6-ce8254c3db77\") " pod="openstack/ceilometer-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.167490 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/eb6c9acf-d69c-4788-a85d-b72135de1ef3-etc-machine-id\") pod \"cinder-db-sync-8bp9n\" (UID: \"eb6c9acf-d69c-4788-a85d-b72135de1ef3\") " pod="openstack/cinder-db-sync-8bp9n" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.167534 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9cb63eff-3e7e-436e-aea6-ce8254c3db77-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9cb63eff-3e7e-436e-aea6-ce8254c3db77\") " pod="openstack/ceilometer-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.178208 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-tq9mv"] Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.183842 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/eb6c9acf-d69c-4788-a85d-b72135de1ef3-etc-machine-id\") pod \"cinder-db-sync-8bp9n\" (UID: \"eb6c9acf-d69c-4788-a85d-b72135de1ef3\") " pod="openstack/cinder-db-sync-8bp9n" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.191708 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/eb6c9acf-d69c-4788-a85d-b72135de1ef3-db-sync-config-data\") pod \"cinder-db-sync-8bp9n\" (UID: \"eb6c9acf-d69c-4788-a85d-b72135de1ef3\") " pod="openstack/cinder-db-sync-8bp9n" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.200680 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb6c9acf-d69c-4788-a85d-b72135de1ef3-scripts\") pod \"cinder-db-sync-8bp9n\" (UID: \"eb6c9acf-d69c-4788-a85d-b72135de1ef3\") " pod="openstack/cinder-db-sync-8bp9n" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.246971 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb6c9acf-d69c-4788-a85d-b72135de1ef3-combined-ca-bundle\") pod \"cinder-db-sync-8bp9n\" (UID: \"eb6c9acf-d69c-4788-a85d-b72135de1ef3\") " pod="openstack/cinder-db-sync-8bp9n" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.267515 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb6c9acf-d69c-4788-a85d-b72135de1ef3-config-data\") pod \"cinder-db-sync-8bp9n\" (UID: \"eb6c9acf-d69c-4788-a85d-b72135de1ef3\") " pod="openstack/cinder-db-sync-8bp9n" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.269429 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5cz7\" (UniqueName: \"kubernetes.io/projected/9cb63eff-3e7e-436e-aea6-ce8254c3db77-kube-api-access-f5cz7\") pod \"ceilometer-0\" (UID: \"9cb63eff-3e7e-436e-aea6-ce8254c3db77\") " pod="openstack/ceilometer-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.269488 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cb63eff-3e7e-436e-aea6-ce8254c3db77-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9cb63eff-3e7e-436e-aea6-ce8254c3db77\") " pod="openstack/ceilometer-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.269509 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/325653d2-264c-4c6c-851f-6e38fe3e61a5-config\") pod \"neutron-db-sync-tq9mv\" (UID: \"325653d2-264c-4c6c-851f-6e38fe3e61a5\") " pod="openstack/neutron-db-sync-tq9mv" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.269533 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9cb63eff-3e7e-436e-aea6-ce8254c3db77-scripts\") pod \"ceilometer-0\" (UID: \"9cb63eff-3e7e-436e-aea6-ce8254c3db77\") " pod="openstack/ceilometer-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.270688 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/79a12e89-074b-48f2-bf44-b177fe2e84e2-horizon-secret-key\") pod \"horizon-574c6969fc-sjxzg\" (UID: \"79a12e89-074b-48f2-bf44-b177fe2e84e2\") " pod="openstack/horizon-574c6969fc-sjxzg" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.270780 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvdv9\" (UniqueName: \"kubernetes.io/projected/325653d2-264c-4c6c-851f-6e38fe3e61a5-kube-api-access-hvdv9\") pod \"neutron-db-sync-tq9mv\" (UID: \"325653d2-264c-4c6c-851f-6e38fe3e61a5\") " pod="openstack/neutron-db-sync-tq9mv" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.270890 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/325653d2-264c-4c6c-851f-6e38fe3e61a5-combined-ca-bundle\") pod \"neutron-db-sync-tq9mv\" (UID: \"325653d2-264c-4c6c-851f-6e38fe3e61a5\") " pod="openstack/neutron-db-sync-tq9mv" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.270996 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-frk9p\" (UniqueName: \"kubernetes.io/projected/79a12e89-074b-48f2-bf44-b177fe2e84e2-kube-api-access-frk9p\") pod \"horizon-574c6969fc-sjxzg\" (UID: \"79a12e89-074b-48f2-bf44-b177fe2e84e2\") " pod="openstack/horizon-574c6969fc-sjxzg" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.271049 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9cb63eff-3e7e-436e-aea6-ce8254c3db77-run-httpd\") pod \"ceilometer-0\" (UID: \"9cb63eff-3e7e-436e-aea6-ce8254c3db77\") " pod="openstack/ceilometer-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.271099 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rfcbt\" (UniqueName: \"kubernetes.io/projected/eb6c9acf-d69c-4788-a85d-b72135de1ef3-kube-api-access-rfcbt\") pod \"cinder-db-sync-8bp9n\" (UID: \"eb6c9acf-d69c-4788-a85d-b72135de1ef3\") " pod="openstack/cinder-db-sync-8bp9n" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.271215 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9cb63eff-3e7e-436e-aea6-ce8254c3db77-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9cb63eff-3e7e-436e-aea6-ce8254c3db77\") " pod="openstack/ceilometer-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.271261 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9cb63eff-3e7e-436e-aea6-ce8254c3db77-log-httpd\") pod \"ceilometer-0\" (UID: \"9cb63eff-3e7e-436e-aea6-ce8254c3db77\") " pod="openstack/ceilometer-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.271311 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/79a12e89-074b-48f2-bf44-b177fe2e84e2-scripts\") pod \"horizon-574c6969fc-sjxzg\" (UID: \"79a12e89-074b-48f2-bf44-b177fe2e84e2\") " pod="openstack/horizon-574c6969fc-sjxzg" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.271449 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/79a12e89-074b-48f2-bf44-b177fe2e84e2-config-data\") pod \"horizon-574c6969fc-sjxzg\" (UID: \"79a12e89-074b-48f2-bf44-b177fe2e84e2\") " pod="openstack/horizon-574c6969fc-sjxzg" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.271519 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cb63eff-3e7e-436e-aea6-ce8254c3db77-config-data\") pod \"ceilometer-0\" (UID: \"9cb63eff-3e7e-436e-aea6-ce8254c3db77\") " pod="openstack/ceilometer-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.271693 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/79a12e89-074b-48f2-bf44-b177fe2e84e2-logs\") pod \"horizon-574c6969fc-sjxzg\" (UID: \"79a12e89-074b-48f2-bf44-b177fe2e84e2\") " pod="openstack/horizon-574c6969fc-sjxzg" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.286995 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9cb63eff-3e7e-436e-aea6-ce8254c3db77-run-httpd\") pod \"ceilometer-0\" (UID: \"9cb63eff-3e7e-436e-aea6-ce8254c3db77\") " pod="openstack/ceilometer-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.291765 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9cb63eff-3e7e-436e-aea6-ce8254c3db77-log-httpd\") pod \"ceilometer-0\" (UID: \"9cb63eff-3e7e-436e-aea6-ce8254c3db77\") " pod="openstack/ceilometer-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.341078 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cb63eff-3e7e-436e-aea6-ce8254c3db77-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9cb63eff-3e7e-436e-aea6-ce8254c3db77\") " pod="openstack/ceilometer-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.341154 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/325653d2-264c-4c6c-851f-6e38fe3e61a5-config\") pod \"neutron-db-sync-tq9mv\" (UID: \"325653d2-264c-4c6c-851f-6e38fe3e61a5\") " pod="openstack/neutron-db-sync-tq9mv" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.341396 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/79a12e89-074b-48f2-bf44-b177fe2e84e2-logs\") pod \"horizon-574c6969fc-sjxzg\" (UID: \"79a12e89-074b-48f2-bf44-b177fe2e84e2\") " pod="openstack/horizon-574c6969fc-sjxzg" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.344053 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9cb63eff-3e7e-436e-aea6-ce8254c3db77-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9cb63eff-3e7e-436e-aea6-ce8254c3db77\") " pod="openstack/ceilometer-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.344669 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/325653d2-264c-4c6c-851f-6e38fe3e61a5-combined-ca-bundle\") pod \"neutron-db-sync-tq9mv\" (UID: \"325653d2-264c-4c6c-851f-6e38fe3e61a5\") " pod="openstack/neutron-db-sync-tq9mv" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.345189 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/79a12e89-074b-48f2-bf44-b177fe2e84e2-scripts\") pod \"horizon-574c6969fc-sjxzg\" (UID: \"79a12e89-074b-48f2-bf44-b177fe2e84e2\") " pod="openstack/horizon-574c6969fc-sjxzg" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.345833 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9cb63eff-3e7e-436e-aea6-ce8254c3db77-scripts\") pod \"ceilometer-0\" (UID: \"9cb63eff-3e7e-436e-aea6-ce8254c3db77\") " pod="openstack/ceilometer-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.346414 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/79a12e89-074b-48f2-bf44-b177fe2e84e2-horizon-secret-key\") pod \"horizon-574c6969fc-sjxzg\" (UID: \"79a12e89-074b-48f2-bf44-b177fe2e84e2\") " pod="openstack/horizon-574c6969fc-sjxzg" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.347463 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/79a12e89-074b-48f2-bf44-b177fe2e84e2-config-data\") pod \"horizon-574c6969fc-sjxzg\" (UID: \"79a12e89-074b-48f2-bf44-b177fe2e84e2\") " pod="openstack/horizon-574c6969fc-sjxzg" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.348883 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-zbmpp"] Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.351844 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-zbmpp" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.368788 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cb63eff-3e7e-436e-aea6-ce8254c3db77-config-data\") pod \"ceilometer-0\" (UID: \"9cb63eff-3e7e-436e-aea6-ce8254c3db77\") " pod="openstack/ceilometer-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.375640 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d7c7075f-28ad-46d5-b301-98f3868873db-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-zbmpp\" (UID: \"d7c7075f-28ad-46d5-b301-98f3868873db\") " pod="openstack/dnsmasq-dns-56df8fb6b7-zbmpp" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.375702 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d7c7075f-28ad-46d5-b301-98f3868873db-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-zbmpp\" (UID: \"d7c7075f-28ad-46d5-b301-98f3868873db\") " pod="openstack/dnsmasq-dns-56df8fb6b7-zbmpp" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.375781 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7c7075f-28ad-46d5-b301-98f3868873db-config\") pod \"dnsmasq-dns-56df8fb6b7-zbmpp\" (UID: \"d7c7075f-28ad-46d5-b301-98f3868873db\") " pod="openstack/dnsmasq-dns-56df8fb6b7-zbmpp" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.375806 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d7c7075f-28ad-46d5-b301-98f3868873db-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-zbmpp\" (UID: \"d7c7075f-28ad-46d5-b301-98f3868873db\") " pod="openstack/dnsmasq-dns-56df8fb6b7-zbmpp" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.376146 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vq2kd\" (UniqueName: \"kubernetes.io/projected/d7c7075f-28ad-46d5-b301-98f3868873db-kube-api-access-vq2kd\") pod \"dnsmasq-dns-56df8fb6b7-zbmpp\" (UID: \"d7c7075f-28ad-46d5-b301-98f3868873db\") " pod="openstack/dnsmasq-dns-56df8fb6b7-zbmpp" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.376311 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d7c7075f-28ad-46d5-b301-98f3868873db-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-zbmpp\" (UID: \"d7c7075f-28ad-46d5-b301-98f3868873db\") " pod="openstack/dnsmasq-dns-56df8fb6b7-zbmpp" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.388949 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvdv9\" (UniqueName: \"kubernetes.io/projected/325653d2-264c-4c6c-851f-6e38fe3e61a5-kube-api-access-hvdv9\") pod \"neutron-db-sync-tq9mv\" (UID: \"325653d2-264c-4c6c-851f-6e38fe3e61a5\") " pod="openstack/neutron-db-sync-tq9mv" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.389441 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5cz7\" (UniqueName: \"kubernetes.io/projected/9cb63eff-3e7e-436e-aea6-ce8254c3db77-kube-api-access-f5cz7\") pod \"ceilometer-0\" (UID: \"9cb63eff-3e7e-436e-aea6-ce8254c3db77\") " pod="openstack/ceilometer-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.395477 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-frk9p\" (UniqueName: \"kubernetes.io/projected/79a12e89-074b-48f2-bf44-b177fe2e84e2-kube-api-access-frk9p\") pod \"horizon-574c6969fc-sjxzg\" (UID: \"79a12e89-074b-48f2-bf44-b177fe2e84e2\") " pod="openstack/horizon-574c6969fc-sjxzg" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.405987 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7558555969-g2z2s"] Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.407451 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7558555969-g2z2s" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.408552 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.443402 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-xxdmc"] Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.444657 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-xxdmc" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.455774 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.456070 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.456225 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-ndsbm" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.462335 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-9qt9t" event={"ID":"2351168e-5815-4d63-b87d-e65011824547","Type":"ContainerStarted","Data":"7bb591d449410556307091427008c9c66dab10518332558ccb2c63301db45583"} Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.462378 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-9qt9t" event={"ID":"2351168e-5815-4d63-b87d-e65011824547","Type":"ContainerStarted","Data":"5528b12573989025e5ba048441f3a409fd07232de5daa53360df0b789a08cb12"} Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.462499 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5f59b8f679-9qt9t" podUID="2351168e-5815-4d63-b87d-e65011824547" containerName="init" containerID="cri-o://7bb591d449410556307091427008c9c66dab10518332558ccb2c63301db45583" gracePeriod=10 Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.472785 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7558555969-g2z2s"] Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.482733 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bee9a405-f60c-449e-beba-71cd802daa79-logs\") pod \"placement-db-sync-xxdmc\" (UID: \"bee9a405-f60c-449e-beba-71cd802daa79\") " pod="openstack/placement-db-sync-xxdmc" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.482905 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bee9a405-f60c-449e-beba-71cd802daa79-config-data\") pod \"placement-db-sync-xxdmc\" (UID: \"bee9a405-f60c-449e-beba-71cd802daa79\") " pod="openstack/placement-db-sync-xxdmc" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.482931 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bee9a405-f60c-449e-beba-71cd802daa79-combined-ca-bundle\") pod \"placement-db-sync-xxdmc\" (UID: \"bee9a405-f60c-449e-beba-71cd802daa79\") " pod="openstack/placement-db-sync-xxdmc" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.482969 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bee9a405-f60c-449e-beba-71cd802daa79-scripts\") pod \"placement-db-sync-xxdmc\" (UID: \"bee9a405-f60c-449e-beba-71cd802daa79\") " pod="openstack/placement-db-sync-xxdmc" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.483017 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2wrj\" (UniqueName: \"kubernetes.io/projected/77e0b5ce-8867-4d7e-b89e-170e67f3a548-kube-api-access-v2wrj\") pod \"horizon-7558555969-g2z2s\" (UID: \"77e0b5ce-8867-4d7e-b89e-170e67f3a548\") " pod="openstack/horizon-7558555969-g2z2s" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.483054 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d7c7075f-28ad-46d5-b301-98f3868873db-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-zbmpp\" (UID: \"d7c7075f-28ad-46d5-b301-98f3868873db\") " pod="openstack/dnsmasq-dns-56df8fb6b7-zbmpp" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.483075 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d7c7075f-28ad-46d5-b301-98f3868873db-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-zbmpp\" (UID: \"d7c7075f-28ad-46d5-b301-98f3868873db\") " pod="openstack/dnsmasq-dns-56df8fb6b7-zbmpp" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.483107 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7c7075f-28ad-46d5-b301-98f3868873db-config\") pod \"dnsmasq-dns-56df8fb6b7-zbmpp\" (UID: \"d7c7075f-28ad-46d5-b301-98f3868873db\") " pod="openstack/dnsmasq-dns-56df8fb6b7-zbmpp" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.483127 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d7c7075f-28ad-46d5-b301-98f3868873db-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-zbmpp\" (UID: \"d7c7075f-28ad-46d5-b301-98f3868873db\") " pod="openstack/dnsmasq-dns-56df8fb6b7-zbmpp" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.483154 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/77e0b5ce-8867-4d7e-b89e-170e67f3a548-horizon-secret-key\") pod \"horizon-7558555969-g2z2s\" (UID: \"77e0b5ce-8867-4d7e-b89e-170e67f3a548\") " pod="openstack/horizon-7558555969-g2z2s" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.483216 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/77e0b5ce-8867-4d7e-b89e-170e67f3a548-config-data\") pod \"horizon-7558555969-g2z2s\" (UID: \"77e0b5ce-8867-4d7e-b89e-170e67f3a548\") " pod="openstack/horizon-7558555969-g2z2s" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.483243 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/77e0b5ce-8867-4d7e-b89e-170e67f3a548-scripts\") pod \"horizon-7558555969-g2z2s\" (UID: \"77e0b5ce-8867-4d7e-b89e-170e67f3a548\") " pod="openstack/horizon-7558555969-g2z2s" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.483274 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/77e0b5ce-8867-4d7e-b89e-170e67f3a548-logs\") pod \"horizon-7558555969-g2z2s\" (UID: \"77e0b5ce-8867-4d7e-b89e-170e67f3a548\") " pod="openstack/horizon-7558555969-g2z2s" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.483305 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g4h77\" (UniqueName: \"kubernetes.io/projected/bee9a405-f60c-449e-beba-71cd802daa79-kube-api-access-g4h77\") pod \"placement-db-sync-xxdmc\" (UID: \"bee9a405-f60c-449e-beba-71cd802daa79\") " pod="openstack/placement-db-sync-xxdmc" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.483338 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vq2kd\" (UniqueName: \"kubernetes.io/projected/d7c7075f-28ad-46d5-b301-98f3868873db-kube-api-access-vq2kd\") pod \"dnsmasq-dns-56df8fb6b7-zbmpp\" (UID: \"d7c7075f-28ad-46d5-b301-98f3868873db\") " pod="openstack/dnsmasq-dns-56df8fb6b7-zbmpp" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.483361 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d7c7075f-28ad-46d5-b301-98f3868873db-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-zbmpp\" (UID: \"d7c7075f-28ad-46d5-b301-98f3868873db\") " pod="openstack/dnsmasq-dns-56df8fb6b7-zbmpp" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.484302 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d7c7075f-28ad-46d5-b301-98f3868873db-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-zbmpp\" (UID: \"d7c7075f-28ad-46d5-b301-98f3868873db\") " pod="openstack/dnsmasq-dns-56df8fb6b7-zbmpp" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.484996 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d7c7075f-28ad-46d5-b301-98f3868873db-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-zbmpp\" (UID: \"d7c7075f-28ad-46d5-b301-98f3868873db\") " pod="openstack/dnsmasq-dns-56df8fb6b7-zbmpp" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.486919 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7c7075f-28ad-46d5-b301-98f3868873db-config\") pod \"dnsmasq-dns-56df8fb6b7-zbmpp\" (UID: \"d7c7075f-28ad-46d5-b301-98f3868873db\") " pod="openstack/dnsmasq-dns-56df8fb6b7-zbmpp" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.489094 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-tq9mv" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.491492 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d7c7075f-28ad-46d5-b301-98f3868873db-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-zbmpp\" (UID: \"d7c7075f-28ad-46d5-b301-98f3868873db\") " pod="openstack/dnsmasq-dns-56df8fb6b7-zbmpp" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.496856 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d7c7075f-28ad-46d5-b301-98f3868873db-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-zbmpp\" (UID: \"d7c7075f-28ad-46d5-b301-98f3868873db\") " pod="openstack/dnsmasq-dns-56df8fb6b7-zbmpp" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.528651 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.530096 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.536894 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vq2kd\" (UniqueName: \"kubernetes.io/projected/d7c7075f-28ad-46d5-b301-98f3868873db-kube-api-access-vq2kd\") pod \"dnsmasq-dns-56df8fb6b7-zbmpp\" (UID: \"d7c7075f-28ad-46d5-b301-98f3868873db\") " pod="openstack/dnsmasq-dns-56df8fb6b7-zbmpp" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.537322 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.537548 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-lqdn5" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.537776 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.537919 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.542262 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-zbmpp"] Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.542397 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-8bp9n" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.579755 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-xxdmc"] Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.586192 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2d41a59-94d9-4440-bfa6-f88e49b97aff-config-data\") pod \"glance-default-external-api-0\" (UID: \"f2d41a59-94d9-4440-bfa6-f88e49b97aff\") " pod="openstack/glance-default-external-api-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.586242 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2wrj\" (UniqueName: \"kubernetes.io/projected/77e0b5ce-8867-4d7e-b89e-170e67f3a548-kube-api-access-v2wrj\") pod \"horizon-7558555969-g2z2s\" (UID: \"77e0b5ce-8867-4d7e-b89e-170e67f3a548\") " pod="openstack/horizon-7558555969-g2z2s" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.586281 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f2d41a59-94d9-4440-bfa6-f88e49b97aff-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f2d41a59-94d9-4440-bfa6-f88e49b97aff\") " pod="openstack/glance-default-external-api-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.586312 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2d41a59-94d9-4440-bfa6-f88e49b97aff-scripts\") pod \"glance-default-external-api-0\" (UID: \"f2d41a59-94d9-4440-bfa6-f88e49b97aff\") " pod="openstack/glance-default-external-api-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.586340 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/77e0b5ce-8867-4d7e-b89e-170e67f3a548-horizon-secret-key\") pod \"horizon-7558555969-g2z2s\" (UID: \"77e0b5ce-8867-4d7e-b89e-170e67f3a548\") " pod="openstack/horizon-7558555969-g2z2s" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.586358 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"f2d41a59-94d9-4440-bfa6-f88e49b97aff\") " pod="openstack/glance-default-external-api-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.586377 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f2d41a59-94d9-4440-bfa6-f88e49b97aff-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"f2d41a59-94d9-4440-bfa6-f88e49b97aff\") " pod="openstack/glance-default-external-api-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.586403 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f2d41a59-94d9-4440-bfa6-f88e49b97aff-logs\") pod \"glance-default-external-api-0\" (UID: \"f2d41a59-94d9-4440-bfa6-f88e49b97aff\") " pod="openstack/glance-default-external-api-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.586421 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/77e0b5ce-8867-4d7e-b89e-170e67f3a548-config-data\") pod \"horizon-7558555969-g2z2s\" (UID: \"77e0b5ce-8867-4d7e-b89e-170e67f3a548\") " pod="openstack/horizon-7558555969-g2z2s" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.586441 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/77e0b5ce-8867-4d7e-b89e-170e67f3a548-scripts\") pod \"horizon-7558555969-g2z2s\" (UID: \"77e0b5ce-8867-4d7e-b89e-170e67f3a548\") " pod="openstack/horizon-7558555969-g2z2s" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.586476 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/77e0b5ce-8867-4d7e-b89e-170e67f3a548-logs\") pod \"horizon-7558555969-g2z2s\" (UID: \"77e0b5ce-8867-4d7e-b89e-170e67f3a548\") " pod="openstack/horizon-7558555969-g2z2s" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.586500 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g4h77\" (UniqueName: \"kubernetes.io/projected/bee9a405-f60c-449e-beba-71cd802daa79-kube-api-access-g4h77\") pod \"placement-db-sync-xxdmc\" (UID: \"bee9a405-f60c-449e-beba-71cd802daa79\") " pod="openstack/placement-db-sync-xxdmc" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.586533 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bee9a405-f60c-449e-beba-71cd802daa79-logs\") pod \"placement-db-sync-xxdmc\" (UID: \"bee9a405-f60c-449e-beba-71cd802daa79\") " pod="openstack/placement-db-sync-xxdmc" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.586551 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bee9a405-f60c-449e-beba-71cd802daa79-config-data\") pod \"placement-db-sync-xxdmc\" (UID: \"bee9a405-f60c-449e-beba-71cd802daa79\") " pod="openstack/placement-db-sync-xxdmc" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.586572 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bee9a405-f60c-449e-beba-71cd802daa79-combined-ca-bundle\") pod \"placement-db-sync-xxdmc\" (UID: \"bee9a405-f60c-449e-beba-71cd802daa79\") " pod="openstack/placement-db-sync-xxdmc" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.586603 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2d41a59-94d9-4440-bfa6-f88e49b97aff-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f2d41a59-94d9-4440-bfa6-f88e49b97aff\") " pod="openstack/glance-default-external-api-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.586627 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bee9a405-f60c-449e-beba-71cd802daa79-scripts\") pod \"placement-db-sync-xxdmc\" (UID: \"bee9a405-f60c-449e-beba-71cd802daa79\") " pod="openstack/placement-db-sync-xxdmc" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.586647 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tvsg4\" (UniqueName: \"kubernetes.io/projected/f2d41a59-94d9-4440-bfa6-f88e49b97aff-kube-api-access-tvsg4\") pod \"glance-default-external-api-0\" (UID: \"f2d41a59-94d9-4440-bfa6-f88e49b97aff\") " pod="openstack/glance-default-external-api-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.588550 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/77e0b5ce-8867-4d7e-b89e-170e67f3a548-logs\") pod \"horizon-7558555969-g2z2s\" (UID: \"77e0b5ce-8867-4d7e-b89e-170e67f3a548\") " pod="openstack/horizon-7558555969-g2z2s" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.589391 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/77e0b5ce-8867-4d7e-b89e-170e67f3a548-scripts\") pod \"horizon-7558555969-g2z2s\" (UID: \"77e0b5ce-8867-4d7e-b89e-170e67f3a548\") " pod="openstack/horizon-7558555969-g2z2s" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.589955 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bee9a405-f60c-449e-beba-71cd802daa79-logs\") pod \"placement-db-sync-xxdmc\" (UID: \"bee9a405-f60c-449e-beba-71cd802daa79\") " pod="openstack/placement-db-sync-xxdmc" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.592504 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/77e0b5ce-8867-4d7e-b89e-170e67f3a548-horizon-secret-key\") pod \"horizon-7558555969-g2z2s\" (UID: \"77e0b5ce-8867-4d7e-b89e-170e67f3a548\") " pod="openstack/horizon-7558555969-g2z2s" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.593346 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/77e0b5ce-8867-4d7e-b89e-170e67f3a548-config-data\") pod \"horizon-7558555969-g2z2s\" (UID: \"77e0b5ce-8867-4d7e-b89e-170e67f3a548\") " pod="openstack/horizon-7558555969-g2z2s" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.597663 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.600393 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bee9a405-f60c-449e-beba-71cd802daa79-combined-ca-bundle\") pod \"placement-db-sync-xxdmc\" (UID: \"bee9a405-f60c-449e-beba-71cd802daa79\") " pod="openstack/placement-db-sync-xxdmc" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.602281 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bee9a405-f60c-449e-beba-71cd802daa79-config-data\") pod \"placement-db-sync-xxdmc\" (UID: \"bee9a405-f60c-449e-beba-71cd802daa79\") " pod="openstack/placement-db-sync-xxdmc" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.602555 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bee9a405-f60c-449e-beba-71cd802daa79-scripts\") pod \"placement-db-sync-xxdmc\" (UID: \"bee9a405-f60c-449e-beba-71cd802daa79\") " pod="openstack/placement-db-sync-xxdmc" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.604549 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-7znwh"] Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.605898 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-7znwh" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.610030 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-7znwh"] Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.610367 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-r9x5r" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.616575 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g4h77\" (UniqueName: \"kubernetes.io/projected/bee9a405-f60c-449e-beba-71cd802daa79-kube-api-access-g4h77\") pod \"placement-db-sync-xxdmc\" (UID: \"bee9a405-f60c-449e-beba-71cd802daa79\") " pod="openstack/placement-db-sync-xxdmc" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.619646 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.621850 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2wrj\" (UniqueName: \"kubernetes.io/projected/77e0b5ce-8867-4d7e-b89e-170e67f3a548-kube-api-access-v2wrj\") pod \"horizon-7558555969-g2z2s\" (UID: \"77e0b5ce-8867-4d7e-b89e-170e67f3a548\") " pod="openstack/horizon-7558555969-g2z2s" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.652939 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.655616 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.659625 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.659981 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.660536 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.688490 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bf0e0161-6d4e-4149-a1f2-7b96ba1f736b-logs\") pod \"glance-default-internal-api-0\" (UID: \"bf0e0161-6d4e-4149-a1f2-7b96ba1f736b\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.688699 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf0e0161-6d4e-4149-a1f2-7b96ba1f736b-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"bf0e0161-6d4e-4149-a1f2-7b96ba1f736b\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.688786 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dfhml\" (UniqueName: \"kubernetes.io/projected/bf0e0161-6d4e-4149-a1f2-7b96ba1f736b-kube-api-access-dfhml\") pod \"glance-default-internal-api-0\" (UID: \"bf0e0161-6d4e-4149-a1f2-7b96ba1f736b\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.688862 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"bf0e0161-6d4e-4149-a1f2-7b96ba1f736b\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.688953 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2d41a59-94d9-4440-bfa6-f88e49b97aff-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f2d41a59-94d9-4440-bfa6-f88e49b97aff\") " pod="openstack/glance-default-external-api-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.689035 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tvsg4\" (UniqueName: \"kubernetes.io/projected/f2d41a59-94d9-4440-bfa6-f88e49b97aff-kube-api-access-tvsg4\") pod \"glance-default-external-api-0\" (UID: \"f2d41a59-94d9-4440-bfa6-f88e49b97aff\") " pod="openstack/glance-default-external-api-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.689108 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2d41a59-94d9-4440-bfa6-f88e49b97aff-config-data\") pod \"glance-default-external-api-0\" (UID: \"f2d41a59-94d9-4440-bfa6-f88e49b97aff\") " pod="openstack/glance-default-external-api-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.689266 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf0e0161-6d4e-4149-a1f2-7b96ba1f736b-config-data\") pod \"glance-default-internal-api-0\" (UID: \"bf0e0161-6d4e-4149-a1f2-7b96ba1f736b\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.689370 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f2d41a59-94d9-4440-bfa6-f88e49b97aff-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f2d41a59-94d9-4440-bfa6-f88e49b97aff\") " pod="openstack/glance-default-external-api-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.689448 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2d41a59-94d9-4440-bfa6-f88e49b97aff-scripts\") pod \"glance-default-external-api-0\" (UID: \"f2d41a59-94d9-4440-bfa6-f88e49b97aff\") " pod="openstack/glance-default-external-api-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.689519 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf0e0161-6d4e-4149-a1f2-7b96ba1f736b-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"bf0e0161-6d4e-4149-a1f2-7b96ba1f736b\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.689597 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"f2d41a59-94d9-4440-bfa6-f88e49b97aff\") " pod="openstack/glance-default-external-api-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.689675 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf0e0161-6d4e-4149-a1f2-7b96ba1f736b-scripts\") pod \"glance-default-internal-api-0\" (UID: \"bf0e0161-6d4e-4149-a1f2-7b96ba1f736b\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.689782 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f2d41a59-94d9-4440-bfa6-f88e49b97aff-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"f2d41a59-94d9-4440-bfa6-f88e49b97aff\") " pod="openstack/glance-default-external-api-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.689863 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bf0e0161-6d4e-4149-a1f2-7b96ba1f736b-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"bf0e0161-6d4e-4149-a1f2-7b96ba1f736b\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.689937 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f2d41a59-94d9-4440-bfa6-f88e49b97aff-logs\") pod \"glance-default-external-api-0\" (UID: \"f2d41a59-94d9-4440-bfa6-f88e49b97aff\") " pod="openstack/glance-default-external-api-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.690488 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f2d41a59-94d9-4440-bfa6-f88e49b97aff-logs\") pod \"glance-default-external-api-0\" (UID: \"f2d41a59-94d9-4440-bfa6-f88e49b97aff\") " pod="openstack/glance-default-external-api-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.690880 4821 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"f2d41a59-94d9-4440-bfa6-f88e49b97aff\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-external-api-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.688908 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-574c6969fc-sjxzg" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.692107 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f2d41a59-94d9-4440-bfa6-f88e49b97aff-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f2d41a59-94d9-4440-bfa6-f88e49b97aff\") " pod="openstack/glance-default-external-api-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.692698 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2d41a59-94d9-4440-bfa6-f88e49b97aff-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f2d41a59-94d9-4440-bfa6-f88e49b97aff\") " pod="openstack/glance-default-external-api-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.697516 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2d41a59-94d9-4440-bfa6-f88e49b97aff-config-data\") pod \"glance-default-external-api-0\" (UID: \"f2d41a59-94d9-4440-bfa6-f88e49b97aff\") " pod="openstack/glance-default-external-api-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.698055 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f2d41a59-94d9-4440-bfa6-f88e49b97aff-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"f2d41a59-94d9-4440-bfa6-f88e49b97aff\") " pod="openstack/glance-default-external-api-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.706780 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-zbmpp" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.715068 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2d41a59-94d9-4440-bfa6-f88e49b97aff-scripts\") pod \"glance-default-external-api-0\" (UID: \"f2d41a59-94d9-4440-bfa6-f88e49b97aff\") " pod="openstack/glance-default-external-api-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.717835 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tvsg4\" (UniqueName: \"kubernetes.io/projected/f2d41a59-94d9-4440-bfa6-f88e49b97aff-kube-api-access-tvsg4\") pod \"glance-default-external-api-0\" (UID: \"f2d41a59-94d9-4440-bfa6-f88e49b97aff\") " pod="openstack/glance-default-external-api-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.732350 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7558555969-g2z2s" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.733493 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"f2d41a59-94d9-4440-bfa6-f88e49b97aff\") " pod="openstack/glance-default-external-api-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.785724 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-xxdmc" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.799701 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"bf0e0161-6d4e-4149-a1f2-7b96ba1f736b\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.799791 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf0e0161-6d4e-4149-a1f2-7b96ba1f736b-config-data\") pod \"glance-default-internal-api-0\" (UID: \"bf0e0161-6d4e-4149-a1f2-7b96ba1f736b\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.799840 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf0e0161-6d4e-4149-a1f2-7b96ba1f736b-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"bf0e0161-6d4e-4149-a1f2-7b96ba1f736b\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.799895 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf0e0161-6d4e-4149-a1f2-7b96ba1f736b-scripts\") pod \"glance-default-internal-api-0\" (UID: \"bf0e0161-6d4e-4149-a1f2-7b96ba1f736b\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.799921 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bf0e0161-6d4e-4149-a1f2-7b96ba1f736b-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"bf0e0161-6d4e-4149-a1f2-7b96ba1f736b\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.799956 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2eb6852b-84a1-4d4a-8402-77a7e2252ab2-combined-ca-bundle\") pod \"barbican-db-sync-7znwh\" (UID: \"2eb6852b-84a1-4d4a-8402-77a7e2252ab2\") " pod="openstack/barbican-db-sync-7znwh" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.799989 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l58nm\" (UniqueName: \"kubernetes.io/projected/2eb6852b-84a1-4d4a-8402-77a7e2252ab2-kube-api-access-l58nm\") pod \"barbican-db-sync-7znwh\" (UID: \"2eb6852b-84a1-4d4a-8402-77a7e2252ab2\") " pod="openstack/barbican-db-sync-7znwh" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.800024 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2eb6852b-84a1-4d4a-8402-77a7e2252ab2-db-sync-config-data\") pod \"barbican-db-sync-7znwh\" (UID: \"2eb6852b-84a1-4d4a-8402-77a7e2252ab2\") " pod="openstack/barbican-db-sync-7znwh" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.800061 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bf0e0161-6d4e-4149-a1f2-7b96ba1f736b-logs\") pod \"glance-default-internal-api-0\" (UID: \"bf0e0161-6d4e-4149-a1f2-7b96ba1f736b\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.800086 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf0e0161-6d4e-4149-a1f2-7b96ba1f736b-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"bf0e0161-6d4e-4149-a1f2-7b96ba1f736b\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.800102 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dfhml\" (UniqueName: \"kubernetes.io/projected/bf0e0161-6d4e-4149-a1f2-7b96ba1f736b-kube-api-access-dfhml\") pod \"glance-default-internal-api-0\" (UID: \"bf0e0161-6d4e-4149-a1f2-7b96ba1f736b\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.800470 4821 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"bf0e0161-6d4e-4149-a1f2-7b96ba1f736b\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-internal-api-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.803293 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bf0e0161-6d4e-4149-a1f2-7b96ba1f736b-logs\") pod \"glance-default-internal-api-0\" (UID: \"bf0e0161-6d4e-4149-a1f2-7b96ba1f736b\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.803455 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bf0e0161-6d4e-4149-a1f2-7b96ba1f736b-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"bf0e0161-6d4e-4149-a1f2-7b96ba1f736b\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.805463 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf0e0161-6d4e-4149-a1f2-7b96ba1f736b-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"bf0e0161-6d4e-4149-a1f2-7b96ba1f736b\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.806424 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf0e0161-6d4e-4149-a1f2-7b96ba1f736b-scripts\") pod \"glance-default-internal-api-0\" (UID: \"bf0e0161-6d4e-4149-a1f2-7b96ba1f736b\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.806636 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf0e0161-6d4e-4149-a1f2-7b96ba1f736b-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"bf0e0161-6d4e-4149-a1f2-7b96ba1f736b\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.807051 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf0e0161-6d4e-4149-a1f2-7b96ba1f736b-config-data\") pod \"glance-default-internal-api-0\" (UID: \"bf0e0161-6d4e-4149-a1f2-7b96ba1f736b\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.817886 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dfhml\" (UniqueName: \"kubernetes.io/projected/bf0e0161-6d4e-4149-a1f2-7b96ba1f736b-kube-api-access-dfhml\") pod \"glance-default-internal-api-0\" (UID: \"bf0e0161-6d4e-4149-a1f2-7b96ba1f736b\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.841333 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"bf0e0161-6d4e-4149-a1f2-7b96ba1f736b\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.901453 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2eb6852b-84a1-4d4a-8402-77a7e2252ab2-combined-ca-bundle\") pod \"barbican-db-sync-7znwh\" (UID: \"2eb6852b-84a1-4d4a-8402-77a7e2252ab2\") " pod="openstack/barbican-db-sync-7znwh" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.901523 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l58nm\" (UniqueName: \"kubernetes.io/projected/2eb6852b-84a1-4d4a-8402-77a7e2252ab2-kube-api-access-l58nm\") pod \"barbican-db-sync-7znwh\" (UID: \"2eb6852b-84a1-4d4a-8402-77a7e2252ab2\") " pod="openstack/barbican-db-sync-7znwh" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.901598 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2eb6852b-84a1-4d4a-8402-77a7e2252ab2-db-sync-config-data\") pod \"barbican-db-sync-7znwh\" (UID: \"2eb6852b-84a1-4d4a-8402-77a7e2252ab2\") " pod="openstack/barbican-db-sync-7znwh" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.903293 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.907700 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2eb6852b-84a1-4d4a-8402-77a7e2252ab2-db-sync-config-data\") pod \"barbican-db-sync-7znwh\" (UID: \"2eb6852b-84a1-4d4a-8402-77a7e2252ab2\") " pod="openstack/barbican-db-sync-7znwh" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.908527 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2eb6852b-84a1-4d4a-8402-77a7e2252ab2-combined-ca-bundle\") pod \"barbican-db-sync-7znwh\" (UID: \"2eb6852b-84a1-4d4a-8402-77a7e2252ab2\") " pod="openstack/barbican-db-sync-7znwh" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.918470 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l58nm\" (UniqueName: \"kubernetes.io/projected/2eb6852b-84a1-4d4a-8402-77a7e2252ab2-kube-api-access-l58nm\") pod \"barbican-db-sync-7znwh\" (UID: \"2eb6852b-84a1-4d4a-8402-77a7e2252ab2\") " pod="openstack/barbican-db-sync-7znwh" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.967333 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-7znwh" Nov 25 10:48:51 crc kubenswrapper[4821]: I1125 10:48:51.991793 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 25 10:48:52 crc kubenswrapper[4821]: I1125 10:48:52.174736 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-2mjz9"] Nov 25 10:48:52 crc kubenswrapper[4821]: I1125 10:48:52.263543 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-b5qsb"] Nov 25 10:48:52 crc kubenswrapper[4821]: W1125 10:48:52.277469 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod099443e1_44f8_4fa0_a567_2a8675d1d740.slice/crio-38f55b56330ea0a7d8a8d252271f17ac9666a5f663f11b8064b51fe6c00e522d WatchSource:0}: Error finding container 38f55b56330ea0a7d8a8d252271f17ac9666a5f663f11b8064b51fe6c00e522d: Status 404 returned error can't find the container with id 38f55b56330ea0a7d8a8d252271f17ac9666a5f663f11b8064b51fe6c00e522d Nov 25 10:48:52 crc kubenswrapper[4821]: I1125 10:48:52.370074 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f59b8f679-9qt9t" Nov 25 10:48:52 crc kubenswrapper[4821]: I1125 10:48:52.474246 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2mjz9" event={"ID":"a8462fa1-9270-4fa6-ae89-ea51e6246764","Type":"ContainerStarted","Data":"ac7e7a6e7fa3583d81e45e8d909f8b18b87b37f18db9fdb40be515f4278bf153"} Nov 25 10:48:52 crc kubenswrapper[4821]: I1125 10:48:52.474589 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2mjz9" event={"ID":"a8462fa1-9270-4fa6-ae89-ea51e6246764","Type":"ContainerStarted","Data":"e82d35ac02d039edd98eca2a377749fec8b2db0dd99f4485e59f34b804cecede"} Nov 25 10:48:52 crc kubenswrapper[4821]: I1125 10:48:52.481571 4821 generic.go:334] "Generic (PLEG): container finished" podID="2351168e-5815-4d63-b87d-e65011824547" containerID="7bb591d449410556307091427008c9c66dab10518332558ccb2c63301db45583" exitCode=0 Nov 25 10:48:52 crc kubenswrapper[4821]: I1125 10:48:52.481683 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-9qt9t" event={"ID":"2351168e-5815-4d63-b87d-e65011824547","Type":"ContainerDied","Data":"7bb591d449410556307091427008c9c66dab10518332558ccb2c63301db45583"} Nov 25 10:48:52 crc kubenswrapper[4821]: I1125 10:48:52.481728 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-9qt9t" event={"ID":"2351168e-5815-4d63-b87d-e65011824547","Type":"ContainerDied","Data":"5528b12573989025e5ba048441f3a409fd07232de5daa53360df0b789a08cb12"} Nov 25 10:48:52 crc kubenswrapper[4821]: I1125 10:48:52.481755 4821 scope.go:117] "RemoveContainer" containerID="7bb591d449410556307091427008c9c66dab10518332558ccb2c63301db45583" Nov 25 10:48:52 crc kubenswrapper[4821]: I1125 10:48:52.481893 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f59b8f679-9qt9t" Nov 25 10:48:52 crc kubenswrapper[4821]: I1125 10:48:52.493320 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bbf5cc879-b5qsb" event={"ID":"099443e1-44f8-4fa0-a567-2a8675d1d740","Type":"ContainerStarted","Data":"38f55b56330ea0a7d8a8d252271f17ac9666a5f663f11b8064b51fe6c00e522d"} Nov 25 10:48:52 crc kubenswrapper[4821]: I1125 10:48:52.505298 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-2mjz9" podStartSLOduration=2.505280869 podStartE2EDuration="2.505280869s" podCreationTimestamp="2025-11-25 10:48:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:48:52.503046333 +0000 UTC m=+1003.039366180" watchObservedRunningTime="2025-11-25 10:48:52.505280869 +0000 UTC m=+1003.041600706" Nov 25 10:48:52 crc kubenswrapper[4821]: I1125 10:48:52.513821 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2351168e-5815-4d63-b87d-e65011824547-ovsdbserver-sb\") pod \"2351168e-5815-4d63-b87d-e65011824547\" (UID: \"2351168e-5815-4d63-b87d-e65011824547\") " Nov 25 10:48:52 crc kubenswrapper[4821]: I1125 10:48:52.513891 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2351168e-5815-4d63-b87d-e65011824547-dns-svc\") pod \"2351168e-5815-4d63-b87d-e65011824547\" (UID: \"2351168e-5815-4d63-b87d-e65011824547\") " Nov 25 10:48:52 crc kubenswrapper[4821]: I1125 10:48:52.513966 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2351168e-5815-4d63-b87d-e65011824547-dns-swift-storage-0\") pod \"2351168e-5815-4d63-b87d-e65011824547\" (UID: \"2351168e-5815-4d63-b87d-e65011824547\") " Nov 25 10:48:52 crc kubenswrapper[4821]: I1125 10:48:52.514001 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6slbb\" (UniqueName: \"kubernetes.io/projected/2351168e-5815-4d63-b87d-e65011824547-kube-api-access-6slbb\") pod \"2351168e-5815-4d63-b87d-e65011824547\" (UID: \"2351168e-5815-4d63-b87d-e65011824547\") " Nov 25 10:48:52 crc kubenswrapper[4821]: I1125 10:48:52.514054 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2351168e-5815-4d63-b87d-e65011824547-config\") pod \"2351168e-5815-4d63-b87d-e65011824547\" (UID: \"2351168e-5815-4d63-b87d-e65011824547\") " Nov 25 10:48:52 crc kubenswrapper[4821]: I1125 10:48:52.514153 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2351168e-5815-4d63-b87d-e65011824547-ovsdbserver-nb\") pod \"2351168e-5815-4d63-b87d-e65011824547\" (UID: \"2351168e-5815-4d63-b87d-e65011824547\") " Nov 25 10:48:52 crc kubenswrapper[4821]: I1125 10:48:52.529251 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2351168e-5815-4d63-b87d-e65011824547-kube-api-access-6slbb" (OuterVolumeSpecName: "kube-api-access-6slbb") pod "2351168e-5815-4d63-b87d-e65011824547" (UID: "2351168e-5815-4d63-b87d-e65011824547"). InnerVolumeSpecName "kube-api-access-6slbb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:48:52 crc kubenswrapper[4821]: I1125 10:48:52.530644 4821 scope.go:117] "RemoveContainer" containerID="7bb591d449410556307091427008c9c66dab10518332558ccb2c63301db45583" Nov 25 10:48:52 crc kubenswrapper[4821]: E1125 10:48:52.531078 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7bb591d449410556307091427008c9c66dab10518332558ccb2c63301db45583\": container with ID starting with 7bb591d449410556307091427008c9c66dab10518332558ccb2c63301db45583 not found: ID does not exist" containerID="7bb591d449410556307091427008c9c66dab10518332558ccb2c63301db45583" Nov 25 10:48:52 crc kubenswrapper[4821]: I1125 10:48:52.531130 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7bb591d449410556307091427008c9c66dab10518332558ccb2c63301db45583"} err="failed to get container status \"7bb591d449410556307091427008c9c66dab10518332558ccb2c63301db45583\": rpc error: code = NotFound desc = could not find container \"7bb591d449410556307091427008c9c66dab10518332558ccb2c63301db45583\": container with ID starting with 7bb591d449410556307091427008c9c66dab10518332558ccb2c63301db45583 not found: ID does not exist" Nov 25 10:48:52 crc kubenswrapper[4821]: I1125 10:48:52.539346 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:48:52 crc kubenswrapper[4821]: I1125 10:48:52.551075 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-tq9mv"] Nov 25 10:48:52 crc kubenswrapper[4821]: I1125 10:48:52.602196 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2351168e-5815-4d63-b87d-e65011824547-config" (OuterVolumeSpecName: "config") pod "2351168e-5815-4d63-b87d-e65011824547" (UID: "2351168e-5815-4d63-b87d-e65011824547"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:48:52 crc kubenswrapper[4821]: I1125 10:48:52.610820 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2351168e-5815-4d63-b87d-e65011824547-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "2351168e-5815-4d63-b87d-e65011824547" (UID: "2351168e-5815-4d63-b87d-e65011824547"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:48:52 crc kubenswrapper[4821]: I1125 10:48:52.617020 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2351168e-5815-4d63-b87d-e65011824547-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "2351168e-5815-4d63-b87d-e65011824547" (UID: "2351168e-5815-4d63-b87d-e65011824547"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:48:52 crc kubenswrapper[4821]: I1125 10:48:52.617446 4821 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2351168e-5815-4d63-b87d-e65011824547-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 10:48:52 crc kubenswrapper[4821]: I1125 10:48:52.617465 4821 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2351168e-5815-4d63-b87d-e65011824547-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 10:48:52 crc kubenswrapper[4821]: I1125 10:48:52.617475 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6slbb\" (UniqueName: \"kubernetes.io/projected/2351168e-5815-4d63-b87d-e65011824547-kube-api-access-6slbb\") on node \"crc\" DevicePath \"\"" Nov 25 10:48:52 crc kubenswrapper[4821]: I1125 10:48:52.617489 4821 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2351168e-5815-4d63-b87d-e65011824547-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:48:52 crc kubenswrapper[4821]: I1125 10:48:52.627685 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2351168e-5815-4d63-b87d-e65011824547-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2351168e-5815-4d63-b87d-e65011824547" (UID: "2351168e-5815-4d63-b87d-e65011824547"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:48:52 crc kubenswrapper[4821]: I1125 10:48:52.632485 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2351168e-5815-4d63-b87d-e65011824547-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "2351168e-5815-4d63-b87d-e65011824547" (UID: "2351168e-5815-4d63-b87d-e65011824547"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:48:52 crc kubenswrapper[4821]: I1125 10:48:52.719022 4821 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2351168e-5815-4d63-b87d-e65011824547-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 25 10:48:52 crc kubenswrapper[4821]: I1125 10:48:52.719368 4821 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2351168e-5815-4d63-b87d-e65011824547-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 10:48:52 crc kubenswrapper[4821]: I1125 10:48:52.722389 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-8bp9n"] Nov 25 10:48:52 crc kubenswrapper[4821]: I1125 10:48:52.866576 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-9qt9t"] Nov 25 10:48:52 crc kubenswrapper[4821]: I1125 10:48:52.873387 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-9qt9t"] Nov 25 10:48:53 crc kubenswrapper[4821]: I1125 10:48:53.014705 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-zbmpp"] Nov 25 10:48:53 crc kubenswrapper[4821]: I1125 10:48:53.031866 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7558555969-g2z2s"] Nov 25 10:48:53 crc kubenswrapper[4821]: I1125 10:48:53.048873 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-574c6969fc-sjxzg"] Nov 25 10:48:53 crc kubenswrapper[4821]: I1125 10:48:53.063066 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-xxdmc"] Nov 25 10:48:53 crc kubenswrapper[4821]: I1125 10:48:53.083787 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-7znwh"] Nov 25 10:48:53 crc kubenswrapper[4821]: I1125 10:48:53.178086 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 10:48:53 crc kubenswrapper[4821]: I1125 10:48:53.509332 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9cb63eff-3e7e-436e-aea6-ce8254c3db77","Type":"ContainerStarted","Data":"3b31424b0b3432b8ccdba74c9228c79dbbfde9d71e6fd17d2545670fd87680df"} Nov 25 10:48:53 crc kubenswrapper[4821]: I1125 10:48:53.512883 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-574c6969fc-sjxzg" event={"ID":"79a12e89-074b-48f2-bf44-b177fe2e84e2","Type":"ContainerStarted","Data":"a796dab8becc201a1a5893ba83b9780af86245e2776ce6e5ed298b535f85e975"} Nov 25 10:48:53 crc kubenswrapper[4821]: I1125 10:48:53.514838 4821 generic.go:334] "Generic (PLEG): container finished" podID="d7c7075f-28ad-46d5-b301-98f3868873db" containerID="0a2ed13b5371cbf233dfa5f8b5c412fad3488bad7b3d4bc5787335ea66f93a7a" exitCode=0 Nov 25 10:48:53 crc kubenswrapper[4821]: I1125 10:48:53.514888 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-zbmpp" event={"ID":"d7c7075f-28ad-46d5-b301-98f3868873db","Type":"ContainerDied","Data":"0a2ed13b5371cbf233dfa5f8b5c412fad3488bad7b3d4bc5787335ea66f93a7a"} Nov 25 10:48:53 crc kubenswrapper[4821]: I1125 10:48:53.514906 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-zbmpp" event={"ID":"d7c7075f-28ad-46d5-b301-98f3868873db","Type":"ContainerStarted","Data":"f6cd48561cb54a03ba2af584e88ca05798dac2fdd347d3343872d689cfcdce5a"} Nov 25 10:48:53 crc kubenswrapper[4821]: I1125 10:48:53.518602 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"bf0e0161-6d4e-4149-a1f2-7b96ba1f736b","Type":"ContainerStarted","Data":"5259f06a31a8ce666b14c694ab5cd9ecf142cc6e6c37940d94fd6e1f47eae93b"} Nov 25 10:48:53 crc kubenswrapper[4821]: I1125 10:48:53.526535 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-8bp9n" event={"ID":"eb6c9acf-d69c-4788-a85d-b72135de1ef3","Type":"ContainerStarted","Data":"9c4305acf12d8d1e49951f88e49be96bacbd6239479548174634d1401b0bf93f"} Nov 25 10:48:53 crc kubenswrapper[4821]: I1125 10:48:53.528654 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-7znwh" event={"ID":"2eb6852b-84a1-4d4a-8402-77a7e2252ab2","Type":"ContainerStarted","Data":"f72d78a2fdc159af1f6b1a9f9d141b637201e9ebc83f8e48b6f743759e1b04fd"} Nov 25 10:48:53 crc kubenswrapper[4821]: I1125 10:48:53.530458 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-tq9mv" event={"ID":"325653d2-264c-4c6c-851f-6e38fe3e61a5","Type":"ContainerStarted","Data":"c9151e75ad24f3fd80b6828468427f0883ab2f705ae36476b0fd3c2fa4619697"} Nov 25 10:48:53 crc kubenswrapper[4821]: I1125 10:48:53.530493 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-tq9mv" event={"ID":"325653d2-264c-4c6c-851f-6e38fe3e61a5","Type":"ContainerStarted","Data":"d051074367e8fb3a3803c8649e596398c8187a81d4c341c26e7ff844f3526c4c"} Nov 25 10:48:53 crc kubenswrapper[4821]: I1125 10:48:53.542502 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-xxdmc" event={"ID":"bee9a405-f60c-449e-beba-71cd802daa79","Type":"ContainerStarted","Data":"14e27cff5971cdc47d84904e5d760f78db7c0348d4da8f82991561d3d2924af3"} Nov 25 10:48:53 crc kubenswrapper[4821]: I1125 10:48:53.552725 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7558555969-g2z2s" event={"ID":"77e0b5ce-8867-4d7e-b89e-170e67f3a548","Type":"ContainerStarted","Data":"f6e1c6c23c32f47b087d392cbf5933988edde50c045e7baa12e5de4c8bf6982e"} Nov 25 10:48:53 crc kubenswrapper[4821]: I1125 10:48:53.556210 4821 generic.go:334] "Generic (PLEG): container finished" podID="099443e1-44f8-4fa0-a567-2a8675d1d740" containerID="9ba5cf262eb76c3b0c6b8be404b80ad64c5b399c21e1af62764068a68d31852e" exitCode=0 Nov 25 10:48:53 crc kubenswrapper[4821]: I1125 10:48:53.557071 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bbf5cc879-b5qsb" event={"ID":"099443e1-44f8-4fa0-a567-2a8675d1d740","Type":"ContainerDied","Data":"9ba5cf262eb76c3b0c6b8be404b80ad64c5b399c21e1af62764068a68d31852e"} Nov 25 10:48:53 crc kubenswrapper[4821]: I1125 10:48:53.566473 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-tq9mv" podStartSLOduration=3.566448198 podStartE2EDuration="3.566448198s" podCreationTimestamp="2025-11-25 10:48:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:48:53.553741476 +0000 UTC m=+1004.090061323" watchObservedRunningTime="2025-11-25 10:48:53.566448198 +0000 UTC m=+1004.102768045" Nov 25 10:48:54 crc kubenswrapper[4821]: I1125 10:48:54.001700 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bbf5cc879-b5qsb" Nov 25 10:48:54 crc kubenswrapper[4821]: I1125 10:48:54.048968 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/099443e1-44f8-4fa0-a567-2a8675d1d740-dns-svc\") pod \"099443e1-44f8-4fa0-a567-2a8675d1d740\" (UID: \"099443e1-44f8-4fa0-a567-2a8675d1d740\") " Nov 25 10:48:54 crc kubenswrapper[4821]: I1125 10:48:54.049028 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-smwmq\" (UniqueName: \"kubernetes.io/projected/099443e1-44f8-4fa0-a567-2a8675d1d740-kube-api-access-smwmq\") pod \"099443e1-44f8-4fa0-a567-2a8675d1d740\" (UID: \"099443e1-44f8-4fa0-a567-2a8675d1d740\") " Nov 25 10:48:54 crc kubenswrapper[4821]: I1125 10:48:54.049071 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/099443e1-44f8-4fa0-a567-2a8675d1d740-config\") pod \"099443e1-44f8-4fa0-a567-2a8675d1d740\" (UID: \"099443e1-44f8-4fa0-a567-2a8675d1d740\") " Nov 25 10:48:54 crc kubenswrapper[4821]: I1125 10:48:54.049105 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/099443e1-44f8-4fa0-a567-2a8675d1d740-ovsdbserver-sb\") pod \"099443e1-44f8-4fa0-a567-2a8675d1d740\" (UID: \"099443e1-44f8-4fa0-a567-2a8675d1d740\") " Nov 25 10:48:54 crc kubenswrapper[4821]: I1125 10:48:54.049124 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/099443e1-44f8-4fa0-a567-2a8675d1d740-dns-swift-storage-0\") pod \"099443e1-44f8-4fa0-a567-2a8675d1d740\" (UID: \"099443e1-44f8-4fa0-a567-2a8675d1d740\") " Nov 25 10:48:54 crc kubenswrapper[4821]: I1125 10:48:54.049225 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/099443e1-44f8-4fa0-a567-2a8675d1d740-ovsdbserver-nb\") pod \"099443e1-44f8-4fa0-a567-2a8675d1d740\" (UID: \"099443e1-44f8-4fa0-a567-2a8675d1d740\") " Nov 25 10:48:54 crc kubenswrapper[4821]: I1125 10:48:54.062464 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/099443e1-44f8-4fa0-a567-2a8675d1d740-kube-api-access-smwmq" (OuterVolumeSpecName: "kube-api-access-smwmq") pod "099443e1-44f8-4fa0-a567-2a8675d1d740" (UID: "099443e1-44f8-4fa0-a567-2a8675d1d740"). InnerVolumeSpecName "kube-api-access-smwmq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:48:54 crc kubenswrapper[4821]: I1125 10:48:54.076897 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/099443e1-44f8-4fa0-a567-2a8675d1d740-config" (OuterVolumeSpecName: "config") pod "099443e1-44f8-4fa0-a567-2a8675d1d740" (UID: "099443e1-44f8-4fa0-a567-2a8675d1d740"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:48:54 crc kubenswrapper[4821]: I1125 10:48:54.077204 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/099443e1-44f8-4fa0-a567-2a8675d1d740-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "099443e1-44f8-4fa0-a567-2a8675d1d740" (UID: "099443e1-44f8-4fa0-a567-2a8675d1d740"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:48:54 crc kubenswrapper[4821]: I1125 10:48:54.080883 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/099443e1-44f8-4fa0-a567-2a8675d1d740-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "099443e1-44f8-4fa0-a567-2a8675d1d740" (UID: "099443e1-44f8-4fa0-a567-2a8675d1d740"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:48:54 crc kubenswrapper[4821]: I1125 10:48:54.084776 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/099443e1-44f8-4fa0-a567-2a8675d1d740-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "099443e1-44f8-4fa0-a567-2a8675d1d740" (UID: "099443e1-44f8-4fa0-a567-2a8675d1d740"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:48:54 crc kubenswrapper[4821]: I1125 10:48:54.085266 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/099443e1-44f8-4fa0-a567-2a8675d1d740-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "099443e1-44f8-4fa0-a567-2a8675d1d740" (UID: "099443e1-44f8-4fa0-a567-2a8675d1d740"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:48:54 crc kubenswrapper[4821]: I1125 10:48:54.138204 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2351168e-5815-4d63-b87d-e65011824547" path="/var/lib/kubelet/pods/2351168e-5815-4d63-b87d-e65011824547/volumes" Nov 25 10:48:54 crc kubenswrapper[4821]: I1125 10:48:54.150872 4821 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/099443e1-44f8-4fa0-a567-2a8675d1d740-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 10:48:54 crc kubenswrapper[4821]: I1125 10:48:54.150899 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-smwmq\" (UniqueName: \"kubernetes.io/projected/099443e1-44f8-4fa0-a567-2a8675d1d740-kube-api-access-smwmq\") on node \"crc\" DevicePath \"\"" Nov 25 10:48:54 crc kubenswrapper[4821]: I1125 10:48:54.150910 4821 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/099443e1-44f8-4fa0-a567-2a8675d1d740-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:48:54 crc kubenswrapper[4821]: I1125 10:48:54.150919 4821 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/099443e1-44f8-4fa0-a567-2a8675d1d740-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 10:48:54 crc kubenswrapper[4821]: I1125 10:48:54.150927 4821 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/099443e1-44f8-4fa0-a567-2a8675d1d740-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 25 10:48:54 crc kubenswrapper[4821]: I1125 10:48:54.150964 4821 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/099443e1-44f8-4fa0-a567-2a8675d1d740-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 10:48:54 crc kubenswrapper[4821]: I1125 10:48:54.261002 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 10:48:54 crc kubenswrapper[4821]: W1125 10:48:54.285433 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf2d41a59_94d9_4440_bfa6_f88e49b97aff.slice/crio-6f1ba2d6f3226e024f3cdd812e1d65127eedddcc9cfa8d63cb7c3899fcafe6d9 WatchSource:0}: Error finding container 6f1ba2d6f3226e024f3cdd812e1d65127eedddcc9cfa8d63cb7c3899fcafe6d9: Status 404 returned error can't find the container with id 6f1ba2d6f3226e024f3cdd812e1d65127eedddcc9cfa8d63cb7c3899fcafe6d9 Nov 25 10:48:54 crc kubenswrapper[4821]: I1125 10:48:54.573578 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bbf5cc879-b5qsb" event={"ID":"099443e1-44f8-4fa0-a567-2a8675d1d740","Type":"ContainerDied","Data":"38f55b56330ea0a7d8a8d252271f17ac9666a5f663f11b8064b51fe6c00e522d"} Nov 25 10:48:54 crc kubenswrapper[4821]: I1125 10:48:54.573669 4821 scope.go:117] "RemoveContainer" containerID="9ba5cf262eb76c3b0c6b8be404b80ad64c5b399c21e1af62764068a68d31852e" Nov 25 10:48:54 crc kubenswrapper[4821]: I1125 10:48:54.573870 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bbf5cc879-b5qsb" Nov 25 10:48:54 crc kubenswrapper[4821]: I1125 10:48:54.577198 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-zbmpp" event={"ID":"d7c7075f-28ad-46d5-b301-98f3868873db","Type":"ContainerStarted","Data":"369a7f9ccf800c63e11bd5c783755f2d86cc3c245075d4e9988f5bfb73f07b61"} Nov 25 10:48:54 crc kubenswrapper[4821]: I1125 10:48:54.578113 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-56df8fb6b7-zbmpp" Nov 25 10:48:54 crc kubenswrapper[4821]: I1125 10:48:54.579251 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f2d41a59-94d9-4440-bfa6-f88e49b97aff","Type":"ContainerStarted","Data":"6f1ba2d6f3226e024f3cdd812e1d65127eedddcc9cfa8d63cb7c3899fcafe6d9"} Nov 25 10:48:54 crc kubenswrapper[4821]: I1125 10:48:54.582873 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"bf0e0161-6d4e-4149-a1f2-7b96ba1f736b","Type":"ContainerStarted","Data":"49cd578d7f42537f45fec5bb96e4f5e441fc4f231f7aeb3564608683e64e988e"} Nov 25 10:48:54 crc kubenswrapper[4821]: I1125 10:48:54.641922 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-56df8fb6b7-zbmpp" podStartSLOduration=3.6419007 podStartE2EDuration="3.6419007s" podCreationTimestamp="2025-11-25 10:48:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:48:54.606605465 +0000 UTC m=+1005.142925312" watchObservedRunningTime="2025-11-25 10:48:54.6419007 +0000 UTC m=+1005.178220557" Nov 25 10:48:54 crc kubenswrapper[4821]: I1125 10:48:54.677719 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-b5qsb"] Nov 25 10:48:54 crc kubenswrapper[4821]: I1125 10:48:54.695788 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-b5qsb"] Nov 25 10:48:54 crc kubenswrapper[4821]: I1125 10:48:54.894136 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-574c6969fc-sjxzg"] Nov 25 10:48:54 crc kubenswrapper[4821]: I1125 10:48:54.906367 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:48:54 crc kubenswrapper[4821]: I1125 10:48:54.947775 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-64dc894fb9-5xf8b"] Nov 25 10:48:54 crc kubenswrapper[4821]: E1125 10:48:54.948194 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="099443e1-44f8-4fa0-a567-2a8675d1d740" containerName="init" Nov 25 10:48:54 crc kubenswrapper[4821]: I1125 10:48:54.948211 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="099443e1-44f8-4fa0-a567-2a8675d1d740" containerName="init" Nov 25 10:48:54 crc kubenswrapper[4821]: E1125 10:48:54.948223 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2351168e-5815-4d63-b87d-e65011824547" containerName="init" Nov 25 10:48:54 crc kubenswrapper[4821]: I1125 10:48:54.948229 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="2351168e-5815-4d63-b87d-e65011824547" containerName="init" Nov 25 10:48:54 crc kubenswrapper[4821]: I1125 10:48:54.948424 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="2351168e-5815-4d63-b87d-e65011824547" containerName="init" Nov 25 10:48:54 crc kubenswrapper[4821]: I1125 10:48:54.948444 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="099443e1-44f8-4fa0-a567-2a8675d1d740" containerName="init" Nov 25 10:48:54 crc kubenswrapper[4821]: I1125 10:48:54.949316 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-64dc894fb9-5xf8b" Nov 25 10:48:54 crc kubenswrapper[4821]: I1125 10:48:54.953373 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-64dc894fb9-5xf8b"] Nov 25 10:48:55 crc kubenswrapper[4821]: I1125 10:48:55.015852 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 10:48:55 crc kubenswrapper[4821]: I1125 10:48:55.079073 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f9398a81-673d-46b4-b560-91c6e2885532-config-data\") pod \"horizon-64dc894fb9-5xf8b\" (UID: \"f9398a81-673d-46b4-b560-91c6e2885532\") " pod="openstack/horizon-64dc894fb9-5xf8b" Nov 25 10:48:55 crc kubenswrapper[4821]: I1125 10:48:55.079182 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f9398a81-673d-46b4-b560-91c6e2885532-scripts\") pod \"horizon-64dc894fb9-5xf8b\" (UID: \"f9398a81-673d-46b4-b560-91c6e2885532\") " pod="openstack/horizon-64dc894fb9-5xf8b" Nov 25 10:48:55 crc kubenswrapper[4821]: I1125 10:48:55.079247 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f9398a81-673d-46b4-b560-91c6e2885532-horizon-secret-key\") pod \"horizon-64dc894fb9-5xf8b\" (UID: \"f9398a81-673d-46b4-b560-91c6e2885532\") " pod="openstack/horizon-64dc894fb9-5xf8b" Nov 25 10:48:55 crc kubenswrapper[4821]: I1125 10:48:55.079327 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xr2t\" (UniqueName: \"kubernetes.io/projected/f9398a81-673d-46b4-b560-91c6e2885532-kube-api-access-4xr2t\") pod \"horizon-64dc894fb9-5xf8b\" (UID: \"f9398a81-673d-46b4-b560-91c6e2885532\") " pod="openstack/horizon-64dc894fb9-5xf8b" Nov 25 10:48:55 crc kubenswrapper[4821]: I1125 10:48:55.079381 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f9398a81-673d-46b4-b560-91c6e2885532-logs\") pod \"horizon-64dc894fb9-5xf8b\" (UID: \"f9398a81-673d-46b4-b560-91c6e2885532\") " pod="openstack/horizon-64dc894fb9-5xf8b" Nov 25 10:48:55 crc kubenswrapper[4821]: I1125 10:48:55.089212 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 10:48:55 crc kubenswrapper[4821]: I1125 10:48:55.180662 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f9398a81-673d-46b4-b560-91c6e2885532-logs\") pod \"horizon-64dc894fb9-5xf8b\" (UID: \"f9398a81-673d-46b4-b560-91c6e2885532\") " pod="openstack/horizon-64dc894fb9-5xf8b" Nov 25 10:48:55 crc kubenswrapper[4821]: I1125 10:48:55.180761 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f9398a81-673d-46b4-b560-91c6e2885532-config-data\") pod \"horizon-64dc894fb9-5xf8b\" (UID: \"f9398a81-673d-46b4-b560-91c6e2885532\") " pod="openstack/horizon-64dc894fb9-5xf8b" Nov 25 10:48:55 crc kubenswrapper[4821]: I1125 10:48:55.180794 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f9398a81-673d-46b4-b560-91c6e2885532-scripts\") pod \"horizon-64dc894fb9-5xf8b\" (UID: \"f9398a81-673d-46b4-b560-91c6e2885532\") " pod="openstack/horizon-64dc894fb9-5xf8b" Nov 25 10:48:55 crc kubenswrapper[4821]: I1125 10:48:55.180837 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f9398a81-673d-46b4-b560-91c6e2885532-horizon-secret-key\") pod \"horizon-64dc894fb9-5xf8b\" (UID: \"f9398a81-673d-46b4-b560-91c6e2885532\") " pod="openstack/horizon-64dc894fb9-5xf8b" Nov 25 10:48:55 crc kubenswrapper[4821]: I1125 10:48:55.180921 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xr2t\" (UniqueName: \"kubernetes.io/projected/f9398a81-673d-46b4-b560-91c6e2885532-kube-api-access-4xr2t\") pod \"horizon-64dc894fb9-5xf8b\" (UID: \"f9398a81-673d-46b4-b560-91c6e2885532\") " pod="openstack/horizon-64dc894fb9-5xf8b" Nov 25 10:48:55 crc kubenswrapper[4821]: I1125 10:48:55.181072 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f9398a81-673d-46b4-b560-91c6e2885532-logs\") pod \"horizon-64dc894fb9-5xf8b\" (UID: \"f9398a81-673d-46b4-b560-91c6e2885532\") " pod="openstack/horizon-64dc894fb9-5xf8b" Nov 25 10:48:55 crc kubenswrapper[4821]: I1125 10:48:55.181546 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f9398a81-673d-46b4-b560-91c6e2885532-scripts\") pod \"horizon-64dc894fb9-5xf8b\" (UID: \"f9398a81-673d-46b4-b560-91c6e2885532\") " pod="openstack/horizon-64dc894fb9-5xf8b" Nov 25 10:48:55 crc kubenswrapper[4821]: I1125 10:48:55.182796 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f9398a81-673d-46b4-b560-91c6e2885532-config-data\") pod \"horizon-64dc894fb9-5xf8b\" (UID: \"f9398a81-673d-46b4-b560-91c6e2885532\") " pod="openstack/horizon-64dc894fb9-5xf8b" Nov 25 10:48:55 crc kubenswrapper[4821]: I1125 10:48:55.187458 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f9398a81-673d-46b4-b560-91c6e2885532-horizon-secret-key\") pod \"horizon-64dc894fb9-5xf8b\" (UID: \"f9398a81-673d-46b4-b560-91c6e2885532\") " pod="openstack/horizon-64dc894fb9-5xf8b" Nov 25 10:48:55 crc kubenswrapper[4821]: I1125 10:48:55.201275 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xr2t\" (UniqueName: \"kubernetes.io/projected/f9398a81-673d-46b4-b560-91c6e2885532-kube-api-access-4xr2t\") pod \"horizon-64dc894fb9-5xf8b\" (UID: \"f9398a81-673d-46b4-b560-91c6e2885532\") " pod="openstack/horizon-64dc894fb9-5xf8b" Nov 25 10:48:55 crc kubenswrapper[4821]: I1125 10:48:55.279569 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-64dc894fb9-5xf8b" Nov 25 10:48:55 crc kubenswrapper[4821]: I1125 10:48:55.619453 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f2d41a59-94d9-4440-bfa6-f88e49b97aff","Type":"ContainerStarted","Data":"a8de04974a2bc9e16f9435e3c9a56a7318997f65d8aa32c8ba43007fc4807cd7"} Nov 25 10:48:55 crc kubenswrapper[4821]: I1125 10:48:55.832366 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-64dc894fb9-5xf8b"] Nov 25 10:48:55 crc kubenswrapper[4821]: W1125 10:48:55.835918 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf9398a81_673d_46b4_b560_91c6e2885532.slice/crio-72b85ac1dd8b375e0cbd78d7a4025b3b5766312305a02ad8d0e6bd5cfeec8ee1 WatchSource:0}: Error finding container 72b85ac1dd8b375e0cbd78d7a4025b3b5766312305a02ad8d0e6bd5cfeec8ee1: Status 404 returned error can't find the container with id 72b85ac1dd8b375e0cbd78d7a4025b3b5766312305a02ad8d0e6bd5cfeec8ee1 Nov 25 10:48:56 crc kubenswrapper[4821]: I1125 10:48:56.126923 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="099443e1-44f8-4fa0-a567-2a8675d1d740" path="/var/lib/kubelet/pods/099443e1-44f8-4fa0-a567-2a8675d1d740/volumes" Nov 25 10:48:56 crc kubenswrapper[4821]: I1125 10:48:56.639661 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"bf0e0161-6d4e-4149-a1f2-7b96ba1f736b","Type":"ContainerStarted","Data":"3e45b3f1fae353f637f1c1a2354f109263e05849d38775faf9c1b8134e76fb3c"} Nov 25 10:48:56 crc kubenswrapper[4821]: I1125 10:48:56.639807 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="bf0e0161-6d4e-4149-a1f2-7b96ba1f736b" containerName="glance-log" containerID="cri-o://49cd578d7f42537f45fec5bb96e4f5e441fc4f231f7aeb3564608683e64e988e" gracePeriod=30 Nov 25 10:48:56 crc kubenswrapper[4821]: I1125 10:48:56.639963 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="bf0e0161-6d4e-4149-a1f2-7b96ba1f736b" containerName="glance-httpd" containerID="cri-o://3e45b3f1fae353f637f1c1a2354f109263e05849d38775faf9c1b8134e76fb3c" gracePeriod=30 Nov 25 10:48:56 crc kubenswrapper[4821]: I1125 10:48:56.641367 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-64dc894fb9-5xf8b" event={"ID":"f9398a81-673d-46b4-b560-91c6e2885532","Type":"ContainerStarted","Data":"72b85ac1dd8b375e0cbd78d7a4025b3b5766312305a02ad8d0e6bd5cfeec8ee1"} Nov 25 10:48:56 crc kubenswrapper[4821]: I1125 10:48:56.660099 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=5.660081666 podStartE2EDuration="5.660081666s" podCreationTimestamp="2025-11-25 10:48:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:48:56.658935007 +0000 UTC m=+1007.195254854" watchObservedRunningTime="2025-11-25 10:48:56.660081666 +0000 UTC m=+1007.196401503" Nov 25 10:48:57 crc kubenswrapper[4821]: I1125 10:48:57.658832 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f2d41a59-94d9-4440-bfa6-f88e49b97aff","Type":"ContainerStarted","Data":"68099be94ad7e6d76b625f0ec140658935ff46e4c9a2f0e5b6cbdd31ba331c97"} Nov 25 10:48:57 crc kubenswrapper[4821]: I1125 10:48:57.658906 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="f2d41a59-94d9-4440-bfa6-f88e49b97aff" containerName="glance-log" containerID="cri-o://a8de04974a2bc9e16f9435e3c9a56a7318997f65d8aa32c8ba43007fc4807cd7" gracePeriod=30 Nov 25 10:48:57 crc kubenswrapper[4821]: I1125 10:48:57.658988 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="f2d41a59-94d9-4440-bfa6-f88e49b97aff" containerName="glance-httpd" containerID="cri-o://68099be94ad7e6d76b625f0ec140658935ff46e4c9a2f0e5b6cbdd31ba331c97" gracePeriod=30 Nov 25 10:48:57 crc kubenswrapper[4821]: I1125 10:48:57.664282 4821 generic.go:334] "Generic (PLEG): container finished" podID="bf0e0161-6d4e-4149-a1f2-7b96ba1f736b" containerID="3e45b3f1fae353f637f1c1a2354f109263e05849d38775faf9c1b8134e76fb3c" exitCode=0 Nov 25 10:48:57 crc kubenswrapper[4821]: I1125 10:48:57.665019 4821 generic.go:334] "Generic (PLEG): container finished" podID="bf0e0161-6d4e-4149-a1f2-7b96ba1f736b" containerID="49cd578d7f42537f45fec5bb96e4f5e441fc4f231f7aeb3564608683e64e988e" exitCode=143 Nov 25 10:48:57 crc kubenswrapper[4821]: I1125 10:48:57.664774 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"bf0e0161-6d4e-4149-a1f2-7b96ba1f736b","Type":"ContainerDied","Data":"3e45b3f1fae353f637f1c1a2354f109263e05849d38775faf9c1b8134e76fb3c"} Nov 25 10:48:57 crc kubenswrapper[4821]: I1125 10:48:57.665068 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"bf0e0161-6d4e-4149-a1f2-7b96ba1f736b","Type":"ContainerDied","Data":"49cd578d7f42537f45fec5bb96e4f5e441fc4f231f7aeb3564608683e64e988e"} Nov 25 10:48:57 crc kubenswrapper[4821]: I1125 10:48:57.687592 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=6.687576201 podStartE2EDuration="6.687576201s" podCreationTimestamp="2025-11-25 10:48:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:48:57.682523103 +0000 UTC m=+1008.218842960" watchObservedRunningTime="2025-11-25 10:48:57.687576201 +0000 UTC m=+1008.223896048" Nov 25 10:48:58 crc kubenswrapper[4821]: I1125 10:48:58.676204 4821 generic.go:334] "Generic (PLEG): container finished" podID="a8462fa1-9270-4fa6-ae89-ea51e6246764" containerID="ac7e7a6e7fa3583d81e45e8d909f8b18b87b37f18db9fdb40be515f4278bf153" exitCode=0 Nov 25 10:48:58 crc kubenswrapper[4821]: I1125 10:48:58.676269 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2mjz9" event={"ID":"a8462fa1-9270-4fa6-ae89-ea51e6246764","Type":"ContainerDied","Data":"ac7e7a6e7fa3583d81e45e8d909f8b18b87b37f18db9fdb40be515f4278bf153"} Nov 25 10:48:58 crc kubenswrapper[4821]: I1125 10:48:58.679495 4821 generic.go:334] "Generic (PLEG): container finished" podID="f2d41a59-94d9-4440-bfa6-f88e49b97aff" containerID="68099be94ad7e6d76b625f0ec140658935ff46e4c9a2f0e5b6cbdd31ba331c97" exitCode=0 Nov 25 10:48:58 crc kubenswrapper[4821]: I1125 10:48:58.679523 4821 generic.go:334] "Generic (PLEG): container finished" podID="f2d41a59-94d9-4440-bfa6-f88e49b97aff" containerID="a8de04974a2bc9e16f9435e3c9a56a7318997f65d8aa32c8ba43007fc4807cd7" exitCode=143 Nov 25 10:48:58 crc kubenswrapper[4821]: I1125 10:48:58.679541 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f2d41a59-94d9-4440-bfa6-f88e49b97aff","Type":"ContainerDied","Data":"68099be94ad7e6d76b625f0ec140658935ff46e4c9a2f0e5b6cbdd31ba331c97"} Nov 25 10:48:58 crc kubenswrapper[4821]: I1125 10:48:58.679562 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f2d41a59-94d9-4440-bfa6-f88e49b97aff","Type":"ContainerDied","Data":"a8de04974a2bc9e16f9435e3c9a56a7318997f65d8aa32c8ba43007fc4807cd7"} Nov 25 10:49:00 crc kubenswrapper[4821]: I1125 10:49:00.592617 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7558555969-g2z2s"] Nov 25 10:49:00 crc kubenswrapper[4821]: I1125 10:49:00.642708 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-869c6f5bfb-l785w"] Nov 25 10:49:00 crc kubenswrapper[4821]: I1125 10:49:00.654004 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-869c6f5bfb-l785w" Nov 25 10:49:00 crc kubenswrapper[4821]: I1125 10:49:00.656510 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-869c6f5bfb-l785w"] Nov 25 10:49:00 crc kubenswrapper[4821]: I1125 10:49:00.659751 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Nov 25 10:49:00 crc kubenswrapper[4821]: I1125 10:49:00.703788 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-64dc894fb9-5xf8b"] Nov 25 10:49:00 crc kubenswrapper[4821]: I1125 10:49:00.739293 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-67f765d856-ncc6t"] Nov 25 10:49:00 crc kubenswrapper[4821]: I1125 10:49:00.741070 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-67f765d856-ncc6t" Nov 25 10:49:00 crc kubenswrapper[4821]: I1125 10:49:00.750532 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-67f765d856-ncc6t"] Nov 25 10:49:00 crc kubenswrapper[4821]: I1125 10:49:00.812895 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e2327ad3-5d88-4f4b-91ea-91deaf7fc249-scripts\") pod \"horizon-869c6f5bfb-l785w\" (UID: \"e2327ad3-5d88-4f4b-91ea-91deaf7fc249\") " pod="openstack/horizon-869c6f5bfb-l785w" Nov 25 10:49:00 crc kubenswrapper[4821]: I1125 10:49:00.812970 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/93b4abfa-c14a-4df9-a6d1-9cdeaec918dd-horizon-tls-certs\") pod \"horizon-67f765d856-ncc6t\" (UID: \"93b4abfa-c14a-4df9-a6d1-9cdeaec918dd\") " pod="openstack/horizon-67f765d856-ncc6t" Nov 25 10:49:00 crc kubenswrapper[4821]: I1125 10:49:00.813014 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e2327ad3-5d88-4f4b-91ea-91deaf7fc249-logs\") pod \"horizon-869c6f5bfb-l785w\" (UID: \"e2327ad3-5d88-4f4b-91ea-91deaf7fc249\") " pod="openstack/horizon-869c6f5bfb-l785w" Nov 25 10:49:00 crc kubenswrapper[4821]: I1125 10:49:00.813057 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e2327ad3-5d88-4f4b-91ea-91deaf7fc249-config-data\") pod \"horizon-869c6f5bfb-l785w\" (UID: \"e2327ad3-5d88-4f4b-91ea-91deaf7fc249\") " pod="openstack/horizon-869c6f5bfb-l785w" Nov 25 10:49:00 crc kubenswrapper[4821]: I1125 10:49:00.813079 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/93b4abfa-c14a-4df9-a6d1-9cdeaec918dd-horizon-secret-key\") pod \"horizon-67f765d856-ncc6t\" (UID: \"93b4abfa-c14a-4df9-a6d1-9cdeaec918dd\") " pod="openstack/horizon-67f765d856-ncc6t" Nov 25 10:49:00 crc kubenswrapper[4821]: I1125 10:49:00.813125 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/93b4abfa-c14a-4df9-a6d1-9cdeaec918dd-logs\") pod \"horizon-67f765d856-ncc6t\" (UID: \"93b4abfa-c14a-4df9-a6d1-9cdeaec918dd\") " pod="openstack/horizon-67f765d856-ncc6t" Nov 25 10:49:00 crc kubenswrapper[4821]: I1125 10:49:00.813188 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6tvvk\" (UniqueName: \"kubernetes.io/projected/e2327ad3-5d88-4f4b-91ea-91deaf7fc249-kube-api-access-6tvvk\") pod \"horizon-869c6f5bfb-l785w\" (UID: \"e2327ad3-5d88-4f4b-91ea-91deaf7fc249\") " pod="openstack/horizon-869c6f5bfb-l785w" Nov 25 10:49:00 crc kubenswrapper[4821]: I1125 10:49:00.813231 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4f8qg\" (UniqueName: \"kubernetes.io/projected/93b4abfa-c14a-4df9-a6d1-9cdeaec918dd-kube-api-access-4f8qg\") pod \"horizon-67f765d856-ncc6t\" (UID: \"93b4abfa-c14a-4df9-a6d1-9cdeaec918dd\") " pod="openstack/horizon-67f765d856-ncc6t" Nov 25 10:49:00 crc kubenswrapper[4821]: I1125 10:49:00.813274 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2327ad3-5d88-4f4b-91ea-91deaf7fc249-combined-ca-bundle\") pod \"horizon-869c6f5bfb-l785w\" (UID: \"e2327ad3-5d88-4f4b-91ea-91deaf7fc249\") " pod="openstack/horizon-869c6f5bfb-l785w" Nov 25 10:49:00 crc kubenswrapper[4821]: I1125 10:49:00.813300 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/e2327ad3-5d88-4f4b-91ea-91deaf7fc249-horizon-tls-certs\") pod \"horizon-869c6f5bfb-l785w\" (UID: \"e2327ad3-5d88-4f4b-91ea-91deaf7fc249\") " pod="openstack/horizon-869c6f5bfb-l785w" Nov 25 10:49:00 crc kubenswrapper[4821]: I1125 10:49:00.813359 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/93b4abfa-c14a-4df9-a6d1-9cdeaec918dd-config-data\") pod \"horizon-67f765d856-ncc6t\" (UID: \"93b4abfa-c14a-4df9-a6d1-9cdeaec918dd\") " pod="openstack/horizon-67f765d856-ncc6t" Nov 25 10:49:00 crc kubenswrapper[4821]: I1125 10:49:00.813381 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/93b4abfa-c14a-4df9-a6d1-9cdeaec918dd-scripts\") pod \"horizon-67f765d856-ncc6t\" (UID: \"93b4abfa-c14a-4df9-a6d1-9cdeaec918dd\") " pod="openstack/horizon-67f765d856-ncc6t" Nov 25 10:49:00 crc kubenswrapper[4821]: I1125 10:49:00.813436 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e2327ad3-5d88-4f4b-91ea-91deaf7fc249-horizon-secret-key\") pod \"horizon-869c6f5bfb-l785w\" (UID: \"e2327ad3-5d88-4f4b-91ea-91deaf7fc249\") " pod="openstack/horizon-869c6f5bfb-l785w" Nov 25 10:49:00 crc kubenswrapper[4821]: I1125 10:49:00.813461 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93b4abfa-c14a-4df9-a6d1-9cdeaec918dd-combined-ca-bundle\") pod \"horizon-67f765d856-ncc6t\" (UID: \"93b4abfa-c14a-4df9-a6d1-9cdeaec918dd\") " pod="openstack/horizon-67f765d856-ncc6t" Nov 25 10:49:00 crc kubenswrapper[4821]: I1125 10:49:00.915142 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e2327ad3-5d88-4f4b-91ea-91deaf7fc249-scripts\") pod \"horizon-869c6f5bfb-l785w\" (UID: \"e2327ad3-5d88-4f4b-91ea-91deaf7fc249\") " pod="openstack/horizon-869c6f5bfb-l785w" Nov 25 10:49:00 crc kubenswrapper[4821]: I1125 10:49:00.915229 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/93b4abfa-c14a-4df9-a6d1-9cdeaec918dd-horizon-tls-certs\") pod \"horizon-67f765d856-ncc6t\" (UID: \"93b4abfa-c14a-4df9-a6d1-9cdeaec918dd\") " pod="openstack/horizon-67f765d856-ncc6t" Nov 25 10:49:00 crc kubenswrapper[4821]: I1125 10:49:00.915287 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e2327ad3-5d88-4f4b-91ea-91deaf7fc249-logs\") pod \"horizon-869c6f5bfb-l785w\" (UID: \"e2327ad3-5d88-4f4b-91ea-91deaf7fc249\") " pod="openstack/horizon-869c6f5bfb-l785w" Nov 25 10:49:00 crc kubenswrapper[4821]: I1125 10:49:00.915314 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e2327ad3-5d88-4f4b-91ea-91deaf7fc249-config-data\") pod \"horizon-869c6f5bfb-l785w\" (UID: \"e2327ad3-5d88-4f4b-91ea-91deaf7fc249\") " pod="openstack/horizon-869c6f5bfb-l785w" Nov 25 10:49:00 crc kubenswrapper[4821]: I1125 10:49:00.915340 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/93b4abfa-c14a-4df9-a6d1-9cdeaec918dd-horizon-secret-key\") pod \"horizon-67f765d856-ncc6t\" (UID: \"93b4abfa-c14a-4df9-a6d1-9cdeaec918dd\") " pod="openstack/horizon-67f765d856-ncc6t" Nov 25 10:49:00 crc kubenswrapper[4821]: I1125 10:49:00.915382 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/93b4abfa-c14a-4df9-a6d1-9cdeaec918dd-logs\") pod \"horizon-67f765d856-ncc6t\" (UID: \"93b4abfa-c14a-4df9-a6d1-9cdeaec918dd\") " pod="openstack/horizon-67f765d856-ncc6t" Nov 25 10:49:00 crc kubenswrapper[4821]: I1125 10:49:00.915424 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6tvvk\" (UniqueName: \"kubernetes.io/projected/e2327ad3-5d88-4f4b-91ea-91deaf7fc249-kube-api-access-6tvvk\") pod \"horizon-869c6f5bfb-l785w\" (UID: \"e2327ad3-5d88-4f4b-91ea-91deaf7fc249\") " pod="openstack/horizon-869c6f5bfb-l785w" Nov 25 10:49:00 crc kubenswrapper[4821]: I1125 10:49:00.915487 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4f8qg\" (UniqueName: \"kubernetes.io/projected/93b4abfa-c14a-4df9-a6d1-9cdeaec918dd-kube-api-access-4f8qg\") pod \"horizon-67f765d856-ncc6t\" (UID: \"93b4abfa-c14a-4df9-a6d1-9cdeaec918dd\") " pod="openstack/horizon-67f765d856-ncc6t" Nov 25 10:49:00 crc kubenswrapper[4821]: I1125 10:49:00.915514 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2327ad3-5d88-4f4b-91ea-91deaf7fc249-combined-ca-bundle\") pod \"horizon-869c6f5bfb-l785w\" (UID: \"e2327ad3-5d88-4f4b-91ea-91deaf7fc249\") " pod="openstack/horizon-869c6f5bfb-l785w" Nov 25 10:49:00 crc kubenswrapper[4821]: I1125 10:49:00.915785 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/e2327ad3-5d88-4f4b-91ea-91deaf7fc249-horizon-tls-certs\") pod \"horizon-869c6f5bfb-l785w\" (UID: \"e2327ad3-5d88-4f4b-91ea-91deaf7fc249\") " pod="openstack/horizon-869c6f5bfb-l785w" Nov 25 10:49:00 crc kubenswrapper[4821]: I1125 10:49:00.915847 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e2327ad3-5d88-4f4b-91ea-91deaf7fc249-logs\") pod \"horizon-869c6f5bfb-l785w\" (UID: \"e2327ad3-5d88-4f4b-91ea-91deaf7fc249\") " pod="openstack/horizon-869c6f5bfb-l785w" Nov 25 10:49:00 crc kubenswrapper[4821]: I1125 10:49:00.915942 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/93b4abfa-c14a-4df9-a6d1-9cdeaec918dd-logs\") pod \"horizon-67f765d856-ncc6t\" (UID: \"93b4abfa-c14a-4df9-a6d1-9cdeaec918dd\") " pod="openstack/horizon-67f765d856-ncc6t" Nov 25 10:49:00 crc kubenswrapper[4821]: I1125 10:49:00.916114 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e2327ad3-5d88-4f4b-91ea-91deaf7fc249-scripts\") pod \"horizon-869c6f5bfb-l785w\" (UID: \"e2327ad3-5d88-4f4b-91ea-91deaf7fc249\") " pod="openstack/horizon-869c6f5bfb-l785w" Nov 25 10:49:00 crc kubenswrapper[4821]: I1125 10:49:00.916532 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/93b4abfa-c14a-4df9-a6d1-9cdeaec918dd-config-data\") pod \"horizon-67f765d856-ncc6t\" (UID: \"93b4abfa-c14a-4df9-a6d1-9cdeaec918dd\") " pod="openstack/horizon-67f765d856-ncc6t" Nov 25 10:49:00 crc kubenswrapper[4821]: I1125 10:49:00.916567 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/93b4abfa-c14a-4df9-a6d1-9cdeaec918dd-scripts\") pod \"horizon-67f765d856-ncc6t\" (UID: \"93b4abfa-c14a-4df9-a6d1-9cdeaec918dd\") " pod="openstack/horizon-67f765d856-ncc6t" Nov 25 10:49:00 crc kubenswrapper[4821]: I1125 10:49:00.916625 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e2327ad3-5d88-4f4b-91ea-91deaf7fc249-horizon-secret-key\") pod \"horizon-869c6f5bfb-l785w\" (UID: \"e2327ad3-5d88-4f4b-91ea-91deaf7fc249\") " pod="openstack/horizon-869c6f5bfb-l785w" Nov 25 10:49:00 crc kubenswrapper[4821]: I1125 10:49:00.916658 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93b4abfa-c14a-4df9-a6d1-9cdeaec918dd-combined-ca-bundle\") pod \"horizon-67f765d856-ncc6t\" (UID: \"93b4abfa-c14a-4df9-a6d1-9cdeaec918dd\") " pod="openstack/horizon-67f765d856-ncc6t" Nov 25 10:49:00 crc kubenswrapper[4821]: I1125 10:49:00.918833 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/93b4abfa-c14a-4df9-a6d1-9cdeaec918dd-scripts\") pod \"horizon-67f765d856-ncc6t\" (UID: \"93b4abfa-c14a-4df9-a6d1-9cdeaec918dd\") " pod="openstack/horizon-67f765d856-ncc6t" Nov 25 10:49:00 crc kubenswrapper[4821]: I1125 10:49:00.918938 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e2327ad3-5d88-4f4b-91ea-91deaf7fc249-config-data\") pod \"horizon-869c6f5bfb-l785w\" (UID: \"e2327ad3-5d88-4f4b-91ea-91deaf7fc249\") " pod="openstack/horizon-869c6f5bfb-l785w" Nov 25 10:49:00 crc kubenswrapper[4821]: I1125 10:49:00.919833 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/93b4abfa-c14a-4df9-a6d1-9cdeaec918dd-config-data\") pod \"horizon-67f765d856-ncc6t\" (UID: \"93b4abfa-c14a-4df9-a6d1-9cdeaec918dd\") " pod="openstack/horizon-67f765d856-ncc6t" Nov 25 10:49:00 crc kubenswrapper[4821]: I1125 10:49:00.921809 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/93b4abfa-c14a-4df9-a6d1-9cdeaec918dd-horizon-secret-key\") pod \"horizon-67f765d856-ncc6t\" (UID: \"93b4abfa-c14a-4df9-a6d1-9cdeaec918dd\") " pod="openstack/horizon-67f765d856-ncc6t" Nov 25 10:49:00 crc kubenswrapper[4821]: I1125 10:49:00.921835 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/93b4abfa-c14a-4df9-a6d1-9cdeaec918dd-horizon-tls-certs\") pod \"horizon-67f765d856-ncc6t\" (UID: \"93b4abfa-c14a-4df9-a6d1-9cdeaec918dd\") " pod="openstack/horizon-67f765d856-ncc6t" Nov 25 10:49:00 crc kubenswrapper[4821]: I1125 10:49:00.922136 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2327ad3-5d88-4f4b-91ea-91deaf7fc249-combined-ca-bundle\") pod \"horizon-869c6f5bfb-l785w\" (UID: \"e2327ad3-5d88-4f4b-91ea-91deaf7fc249\") " pod="openstack/horizon-869c6f5bfb-l785w" Nov 25 10:49:00 crc kubenswrapper[4821]: I1125 10:49:00.923613 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93b4abfa-c14a-4df9-a6d1-9cdeaec918dd-combined-ca-bundle\") pod \"horizon-67f765d856-ncc6t\" (UID: \"93b4abfa-c14a-4df9-a6d1-9cdeaec918dd\") " pod="openstack/horizon-67f765d856-ncc6t" Nov 25 10:49:00 crc kubenswrapper[4821]: I1125 10:49:00.923661 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/e2327ad3-5d88-4f4b-91ea-91deaf7fc249-horizon-tls-certs\") pod \"horizon-869c6f5bfb-l785w\" (UID: \"e2327ad3-5d88-4f4b-91ea-91deaf7fc249\") " pod="openstack/horizon-869c6f5bfb-l785w" Nov 25 10:49:00 crc kubenswrapper[4821]: I1125 10:49:00.925532 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e2327ad3-5d88-4f4b-91ea-91deaf7fc249-horizon-secret-key\") pod \"horizon-869c6f5bfb-l785w\" (UID: \"e2327ad3-5d88-4f4b-91ea-91deaf7fc249\") " pod="openstack/horizon-869c6f5bfb-l785w" Nov 25 10:49:00 crc kubenswrapper[4821]: I1125 10:49:00.930093 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6tvvk\" (UniqueName: \"kubernetes.io/projected/e2327ad3-5d88-4f4b-91ea-91deaf7fc249-kube-api-access-6tvvk\") pod \"horizon-869c6f5bfb-l785w\" (UID: \"e2327ad3-5d88-4f4b-91ea-91deaf7fc249\") " pod="openstack/horizon-869c6f5bfb-l785w" Nov 25 10:49:00 crc kubenswrapper[4821]: I1125 10:49:00.938037 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4f8qg\" (UniqueName: \"kubernetes.io/projected/93b4abfa-c14a-4df9-a6d1-9cdeaec918dd-kube-api-access-4f8qg\") pod \"horizon-67f765d856-ncc6t\" (UID: \"93b4abfa-c14a-4df9-a6d1-9cdeaec918dd\") " pod="openstack/horizon-67f765d856-ncc6t" Nov 25 10:49:00 crc kubenswrapper[4821]: I1125 10:49:00.992497 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-869c6f5bfb-l785w" Nov 25 10:49:01 crc kubenswrapper[4821]: I1125 10:49:01.066948 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-67f765d856-ncc6t" Nov 25 10:49:01 crc kubenswrapper[4821]: I1125 10:49:01.708301 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-56df8fb6b7-zbmpp" Nov 25 10:49:01 crc kubenswrapper[4821]: I1125 10:49:01.789034 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74dc88fc-bqbsk"] Nov 25 10:49:01 crc kubenswrapper[4821]: I1125 10:49:01.789280 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-74dc88fc-bqbsk" podUID="ccd85fd6-b0a0-4184-8082-054d2de0adec" containerName="dnsmasq-dns" containerID="cri-o://67a95be9b3aa84d207a73ce02dd8c799210d3c4070ac21ef95de932bfab350ab" gracePeriod=10 Nov 25 10:49:02 crc kubenswrapper[4821]: I1125 10:49:02.722674 4821 generic.go:334] "Generic (PLEG): container finished" podID="ccd85fd6-b0a0-4184-8082-054d2de0adec" containerID="67a95be9b3aa84d207a73ce02dd8c799210d3c4070ac21ef95de932bfab350ab" exitCode=0 Nov 25 10:49:02 crc kubenswrapper[4821]: I1125 10:49:02.722713 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74dc88fc-bqbsk" event={"ID":"ccd85fd6-b0a0-4184-8082-054d2de0adec","Type":"ContainerDied","Data":"67a95be9b3aa84d207a73ce02dd8c799210d3c4070ac21ef95de932bfab350ab"} Nov 25 10:49:02 crc kubenswrapper[4821]: I1125 10:49:02.965435 4821 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-74dc88fc-bqbsk" podUID="ccd85fd6-b0a0-4184-8082-054d2de0adec" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.133:5353: connect: connection refused" Nov 25 10:49:05 crc kubenswrapper[4821]: I1125 10:49:05.235593 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 25 10:49:05 crc kubenswrapper[4821]: I1125 10:49:05.393778 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dfhml\" (UniqueName: \"kubernetes.io/projected/bf0e0161-6d4e-4149-a1f2-7b96ba1f736b-kube-api-access-dfhml\") pod \"bf0e0161-6d4e-4149-a1f2-7b96ba1f736b\" (UID: \"bf0e0161-6d4e-4149-a1f2-7b96ba1f736b\") " Nov 25 10:49:05 crc kubenswrapper[4821]: I1125 10:49:05.393900 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf0e0161-6d4e-4149-a1f2-7b96ba1f736b-combined-ca-bundle\") pod \"bf0e0161-6d4e-4149-a1f2-7b96ba1f736b\" (UID: \"bf0e0161-6d4e-4149-a1f2-7b96ba1f736b\") " Nov 25 10:49:05 crc kubenswrapper[4821]: I1125 10:49:05.393933 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf0e0161-6d4e-4149-a1f2-7b96ba1f736b-config-data\") pod \"bf0e0161-6d4e-4149-a1f2-7b96ba1f736b\" (UID: \"bf0e0161-6d4e-4149-a1f2-7b96ba1f736b\") " Nov 25 10:49:05 crc kubenswrapper[4821]: I1125 10:49:05.394048 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bf0e0161-6d4e-4149-a1f2-7b96ba1f736b-httpd-run\") pod \"bf0e0161-6d4e-4149-a1f2-7b96ba1f736b\" (UID: \"bf0e0161-6d4e-4149-a1f2-7b96ba1f736b\") " Nov 25 10:49:05 crc kubenswrapper[4821]: I1125 10:49:05.394079 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf0e0161-6d4e-4149-a1f2-7b96ba1f736b-scripts\") pod \"bf0e0161-6d4e-4149-a1f2-7b96ba1f736b\" (UID: \"bf0e0161-6d4e-4149-a1f2-7b96ba1f736b\") " Nov 25 10:49:05 crc kubenswrapper[4821]: I1125 10:49:05.394100 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"bf0e0161-6d4e-4149-a1f2-7b96ba1f736b\" (UID: \"bf0e0161-6d4e-4149-a1f2-7b96ba1f736b\") " Nov 25 10:49:05 crc kubenswrapper[4821]: I1125 10:49:05.394196 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf0e0161-6d4e-4149-a1f2-7b96ba1f736b-internal-tls-certs\") pod \"bf0e0161-6d4e-4149-a1f2-7b96ba1f736b\" (UID: \"bf0e0161-6d4e-4149-a1f2-7b96ba1f736b\") " Nov 25 10:49:05 crc kubenswrapper[4821]: I1125 10:49:05.394231 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bf0e0161-6d4e-4149-a1f2-7b96ba1f736b-logs\") pod \"bf0e0161-6d4e-4149-a1f2-7b96ba1f736b\" (UID: \"bf0e0161-6d4e-4149-a1f2-7b96ba1f736b\") " Nov 25 10:49:05 crc kubenswrapper[4821]: I1125 10:49:05.395158 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf0e0161-6d4e-4149-a1f2-7b96ba1f736b-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "bf0e0161-6d4e-4149-a1f2-7b96ba1f736b" (UID: "bf0e0161-6d4e-4149-a1f2-7b96ba1f736b"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:49:05 crc kubenswrapper[4821]: I1125 10:49:05.395486 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf0e0161-6d4e-4149-a1f2-7b96ba1f736b-logs" (OuterVolumeSpecName: "logs") pod "bf0e0161-6d4e-4149-a1f2-7b96ba1f736b" (UID: "bf0e0161-6d4e-4149-a1f2-7b96ba1f736b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:49:05 crc kubenswrapper[4821]: I1125 10:49:05.402751 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "bf0e0161-6d4e-4149-a1f2-7b96ba1f736b" (UID: "bf0e0161-6d4e-4149-a1f2-7b96ba1f736b"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 10:49:05 crc kubenswrapper[4821]: I1125 10:49:05.402770 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf0e0161-6d4e-4149-a1f2-7b96ba1f736b-kube-api-access-dfhml" (OuterVolumeSpecName: "kube-api-access-dfhml") pod "bf0e0161-6d4e-4149-a1f2-7b96ba1f736b" (UID: "bf0e0161-6d4e-4149-a1f2-7b96ba1f736b"). InnerVolumeSpecName "kube-api-access-dfhml". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:49:05 crc kubenswrapper[4821]: I1125 10:49:05.414319 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf0e0161-6d4e-4149-a1f2-7b96ba1f736b-scripts" (OuterVolumeSpecName: "scripts") pod "bf0e0161-6d4e-4149-a1f2-7b96ba1f736b" (UID: "bf0e0161-6d4e-4149-a1f2-7b96ba1f736b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:49:05 crc kubenswrapper[4821]: I1125 10:49:05.426338 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf0e0161-6d4e-4149-a1f2-7b96ba1f736b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bf0e0161-6d4e-4149-a1f2-7b96ba1f736b" (UID: "bf0e0161-6d4e-4149-a1f2-7b96ba1f736b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:49:05 crc kubenswrapper[4821]: I1125 10:49:05.449373 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf0e0161-6d4e-4149-a1f2-7b96ba1f736b-config-data" (OuterVolumeSpecName: "config-data") pod "bf0e0161-6d4e-4149-a1f2-7b96ba1f736b" (UID: "bf0e0161-6d4e-4149-a1f2-7b96ba1f736b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:49:05 crc kubenswrapper[4821]: I1125 10:49:05.465303 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf0e0161-6d4e-4149-a1f2-7b96ba1f736b-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "bf0e0161-6d4e-4149-a1f2-7b96ba1f736b" (UID: "bf0e0161-6d4e-4149-a1f2-7b96ba1f736b"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:49:05 crc kubenswrapper[4821]: I1125 10:49:05.496951 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dfhml\" (UniqueName: \"kubernetes.io/projected/bf0e0161-6d4e-4149-a1f2-7b96ba1f736b-kube-api-access-dfhml\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:05 crc kubenswrapper[4821]: I1125 10:49:05.496986 4821 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf0e0161-6d4e-4149-a1f2-7b96ba1f736b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:05 crc kubenswrapper[4821]: I1125 10:49:05.497008 4821 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf0e0161-6d4e-4149-a1f2-7b96ba1f736b-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:05 crc kubenswrapper[4821]: I1125 10:49:05.497016 4821 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bf0e0161-6d4e-4149-a1f2-7b96ba1f736b-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:05 crc kubenswrapper[4821]: I1125 10:49:05.497025 4821 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf0e0161-6d4e-4149-a1f2-7b96ba1f736b-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:05 crc kubenswrapper[4821]: I1125 10:49:05.497058 4821 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Nov 25 10:49:05 crc kubenswrapper[4821]: I1125 10:49:05.497070 4821 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf0e0161-6d4e-4149-a1f2-7b96ba1f736b-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:05 crc kubenswrapper[4821]: I1125 10:49:05.497610 4821 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bf0e0161-6d4e-4149-a1f2-7b96ba1f736b-logs\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:05 crc kubenswrapper[4821]: I1125 10:49:05.514795 4821 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Nov 25 10:49:05 crc kubenswrapper[4821]: I1125 10:49:05.600011 4821 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:05 crc kubenswrapper[4821]: I1125 10:49:05.747982 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"bf0e0161-6d4e-4149-a1f2-7b96ba1f736b","Type":"ContainerDied","Data":"5259f06a31a8ce666b14c694ab5cd9ecf142cc6e6c37940d94fd6e1f47eae93b"} Nov 25 10:49:05 crc kubenswrapper[4821]: I1125 10:49:05.748035 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 25 10:49:05 crc kubenswrapper[4821]: I1125 10:49:05.748044 4821 scope.go:117] "RemoveContainer" containerID="3e45b3f1fae353f637f1c1a2354f109263e05849d38775faf9c1b8134e76fb3c" Nov 25 10:49:05 crc kubenswrapper[4821]: I1125 10:49:05.778468 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 10:49:05 crc kubenswrapper[4821]: I1125 10:49:05.785651 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 10:49:05 crc kubenswrapper[4821]: I1125 10:49:05.835030 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 10:49:05 crc kubenswrapper[4821]: E1125 10:49:05.841685 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf0e0161-6d4e-4149-a1f2-7b96ba1f736b" containerName="glance-log" Nov 25 10:49:05 crc kubenswrapper[4821]: I1125 10:49:05.841724 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf0e0161-6d4e-4149-a1f2-7b96ba1f736b" containerName="glance-log" Nov 25 10:49:05 crc kubenswrapper[4821]: E1125 10:49:05.841740 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf0e0161-6d4e-4149-a1f2-7b96ba1f736b" containerName="glance-httpd" Nov 25 10:49:05 crc kubenswrapper[4821]: I1125 10:49:05.841747 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf0e0161-6d4e-4149-a1f2-7b96ba1f736b" containerName="glance-httpd" Nov 25 10:49:05 crc kubenswrapper[4821]: I1125 10:49:05.841915 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf0e0161-6d4e-4149-a1f2-7b96ba1f736b" containerName="glance-httpd" Nov 25 10:49:05 crc kubenswrapper[4821]: I1125 10:49:05.841940 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf0e0161-6d4e-4149-a1f2-7b96ba1f736b" containerName="glance-log" Nov 25 10:49:05 crc kubenswrapper[4821]: I1125 10:49:05.845312 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 25 10:49:05 crc kubenswrapper[4821]: I1125 10:49:05.848113 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 10:49:05 crc kubenswrapper[4821]: I1125 10:49:05.849723 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 25 10:49:05 crc kubenswrapper[4821]: I1125 10:49:05.849932 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 25 10:49:06 crc kubenswrapper[4821]: I1125 10:49:06.017980 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16fdcebd-383a-4945-a5bf-a299b332bf8f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"16fdcebd-383a-4945-a5bf-a299b332bf8f\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:49:06 crc kubenswrapper[4821]: I1125 10:49:06.018206 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16fdcebd-383a-4945-a5bf-a299b332bf8f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"16fdcebd-383a-4945-a5bf-a299b332bf8f\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:49:06 crc kubenswrapper[4821]: I1125 10:49:06.018325 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/16fdcebd-383a-4945-a5bf-a299b332bf8f-logs\") pod \"glance-default-internal-api-0\" (UID: \"16fdcebd-383a-4945-a5bf-a299b332bf8f\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:49:06 crc kubenswrapper[4821]: I1125 10:49:06.018518 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/16fdcebd-383a-4945-a5bf-a299b332bf8f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"16fdcebd-383a-4945-a5bf-a299b332bf8f\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:49:06 crc kubenswrapper[4821]: I1125 10:49:06.018552 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"16fdcebd-383a-4945-a5bf-a299b332bf8f\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:49:06 crc kubenswrapper[4821]: I1125 10:49:06.018589 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nntd\" (UniqueName: \"kubernetes.io/projected/16fdcebd-383a-4945-a5bf-a299b332bf8f-kube-api-access-8nntd\") pod \"glance-default-internal-api-0\" (UID: \"16fdcebd-383a-4945-a5bf-a299b332bf8f\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:49:06 crc kubenswrapper[4821]: I1125 10:49:06.018650 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/16fdcebd-383a-4945-a5bf-a299b332bf8f-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"16fdcebd-383a-4945-a5bf-a299b332bf8f\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:49:06 crc kubenswrapper[4821]: I1125 10:49:06.018704 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16fdcebd-383a-4945-a5bf-a299b332bf8f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"16fdcebd-383a-4945-a5bf-a299b332bf8f\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:49:06 crc kubenswrapper[4821]: I1125 10:49:06.120024 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16fdcebd-383a-4945-a5bf-a299b332bf8f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"16fdcebd-383a-4945-a5bf-a299b332bf8f\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:49:06 crc kubenswrapper[4821]: I1125 10:49:06.120106 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16fdcebd-383a-4945-a5bf-a299b332bf8f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"16fdcebd-383a-4945-a5bf-a299b332bf8f\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:49:06 crc kubenswrapper[4821]: I1125 10:49:06.120146 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16fdcebd-383a-4945-a5bf-a299b332bf8f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"16fdcebd-383a-4945-a5bf-a299b332bf8f\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:49:06 crc kubenswrapper[4821]: I1125 10:49:06.120546 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/16fdcebd-383a-4945-a5bf-a299b332bf8f-logs\") pod \"glance-default-internal-api-0\" (UID: \"16fdcebd-383a-4945-a5bf-a299b332bf8f\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:49:06 crc kubenswrapper[4821]: I1125 10:49:06.120657 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/16fdcebd-383a-4945-a5bf-a299b332bf8f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"16fdcebd-383a-4945-a5bf-a299b332bf8f\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:49:06 crc kubenswrapper[4821]: I1125 10:49:06.120683 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"16fdcebd-383a-4945-a5bf-a299b332bf8f\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:49:06 crc kubenswrapper[4821]: I1125 10:49:06.120709 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8nntd\" (UniqueName: \"kubernetes.io/projected/16fdcebd-383a-4945-a5bf-a299b332bf8f-kube-api-access-8nntd\") pod \"glance-default-internal-api-0\" (UID: \"16fdcebd-383a-4945-a5bf-a299b332bf8f\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:49:06 crc kubenswrapper[4821]: I1125 10:49:06.120751 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/16fdcebd-383a-4945-a5bf-a299b332bf8f-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"16fdcebd-383a-4945-a5bf-a299b332bf8f\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:49:06 crc kubenswrapper[4821]: I1125 10:49:06.121127 4821 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"16fdcebd-383a-4945-a5bf-a299b332bf8f\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-internal-api-0" Nov 25 10:49:06 crc kubenswrapper[4821]: I1125 10:49:06.121859 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/16fdcebd-383a-4945-a5bf-a299b332bf8f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"16fdcebd-383a-4945-a5bf-a299b332bf8f\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:49:06 crc kubenswrapper[4821]: I1125 10:49:06.122003 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/16fdcebd-383a-4945-a5bf-a299b332bf8f-logs\") pod \"glance-default-internal-api-0\" (UID: \"16fdcebd-383a-4945-a5bf-a299b332bf8f\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:49:06 crc kubenswrapper[4821]: I1125 10:49:06.125752 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/16fdcebd-383a-4945-a5bf-a299b332bf8f-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"16fdcebd-383a-4945-a5bf-a299b332bf8f\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:49:06 crc kubenswrapper[4821]: I1125 10:49:06.126117 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16fdcebd-383a-4945-a5bf-a299b332bf8f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"16fdcebd-383a-4945-a5bf-a299b332bf8f\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:49:06 crc kubenswrapper[4821]: I1125 10:49:06.126557 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16fdcebd-383a-4945-a5bf-a299b332bf8f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"16fdcebd-383a-4945-a5bf-a299b332bf8f\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:49:06 crc kubenswrapper[4821]: I1125 10:49:06.129421 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf0e0161-6d4e-4149-a1f2-7b96ba1f736b" path="/var/lib/kubelet/pods/bf0e0161-6d4e-4149-a1f2-7b96ba1f736b/volumes" Nov 25 10:49:06 crc kubenswrapper[4821]: I1125 10:49:06.142898 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nntd\" (UniqueName: \"kubernetes.io/projected/16fdcebd-383a-4945-a5bf-a299b332bf8f-kube-api-access-8nntd\") pod \"glance-default-internal-api-0\" (UID: \"16fdcebd-383a-4945-a5bf-a299b332bf8f\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:49:06 crc kubenswrapper[4821]: I1125 10:49:06.143775 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16fdcebd-383a-4945-a5bf-a299b332bf8f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"16fdcebd-383a-4945-a5bf-a299b332bf8f\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:49:06 crc kubenswrapper[4821]: I1125 10:49:06.154504 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"16fdcebd-383a-4945-a5bf-a299b332bf8f\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:49:06 crc kubenswrapper[4821]: I1125 10:49:06.185198 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 25 10:49:07 crc kubenswrapper[4821]: E1125 10:49:07.331773 4821 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-placement-api:current-podified" Nov 25 10:49:07 crc kubenswrapper[4821]: E1125 10:49:07.332221 4821 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:placement-db-sync,Image:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/placement,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:placement-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-g4h77,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42482,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-db-sync-xxdmc_openstack(bee9a405-f60c-449e-beba-71cd802daa79): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 10:49:07 crc kubenswrapper[4821]: E1125 10:49:07.334977 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/placement-db-sync-xxdmc" podUID="bee9a405-f60c-449e-beba-71cd802daa79" Nov 25 10:49:07 crc kubenswrapper[4821]: E1125 10:49:07.771432 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-placement-api:current-podified\\\"\"" pod="openstack/placement-db-sync-xxdmc" podUID="bee9a405-f60c-449e-beba-71cd802daa79" Nov 25 10:49:07 crc kubenswrapper[4821]: I1125 10:49:07.964691 4821 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-74dc88fc-bqbsk" podUID="ccd85fd6-b0a0-4184-8082-054d2de0adec" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.133:5353: connect: connection refused" Nov 25 10:49:09 crc kubenswrapper[4821]: E1125 10:49:09.435016 4821 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified" Nov 25 10:49:09 crc kubenswrapper[4821]: E1125 10:49:09.435938 4821 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n66bh695h686h7h655h644h5fbh8h566h679h58h576hf4h7chd5h88hf5hf9h57h67fhch5b7h58ch584h689h78h5fh574h664h666h55dh574q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-f5cz7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(9cb63eff-3e7e-436e-aea6-ce8254c3db77): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 10:49:09 crc kubenswrapper[4821]: I1125 10:49:09.501646 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2mjz9" Nov 25 10:49:09 crc kubenswrapper[4821]: I1125 10:49:09.581990 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8462fa1-9270-4fa6-ae89-ea51e6246764-scripts\") pod \"a8462fa1-9270-4fa6-ae89-ea51e6246764\" (UID: \"a8462fa1-9270-4fa6-ae89-ea51e6246764\") " Nov 25 10:49:09 crc kubenswrapper[4821]: I1125 10:49:09.582221 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8462fa1-9270-4fa6-ae89-ea51e6246764-config-data\") pod \"a8462fa1-9270-4fa6-ae89-ea51e6246764\" (UID: \"a8462fa1-9270-4fa6-ae89-ea51e6246764\") " Nov 25 10:49:09 crc kubenswrapper[4821]: I1125 10:49:09.582253 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a8462fa1-9270-4fa6-ae89-ea51e6246764-fernet-keys\") pod \"a8462fa1-9270-4fa6-ae89-ea51e6246764\" (UID: \"a8462fa1-9270-4fa6-ae89-ea51e6246764\") " Nov 25 10:49:09 crc kubenswrapper[4821]: I1125 10:49:09.582363 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hxmf5\" (UniqueName: \"kubernetes.io/projected/a8462fa1-9270-4fa6-ae89-ea51e6246764-kube-api-access-hxmf5\") pod \"a8462fa1-9270-4fa6-ae89-ea51e6246764\" (UID: \"a8462fa1-9270-4fa6-ae89-ea51e6246764\") " Nov 25 10:49:09 crc kubenswrapper[4821]: I1125 10:49:09.582408 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a8462fa1-9270-4fa6-ae89-ea51e6246764-credential-keys\") pod \"a8462fa1-9270-4fa6-ae89-ea51e6246764\" (UID: \"a8462fa1-9270-4fa6-ae89-ea51e6246764\") " Nov 25 10:49:09 crc kubenswrapper[4821]: I1125 10:49:09.582498 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8462fa1-9270-4fa6-ae89-ea51e6246764-combined-ca-bundle\") pod \"a8462fa1-9270-4fa6-ae89-ea51e6246764\" (UID: \"a8462fa1-9270-4fa6-ae89-ea51e6246764\") " Nov 25 10:49:09 crc kubenswrapper[4821]: I1125 10:49:09.590085 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8462fa1-9270-4fa6-ae89-ea51e6246764-scripts" (OuterVolumeSpecName: "scripts") pod "a8462fa1-9270-4fa6-ae89-ea51e6246764" (UID: "a8462fa1-9270-4fa6-ae89-ea51e6246764"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:49:09 crc kubenswrapper[4821]: I1125 10:49:09.604006 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8462fa1-9270-4fa6-ae89-ea51e6246764-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "a8462fa1-9270-4fa6-ae89-ea51e6246764" (UID: "a8462fa1-9270-4fa6-ae89-ea51e6246764"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:49:09 crc kubenswrapper[4821]: I1125 10:49:09.604108 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8462fa1-9270-4fa6-ae89-ea51e6246764-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "a8462fa1-9270-4fa6-ae89-ea51e6246764" (UID: "a8462fa1-9270-4fa6-ae89-ea51e6246764"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:49:09 crc kubenswrapper[4821]: I1125 10:49:09.612816 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8462fa1-9270-4fa6-ae89-ea51e6246764-kube-api-access-hxmf5" (OuterVolumeSpecName: "kube-api-access-hxmf5") pod "a8462fa1-9270-4fa6-ae89-ea51e6246764" (UID: "a8462fa1-9270-4fa6-ae89-ea51e6246764"). InnerVolumeSpecName "kube-api-access-hxmf5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:49:09 crc kubenswrapper[4821]: I1125 10:49:09.617844 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8462fa1-9270-4fa6-ae89-ea51e6246764-config-data" (OuterVolumeSpecName: "config-data") pod "a8462fa1-9270-4fa6-ae89-ea51e6246764" (UID: "a8462fa1-9270-4fa6-ae89-ea51e6246764"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:49:09 crc kubenswrapper[4821]: I1125 10:49:09.632767 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8462fa1-9270-4fa6-ae89-ea51e6246764-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a8462fa1-9270-4fa6-ae89-ea51e6246764" (UID: "a8462fa1-9270-4fa6-ae89-ea51e6246764"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:49:09 crc kubenswrapper[4821]: I1125 10:49:09.686680 4821 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8462fa1-9270-4fa6-ae89-ea51e6246764-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:09 crc kubenswrapper[4821]: I1125 10:49:09.687085 4821 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8462fa1-9270-4fa6-ae89-ea51e6246764-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:09 crc kubenswrapper[4821]: I1125 10:49:09.687112 4821 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8462fa1-9270-4fa6-ae89-ea51e6246764-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:09 crc kubenswrapper[4821]: I1125 10:49:09.687120 4821 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a8462fa1-9270-4fa6-ae89-ea51e6246764-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:09 crc kubenswrapper[4821]: I1125 10:49:09.687129 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hxmf5\" (UniqueName: \"kubernetes.io/projected/a8462fa1-9270-4fa6-ae89-ea51e6246764-kube-api-access-hxmf5\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:09 crc kubenswrapper[4821]: I1125 10:49:09.687139 4821 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a8462fa1-9270-4fa6-ae89-ea51e6246764-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:09 crc kubenswrapper[4821]: I1125 10:49:09.786403 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2mjz9" event={"ID":"a8462fa1-9270-4fa6-ae89-ea51e6246764","Type":"ContainerDied","Data":"e82d35ac02d039edd98eca2a377749fec8b2db0dd99f4485e59f34b804cecede"} Nov 25 10:49:09 crc kubenswrapper[4821]: I1125 10:49:09.786455 4821 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e82d35ac02d039edd98eca2a377749fec8b2db0dd99f4485e59f34b804cecede" Nov 25 10:49:09 crc kubenswrapper[4821]: I1125 10:49:09.786534 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2mjz9" Nov 25 10:49:10 crc kubenswrapper[4821]: I1125 10:49:10.586214 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-2mjz9"] Nov 25 10:49:10 crc kubenswrapper[4821]: I1125 10:49:10.594338 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-2mjz9"] Nov 25 10:49:10 crc kubenswrapper[4821]: I1125 10:49:10.684397 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-bkbhq"] Nov 25 10:49:10 crc kubenswrapper[4821]: E1125 10:49:10.685115 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8462fa1-9270-4fa6-ae89-ea51e6246764" containerName="keystone-bootstrap" Nov 25 10:49:10 crc kubenswrapper[4821]: I1125 10:49:10.685134 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8462fa1-9270-4fa6-ae89-ea51e6246764" containerName="keystone-bootstrap" Nov 25 10:49:10 crc kubenswrapper[4821]: I1125 10:49:10.685434 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8462fa1-9270-4fa6-ae89-ea51e6246764" containerName="keystone-bootstrap" Nov 25 10:49:10 crc kubenswrapper[4821]: I1125 10:49:10.686141 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-bkbhq" Nov 25 10:49:10 crc kubenswrapper[4821]: I1125 10:49:10.688797 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 25 10:49:10 crc kubenswrapper[4821]: I1125 10:49:10.688971 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-rzxc4" Nov 25 10:49:10 crc kubenswrapper[4821]: I1125 10:49:10.689045 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 25 10:49:10 crc kubenswrapper[4821]: I1125 10:49:10.689204 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 25 10:49:10 crc kubenswrapper[4821]: I1125 10:49:10.689862 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 25 10:49:10 crc kubenswrapper[4821]: I1125 10:49:10.701206 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-bkbhq"] Nov 25 10:49:10 crc kubenswrapper[4821]: I1125 10:49:10.812508 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5302114a-3896-429e-9693-3483c106730a-scripts\") pod \"keystone-bootstrap-bkbhq\" (UID: \"5302114a-3896-429e-9693-3483c106730a\") " pod="openstack/keystone-bootstrap-bkbhq" Nov 25 10:49:10 crc kubenswrapper[4821]: I1125 10:49:10.812589 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5302114a-3896-429e-9693-3483c106730a-fernet-keys\") pod \"keystone-bootstrap-bkbhq\" (UID: \"5302114a-3896-429e-9693-3483c106730a\") " pod="openstack/keystone-bootstrap-bkbhq" Nov 25 10:49:10 crc kubenswrapper[4821]: I1125 10:49:10.812689 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5302114a-3896-429e-9693-3483c106730a-config-data\") pod \"keystone-bootstrap-bkbhq\" (UID: \"5302114a-3896-429e-9693-3483c106730a\") " pod="openstack/keystone-bootstrap-bkbhq" Nov 25 10:49:10 crc kubenswrapper[4821]: I1125 10:49:10.813038 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5302114a-3896-429e-9693-3483c106730a-combined-ca-bundle\") pod \"keystone-bootstrap-bkbhq\" (UID: \"5302114a-3896-429e-9693-3483c106730a\") " pod="openstack/keystone-bootstrap-bkbhq" Nov 25 10:49:10 crc kubenswrapper[4821]: I1125 10:49:10.813133 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5302114a-3896-429e-9693-3483c106730a-credential-keys\") pod \"keystone-bootstrap-bkbhq\" (UID: \"5302114a-3896-429e-9693-3483c106730a\") " pod="openstack/keystone-bootstrap-bkbhq" Nov 25 10:49:10 crc kubenswrapper[4821]: I1125 10:49:10.813269 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96bqm\" (UniqueName: \"kubernetes.io/projected/5302114a-3896-429e-9693-3483c106730a-kube-api-access-96bqm\") pod \"keystone-bootstrap-bkbhq\" (UID: \"5302114a-3896-429e-9693-3483c106730a\") " pod="openstack/keystone-bootstrap-bkbhq" Nov 25 10:49:10 crc kubenswrapper[4821]: I1125 10:49:10.915072 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5302114a-3896-429e-9693-3483c106730a-fernet-keys\") pod \"keystone-bootstrap-bkbhq\" (UID: \"5302114a-3896-429e-9693-3483c106730a\") " pod="openstack/keystone-bootstrap-bkbhq" Nov 25 10:49:10 crc kubenswrapper[4821]: I1125 10:49:10.915115 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5302114a-3896-429e-9693-3483c106730a-config-data\") pod \"keystone-bootstrap-bkbhq\" (UID: \"5302114a-3896-429e-9693-3483c106730a\") " pod="openstack/keystone-bootstrap-bkbhq" Nov 25 10:49:10 crc kubenswrapper[4821]: I1125 10:49:10.915207 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5302114a-3896-429e-9693-3483c106730a-combined-ca-bundle\") pod \"keystone-bootstrap-bkbhq\" (UID: \"5302114a-3896-429e-9693-3483c106730a\") " pod="openstack/keystone-bootstrap-bkbhq" Nov 25 10:49:10 crc kubenswrapper[4821]: I1125 10:49:10.915234 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5302114a-3896-429e-9693-3483c106730a-credential-keys\") pod \"keystone-bootstrap-bkbhq\" (UID: \"5302114a-3896-429e-9693-3483c106730a\") " pod="openstack/keystone-bootstrap-bkbhq" Nov 25 10:49:10 crc kubenswrapper[4821]: I1125 10:49:10.915931 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96bqm\" (UniqueName: \"kubernetes.io/projected/5302114a-3896-429e-9693-3483c106730a-kube-api-access-96bqm\") pod \"keystone-bootstrap-bkbhq\" (UID: \"5302114a-3896-429e-9693-3483c106730a\") " pod="openstack/keystone-bootstrap-bkbhq" Nov 25 10:49:10 crc kubenswrapper[4821]: I1125 10:49:10.916037 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5302114a-3896-429e-9693-3483c106730a-scripts\") pod \"keystone-bootstrap-bkbhq\" (UID: \"5302114a-3896-429e-9693-3483c106730a\") " pod="openstack/keystone-bootstrap-bkbhq" Nov 25 10:49:10 crc kubenswrapper[4821]: I1125 10:49:10.921377 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5302114a-3896-429e-9693-3483c106730a-scripts\") pod \"keystone-bootstrap-bkbhq\" (UID: \"5302114a-3896-429e-9693-3483c106730a\") " pod="openstack/keystone-bootstrap-bkbhq" Nov 25 10:49:10 crc kubenswrapper[4821]: I1125 10:49:10.921407 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5302114a-3896-429e-9693-3483c106730a-credential-keys\") pod \"keystone-bootstrap-bkbhq\" (UID: \"5302114a-3896-429e-9693-3483c106730a\") " pod="openstack/keystone-bootstrap-bkbhq" Nov 25 10:49:10 crc kubenswrapper[4821]: I1125 10:49:10.926455 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5302114a-3896-429e-9693-3483c106730a-fernet-keys\") pod \"keystone-bootstrap-bkbhq\" (UID: \"5302114a-3896-429e-9693-3483c106730a\") " pod="openstack/keystone-bootstrap-bkbhq" Nov 25 10:49:10 crc kubenswrapper[4821]: I1125 10:49:10.928431 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5302114a-3896-429e-9693-3483c106730a-combined-ca-bundle\") pod \"keystone-bootstrap-bkbhq\" (UID: \"5302114a-3896-429e-9693-3483c106730a\") " pod="openstack/keystone-bootstrap-bkbhq" Nov 25 10:49:10 crc kubenswrapper[4821]: I1125 10:49:10.931373 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96bqm\" (UniqueName: \"kubernetes.io/projected/5302114a-3896-429e-9693-3483c106730a-kube-api-access-96bqm\") pod \"keystone-bootstrap-bkbhq\" (UID: \"5302114a-3896-429e-9693-3483c106730a\") " pod="openstack/keystone-bootstrap-bkbhq" Nov 25 10:49:10 crc kubenswrapper[4821]: I1125 10:49:10.939739 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5302114a-3896-429e-9693-3483c106730a-config-data\") pod \"keystone-bootstrap-bkbhq\" (UID: \"5302114a-3896-429e-9693-3483c106730a\") " pod="openstack/keystone-bootstrap-bkbhq" Nov 25 10:49:11 crc kubenswrapper[4821]: I1125 10:49:11.015617 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-bkbhq" Nov 25 10:49:12 crc kubenswrapper[4821]: I1125 10:49:12.145396 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a8462fa1-9270-4fa6-ae89-ea51e6246764" path="/var/lib/kubelet/pods/a8462fa1-9270-4fa6-ae89-ea51e6246764/volumes" Nov 25 10:49:17 crc kubenswrapper[4821]: I1125 10:49:17.965229 4821 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-74dc88fc-bqbsk" podUID="ccd85fd6-b0a0-4184-8082-054d2de0adec" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.133:5353: i/o timeout" Nov 25 10:49:17 crc kubenswrapper[4821]: I1125 10:49:17.966020 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-74dc88fc-bqbsk" Nov 25 10:49:18 crc kubenswrapper[4821]: E1125 10:49:18.127730 4821 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Nov 25 10:49:18 crc kubenswrapper[4821]: E1125 10:49:18.127869 4821 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-l58nm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-7znwh_openstack(2eb6852b-84a1-4d4a-8402-77a7e2252ab2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 10:49:18 crc kubenswrapper[4821]: E1125 10:49:18.128960 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-7znwh" podUID="2eb6852b-84a1-4d4a-8402-77a7e2252ab2" Nov 25 10:49:18 crc kubenswrapper[4821]: I1125 10:49:18.224033 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 25 10:49:18 crc kubenswrapper[4821]: I1125 10:49:18.230524 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74dc88fc-bqbsk" Nov 25 10:49:18 crc kubenswrapper[4821]: I1125 10:49:18.358094 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tvsg4\" (UniqueName: \"kubernetes.io/projected/f2d41a59-94d9-4440-bfa6-f88e49b97aff-kube-api-access-tvsg4\") pod \"f2d41a59-94d9-4440-bfa6-f88e49b97aff\" (UID: \"f2d41a59-94d9-4440-bfa6-f88e49b97aff\") " Nov 25 10:49:18 crc kubenswrapper[4821]: I1125 10:49:18.358155 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f2d41a59-94d9-4440-bfa6-f88e49b97aff-httpd-run\") pod \"f2d41a59-94d9-4440-bfa6-f88e49b97aff\" (UID: \"f2d41a59-94d9-4440-bfa6-f88e49b97aff\") " Nov 25 10:49:18 crc kubenswrapper[4821]: I1125 10:49:18.358210 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mwzch\" (UniqueName: \"kubernetes.io/projected/ccd85fd6-b0a0-4184-8082-054d2de0adec-kube-api-access-mwzch\") pod \"ccd85fd6-b0a0-4184-8082-054d2de0adec\" (UID: \"ccd85fd6-b0a0-4184-8082-054d2de0adec\") " Nov 25 10:49:18 crc kubenswrapper[4821]: I1125 10:49:18.358267 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2d41a59-94d9-4440-bfa6-f88e49b97aff-combined-ca-bundle\") pod \"f2d41a59-94d9-4440-bfa6-f88e49b97aff\" (UID: \"f2d41a59-94d9-4440-bfa6-f88e49b97aff\") " Nov 25 10:49:18 crc kubenswrapper[4821]: I1125 10:49:18.358331 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ccd85fd6-b0a0-4184-8082-054d2de0adec-config\") pod \"ccd85fd6-b0a0-4184-8082-054d2de0adec\" (UID: \"ccd85fd6-b0a0-4184-8082-054d2de0adec\") " Nov 25 10:49:18 crc kubenswrapper[4821]: I1125 10:49:18.358389 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ccd85fd6-b0a0-4184-8082-054d2de0adec-dns-svc\") pod \"ccd85fd6-b0a0-4184-8082-054d2de0adec\" (UID: \"ccd85fd6-b0a0-4184-8082-054d2de0adec\") " Nov 25 10:49:18 crc kubenswrapper[4821]: I1125 10:49:18.358465 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2d41a59-94d9-4440-bfa6-f88e49b97aff-config-data\") pod \"f2d41a59-94d9-4440-bfa6-f88e49b97aff\" (UID: \"f2d41a59-94d9-4440-bfa6-f88e49b97aff\") " Nov 25 10:49:18 crc kubenswrapper[4821]: I1125 10:49:18.358492 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2d41a59-94d9-4440-bfa6-f88e49b97aff-scripts\") pod \"f2d41a59-94d9-4440-bfa6-f88e49b97aff\" (UID: \"f2d41a59-94d9-4440-bfa6-f88e49b97aff\") " Nov 25 10:49:18 crc kubenswrapper[4821]: I1125 10:49:18.358516 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f2d41a59-94d9-4440-bfa6-f88e49b97aff-logs\") pod \"f2d41a59-94d9-4440-bfa6-f88e49b97aff\" (UID: \"f2d41a59-94d9-4440-bfa6-f88e49b97aff\") " Nov 25 10:49:18 crc kubenswrapper[4821]: I1125 10:49:18.358582 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f2d41a59-94d9-4440-bfa6-f88e49b97aff-public-tls-certs\") pod \"f2d41a59-94d9-4440-bfa6-f88e49b97aff\" (UID: \"f2d41a59-94d9-4440-bfa6-f88e49b97aff\") " Nov 25 10:49:18 crc kubenswrapper[4821]: I1125 10:49:18.358607 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ccd85fd6-b0a0-4184-8082-054d2de0adec-ovsdbserver-nb\") pod \"ccd85fd6-b0a0-4184-8082-054d2de0adec\" (UID: \"ccd85fd6-b0a0-4184-8082-054d2de0adec\") " Nov 25 10:49:18 crc kubenswrapper[4821]: I1125 10:49:18.358633 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"f2d41a59-94d9-4440-bfa6-f88e49b97aff\" (UID: \"f2d41a59-94d9-4440-bfa6-f88e49b97aff\") " Nov 25 10:49:18 crc kubenswrapper[4821]: I1125 10:49:18.358665 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ccd85fd6-b0a0-4184-8082-054d2de0adec-ovsdbserver-sb\") pod \"ccd85fd6-b0a0-4184-8082-054d2de0adec\" (UID: \"ccd85fd6-b0a0-4184-8082-054d2de0adec\") " Nov 25 10:49:18 crc kubenswrapper[4821]: I1125 10:49:18.358792 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f2d41a59-94d9-4440-bfa6-f88e49b97aff-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "f2d41a59-94d9-4440-bfa6-f88e49b97aff" (UID: "f2d41a59-94d9-4440-bfa6-f88e49b97aff"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:49:18 crc kubenswrapper[4821]: I1125 10:49:18.359096 4821 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f2d41a59-94d9-4440-bfa6-f88e49b97aff-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:18 crc kubenswrapper[4821]: I1125 10:49:18.361242 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f2d41a59-94d9-4440-bfa6-f88e49b97aff-logs" (OuterVolumeSpecName: "logs") pod "f2d41a59-94d9-4440-bfa6-f88e49b97aff" (UID: "f2d41a59-94d9-4440-bfa6-f88e49b97aff"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:49:18 crc kubenswrapper[4821]: I1125 10:49:18.364941 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance") pod "f2d41a59-94d9-4440-bfa6-f88e49b97aff" (UID: "f2d41a59-94d9-4440-bfa6-f88e49b97aff"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 10:49:18 crc kubenswrapper[4821]: I1125 10:49:18.378836 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ccd85fd6-b0a0-4184-8082-054d2de0adec-kube-api-access-mwzch" (OuterVolumeSpecName: "kube-api-access-mwzch") pod "ccd85fd6-b0a0-4184-8082-054d2de0adec" (UID: "ccd85fd6-b0a0-4184-8082-054d2de0adec"). InnerVolumeSpecName "kube-api-access-mwzch". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:49:18 crc kubenswrapper[4821]: I1125 10:49:18.382436 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2d41a59-94d9-4440-bfa6-f88e49b97aff-kube-api-access-tvsg4" (OuterVolumeSpecName: "kube-api-access-tvsg4") pod "f2d41a59-94d9-4440-bfa6-f88e49b97aff" (UID: "f2d41a59-94d9-4440-bfa6-f88e49b97aff"). InnerVolumeSpecName "kube-api-access-tvsg4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:49:18 crc kubenswrapper[4821]: I1125 10:49:18.396310 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2d41a59-94d9-4440-bfa6-f88e49b97aff-scripts" (OuterVolumeSpecName: "scripts") pod "f2d41a59-94d9-4440-bfa6-f88e49b97aff" (UID: "f2d41a59-94d9-4440-bfa6-f88e49b97aff"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:49:18 crc kubenswrapper[4821]: I1125 10:49:18.405042 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2d41a59-94d9-4440-bfa6-f88e49b97aff-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f2d41a59-94d9-4440-bfa6-f88e49b97aff" (UID: "f2d41a59-94d9-4440-bfa6-f88e49b97aff"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:49:18 crc kubenswrapper[4821]: I1125 10:49:18.416136 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ccd85fd6-b0a0-4184-8082-054d2de0adec-config" (OuterVolumeSpecName: "config") pod "ccd85fd6-b0a0-4184-8082-054d2de0adec" (UID: "ccd85fd6-b0a0-4184-8082-054d2de0adec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:49:18 crc kubenswrapper[4821]: I1125 10:49:18.421027 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ccd85fd6-b0a0-4184-8082-054d2de0adec-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ccd85fd6-b0a0-4184-8082-054d2de0adec" (UID: "ccd85fd6-b0a0-4184-8082-054d2de0adec"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:49:18 crc kubenswrapper[4821]: I1125 10:49:18.421881 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ccd85fd6-b0a0-4184-8082-054d2de0adec-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ccd85fd6-b0a0-4184-8082-054d2de0adec" (UID: "ccd85fd6-b0a0-4184-8082-054d2de0adec"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:49:18 crc kubenswrapper[4821]: I1125 10:49:18.423968 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2d41a59-94d9-4440-bfa6-f88e49b97aff-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "f2d41a59-94d9-4440-bfa6-f88e49b97aff" (UID: "f2d41a59-94d9-4440-bfa6-f88e49b97aff"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:49:18 crc kubenswrapper[4821]: I1125 10:49:18.425296 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ccd85fd6-b0a0-4184-8082-054d2de0adec-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ccd85fd6-b0a0-4184-8082-054d2de0adec" (UID: "ccd85fd6-b0a0-4184-8082-054d2de0adec"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:49:18 crc kubenswrapper[4821]: I1125 10:49:18.428846 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2d41a59-94d9-4440-bfa6-f88e49b97aff-config-data" (OuterVolumeSpecName: "config-data") pod "f2d41a59-94d9-4440-bfa6-f88e49b97aff" (UID: "f2d41a59-94d9-4440-bfa6-f88e49b97aff"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:49:18 crc kubenswrapper[4821]: I1125 10:49:18.461155 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mwzch\" (UniqueName: \"kubernetes.io/projected/ccd85fd6-b0a0-4184-8082-054d2de0adec-kube-api-access-mwzch\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:18 crc kubenswrapper[4821]: I1125 10:49:18.461210 4821 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2d41a59-94d9-4440-bfa6-f88e49b97aff-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:18 crc kubenswrapper[4821]: I1125 10:49:18.461223 4821 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ccd85fd6-b0a0-4184-8082-054d2de0adec-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:18 crc kubenswrapper[4821]: I1125 10:49:18.461236 4821 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ccd85fd6-b0a0-4184-8082-054d2de0adec-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:18 crc kubenswrapper[4821]: I1125 10:49:18.461246 4821 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2d41a59-94d9-4440-bfa6-f88e49b97aff-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:18 crc kubenswrapper[4821]: I1125 10:49:18.461256 4821 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2d41a59-94d9-4440-bfa6-f88e49b97aff-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:18 crc kubenswrapper[4821]: I1125 10:49:18.461266 4821 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f2d41a59-94d9-4440-bfa6-f88e49b97aff-logs\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:18 crc kubenswrapper[4821]: I1125 10:49:18.461276 4821 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f2d41a59-94d9-4440-bfa6-f88e49b97aff-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:18 crc kubenswrapper[4821]: I1125 10:49:18.461289 4821 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ccd85fd6-b0a0-4184-8082-054d2de0adec-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:18 crc kubenswrapper[4821]: I1125 10:49:18.461337 4821 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Nov 25 10:49:18 crc kubenswrapper[4821]: I1125 10:49:18.461353 4821 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ccd85fd6-b0a0-4184-8082-054d2de0adec-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:18 crc kubenswrapper[4821]: I1125 10:49:18.461365 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tvsg4\" (UniqueName: \"kubernetes.io/projected/f2d41a59-94d9-4440-bfa6-f88e49b97aff-kube-api-access-tvsg4\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:18 crc kubenswrapper[4821]: I1125 10:49:18.485384 4821 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Nov 25 10:49:18 crc kubenswrapper[4821]: I1125 10:49:18.562986 4821 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:18 crc kubenswrapper[4821]: I1125 10:49:18.869755 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f2d41a59-94d9-4440-bfa6-f88e49b97aff","Type":"ContainerDied","Data":"6f1ba2d6f3226e024f3cdd812e1d65127eedddcc9cfa8d63cb7c3899fcafe6d9"} Nov 25 10:49:18 crc kubenswrapper[4821]: I1125 10:49:18.869818 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 25 10:49:18 crc kubenswrapper[4821]: I1125 10:49:18.872856 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74dc88fc-bqbsk" event={"ID":"ccd85fd6-b0a0-4184-8082-054d2de0adec","Type":"ContainerDied","Data":"6531123427f2e88df3a3269538b8d53b4d6f734bd55cebaf7a49d8b5f65a3169"} Nov 25 10:49:18 crc kubenswrapper[4821]: I1125 10:49:18.872883 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74dc88fc-bqbsk" Nov 25 10:49:18 crc kubenswrapper[4821]: E1125 10:49:18.873932 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-7znwh" podUID="2eb6852b-84a1-4d4a-8402-77a7e2252ab2" Nov 25 10:49:18 crc kubenswrapper[4821]: I1125 10:49:18.906608 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 10:49:18 crc kubenswrapper[4821]: I1125 10:49:18.913206 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 10:49:18 crc kubenswrapper[4821]: I1125 10:49:18.940243 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 10:49:18 crc kubenswrapper[4821]: E1125 10:49:18.940718 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2d41a59-94d9-4440-bfa6-f88e49b97aff" containerName="glance-log" Nov 25 10:49:18 crc kubenswrapper[4821]: I1125 10:49:18.940740 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2d41a59-94d9-4440-bfa6-f88e49b97aff" containerName="glance-log" Nov 25 10:49:18 crc kubenswrapper[4821]: E1125 10:49:18.940779 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccd85fd6-b0a0-4184-8082-054d2de0adec" containerName="init" Nov 25 10:49:18 crc kubenswrapper[4821]: I1125 10:49:18.940786 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccd85fd6-b0a0-4184-8082-054d2de0adec" containerName="init" Nov 25 10:49:18 crc kubenswrapper[4821]: E1125 10:49:18.940805 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccd85fd6-b0a0-4184-8082-054d2de0adec" containerName="dnsmasq-dns" Nov 25 10:49:18 crc kubenswrapper[4821]: I1125 10:49:18.940814 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccd85fd6-b0a0-4184-8082-054d2de0adec" containerName="dnsmasq-dns" Nov 25 10:49:18 crc kubenswrapper[4821]: E1125 10:49:18.940827 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2d41a59-94d9-4440-bfa6-f88e49b97aff" containerName="glance-httpd" Nov 25 10:49:18 crc kubenswrapper[4821]: I1125 10:49:18.940834 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2d41a59-94d9-4440-bfa6-f88e49b97aff" containerName="glance-httpd" Nov 25 10:49:18 crc kubenswrapper[4821]: I1125 10:49:18.941068 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccd85fd6-b0a0-4184-8082-054d2de0adec" containerName="dnsmasq-dns" Nov 25 10:49:18 crc kubenswrapper[4821]: I1125 10:49:18.941088 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2d41a59-94d9-4440-bfa6-f88e49b97aff" containerName="glance-log" Nov 25 10:49:18 crc kubenswrapper[4821]: I1125 10:49:18.941105 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2d41a59-94d9-4440-bfa6-f88e49b97aff" containerName="glance-httpd" Nov 25 10:49:18 crc kubenswrapper[4821]: I1125 10:49:18.942347 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 25 10:49:18 crc kubenswrapper[4821]: I1125 10:49:18.945982 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 25 10:49:18 crc kubenswrapper[4821]: I1125 10:49:18.946245 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 25 10:49:18 crc kubenswrapper[4821]: I1125 10:49:18.958634 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74dc88fc-bqbsk"] Nov 25 10:49:18 crc kubenswrapper[4821]: I1125 10:49:18.983194 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-74dc88fc-bqbsk"] Nov 25 10:49:18 crc kubenswrapper[4821]: I1125 10:49:18.993566 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 10:49:19 crc kubenswrapper[4821]: I1125 10:49:19.071015 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e08ffdf-0b33-475b-b040-c0ecda45ff57-scripts\") pod \"glance-default-external-api-0\" (UID: \"2e08ffdf-0b33-475b-b040-c0ecda45ff57\") " pod="openstack/glance-default-external-api-0" Nov 25 10:49:19 crc kubenswrapper[4821]: I1125 10:49:19.071084 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e08ffdf-0b33-475b-b040-c0ecda45ff57-config-data\") pod \"glance-default-external-api-0\" (UID: \"2e08ffdf-0b33-475b-b040-c0ecda45ff57\") " pod="openstack/glance-default-external-api-0" Nov 25 10:49:19 crc kubenswrapper[4821]: I1125 10:49:19.071178 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e08ffdf-0b33-475b-b040-c0ecda45ff57-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"2e08ffdf-0b33-475b-b040-c0ecda45ff57\") " pod="openstack/glance-default-external-api-0" Nov 25 10:49:19 crc kubenswrapper[4821]: I1125 10:49:19.071240 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2e08ffdf-0b33-475b-b040-c0ecda45ff57-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2e08ffdf-0b33-475b-b040-c0ecda45ff57\") " pod="openstack/glance-default-external-api-0" Nov 25 10:49:19 crc kubenswrapper[4821]: I1125 10:49:19.071300 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2p6qs\" (UniqueName: \"kubernetes.io/projected/2e08ffdf-0b33-475b-b040-c0ecda45ff57-kube-api-access-2p6qs\") pod \"glance-default-external-api-0\" (UID: \"2e08ffdf-0b33-475b-b040-c0ecda45ff57\") " pod="openstack/glance-default-external-api-0" Nov 25 10:49:19 crc kubenswrapper[4821]: I1125 10:49:19.071371 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e08ffdf-0b33-475b-b040-c0ecda45ff57-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2e08ffdf-0b33-475b-b040-c0ecda45ff57\") " pod="openstack/glance-default-external-api-0" Nov 25 10:49:19 crc kubenswrapper[4821]: I1125 10:49:19.071391 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e08ffdf-0b33-475b-b040-c0ecda45ff57-logs\") pod \"glance-default-external-api-0\" (UID: \"2e08ffdf-0b33-475b-b040-c0ecda45ff57\") " pod="openstack/glance-default-external-api-0" Nov 25 10:49:19 crc kubenswrapper[4821]: I1125 10:49:19.071418 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"2e08ffdf-0b33-475b-b040-c0ecda45ff57\") " pod="openstack/glance-default-external-api-0" Nov 25 10:49:19 crc kubenswrapper[4821]: I1125 10:49:19.174506 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e08ffdf-0b33-475b-b040-c0ecda45ff57-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"2e08ffdf-0b33-475b-b040-c0ecda45ff57\") " pod="openstack/glance-default-external-api-0" Nov 25 10:49:19 crc kubenswrapper[4821]: I1125 10:49:19.174936 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2e08ffdf-0b33-475b-b040-c0ecda45ff57-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2e08ffdf-0b33-475b-b040-c0ecda45ff57\") " pod="openstack/glance-default-external-api-0" Nov 25 10:49:19 crc kubenswrapper[4821]: I1125 10:49:19.174990 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2p6qs\" (UniqueName: \"kubernetes.io/projected/2e08ffdf-0b33-475b-b040-c0ecda45ff57-kube-api-access-2p6qs\") pod \"glance-default-external-api-0\" (UID: \"2e08ffdf-0b33-475b-b040-c0ecda45ff57\") " pod="openstack/glance-default-external-api-0" Nov 25 10:49:19 crc kubenswrapper[4821]: I1125 10:49:19.175041 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e08ffdf-0b33-475b-b040-c0ecda45ff57-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2e08ffdf-0b33-475b-b040-c0ecda45ff57\") " pod="openstack/glance-default-external-api-0" Nov 25 10:49:19 crc kubenswrapper[4821]: I1125 10:49:19.175059 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e08ffdf-0b33-475b-b040-c0ecda45ff57-logs\") pod \"glance-default-external-api-0\" (UID: \"2e08ffdf-0b33-475b-b040-c0ecda45ff57\") " pod="openstack/glance-default-external-api-0" Nov 25 10:49:19 crc kubenswrapper[4821]: I1125 10:49:19.175112 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"2e08ffdf-0b33-475b-b040-c0ecda45ff57\") " pod="openstack/glance-default-external-api-0" Nov 25 10:49:19 crc kubenswrapper[4821]: I1125 10:49:19.175136 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e08ffdf-0b33-475b-b040-c0ecda45ff57-scripts\") pod \"glance-default-external-api-0\" (UID: \"2e08ffdf-0b33-475b-b040-c0ecda45ff57\") " pod="openstack/glance-default-external-api-0" Nov 25 10:49:19 crc kubenswrapper[4821]: I1125 10:49:19.175210 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e08ffdf-0b33-475b-b040-c0ecda45ff57-config-data\") pod \"glance-default-external-api-0\" (UID: \"2e08ffdf-0b33-475b-b040-c0ecda45ff57\") " pod="openstack/glance-default-external-api-0" Nov 25 10:49:19 crc kubenswrapper[4821]: I1125 10:49:19.175496 4821 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"2e08ffdf-0b33-475b-b040-c0ecda45ff57\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-external-api-0" Nov 25 10:49:19 crc kubenswrapper[4821]: I1125 10:49:19.175786 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2e08ffdf-0b33-475b-b040-c0ecda45ff57-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2e08ffdf-0b33-475b-b040-c0ecda45ff57\") " pod="openstack/glance-default-external-api-0" Nov 25 10:49:19 crc kubenswrapper[4821]: I1125 10:49:19.176721 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e08ffdf-0b33-475b-b040-c0ecda45ff57-logs\") pod \"glance-default-external-api-0\" (UID: \"2e08ffdf-0b33-475b-b040-c0ecda45ff57\") " pod="openstack/glance-default-external-api-0" Nov 25 10:49:19 crc kubenswrapper[4821]: I1125 10:49:19.180788 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e08ffdf-0b33-475b-b040-c0ecda45ff57-config-data\") pod \"glance-default-external-api-0\" (UID: \"2e08ffdf-0b33-475b-b040-c0ecda45ff57\") " pod="openstack/glance-default-external-api-0" Nov 25 10:49:19 crc kubenswrapper[4821]: I1125 10:49:19.180842 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e08ffdf-0b33-475b-b040-c0ecda45ff57-scripts\") pod \"glance-default-external-api-0\" (UID: \"2e08ffdf-0b33-475b-b040-c0ecda45ff57\") " pod="openstack/glance-default-external-api-0" Nov 25 10:49:19 crc kubenswrapper[4821]: I1125 10:49:19.181801 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e08ffdf-0b33-475b-b040-c0ecda45ff57-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"2e08ffdf-0b33-475b-b040-c0ecda45ff57\") " pod="openstack/glance-default-external-api-0" Nov 25 10:49:19 crc kubenswrapper[4821]: I1125 10:49:19.182908 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e08ffdf-0b33-475b-b040-c0ecda45ff57-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2e08ffdf-0b33-475b-b040-c0ecda45ff57\") " pod="openstack/glance-default-external-api-0" Nov 25 10:49:19 crc kubenswrapper[4821]: I1125 10:49:19.197496 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"2e08ffdf-0b33-475b-b040-c0ecda45ff57\") " pod="openstack/glance-default-external-api-0" Nov 25 10:49:19 crc kubenswrapper[4821]: I1125 10:49:19.205038 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2p6qs\" (UniqueName: \"kubernetes.io/projected/2e08ffdf-0b33-475b-b040-c0ecda45ff57-kube-api-access-2p6qs\") pod \"glance-default-external-api-0\" (UID: \"2e08ffdf-0b33-475b-b040-c0ecda45ff57\") " pod="openstack/glance-default-external-api-0" Nov 25 10:49:19 crc kubenswrapper[4821]: I1125 10:49:19.277771 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 25 10:49:19 crc kubenswrapper[4821]: I1125 10:49:19.476878 4821 scope.go:117] "RemoveContainer" containerID="49cd578d7f42537f45fec5bb96e4f5e441fc4f231f7aeb3564608683e64e988e" Nov 25 10:49:19 crc kubenswrapper[4821]: E1125 10:49:19.506610 4821 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Nov 25 10:49:19 crc kubenswrapper[4821]: E1125 10:49:19.506797 4821 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rfcbt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-8bp9n_openstack(eb6c9acf-d69c-4788-a85d-b72135de1ef3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 10:49:19 crc kubenswrapper[4821]: E1125 10:49:19.509670 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-8bp9n" podUID="eb6c9acf-d69c-4788-a85d-b72135de1ef3" Nov 25 10:49:19 crc kubenswrapper[4821]: I1125 10:49:19.887645 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-67f765d856-ncc6t"] Nov 25 10:49:19 crc kubenswrapper[4821]: E1125 10:49:19.975988 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-8bp9n" podUID="eb6c9acf-d69c-4788-a85d-b72135de1ef3" Nov 25 10:49:19 crc kubenswrapper[4821]: I1125 10:49:19.983856 4821 scope.go:117] "RemoveContainer" containerID="68099be94ad7e6d76b625f0ec140658935ff46e4c9a2f0e5b6cbdd31ba331c97" Nov 25 10:49:20 crc kubenswrapper[4821]: I1125 10:49:20.032834 4821 scope.go:117] "RemoveContainer" containerID="a8de04974a2bc9e16f9435e3c9a56a7318997f65d8aa32c8ba43007fc4807cd7" Nov 25 10:49:20 crc kubenswrapper[4821]: I1125 10:49:20.074693 4821 scope.go:117] "RemoveContainer" containerID="67a95be9b3aa84d207a73ce02dd8c799210d3c4070ac21ef95de932bfab350ab" Nov 25 10:49:20 crc kubenswrapper[4821]: I1125 10:49:20.129420 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ccd85fd6-b0a0-4184-8082-054d2de0adec" path="/var/lib/kubelet/pods/ccd85fd6-b0a0-4184-8082-054d2de0adec/volumes" Nov 25 10:49:20 crc kubenswrapper[4821]: I1125 10:49:20.130039 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f2d41a59-94d9-4440-bfa6-f88e49b97aff" path="/var/lib/kubelet/pods/f2d41a59-94d9-4440-bfa6-f88e49b97aff/volumes" Nov 25 10:49:20 crc kubenswrapper[4821]: I1125 10:49:20.148137 4821 scope.go:117] "RemoveContainer" containerID="9be898024013d7da65251768072c3f8f4bbbaa8d591cd6cc751e41d5ee6da326" Nov 25 10:49:20 crc kubenswrapper[4821]: I1125 10:49:20.341436 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-869c6f5bfb-l785w"] Nov 25 10:49:20 crc kubenswrapper[4821]: I1125 10:49:20.426521 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 10:49:20 crc kubenswrapper[4821]: I1125 10:49:20.443946 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-bkbhq"] Nov 25 10:49:20 crc kubenswrapper[4821]: W1125 10:49:20.465046 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod16fdcebd_383a_4945_a5bf_a299b332bf8f.slice/crio-0df33c820c10e12b24d38e8d4319533b7c9fd7ae73003b4ee9209b4d014aa9bd WatchSource:0}: Error finding container 0df33c820c10e12b24d38e8d4319533b7c9fd7ae73003b4ee9209b4d014aa9bd: Status 404 returned error can't find the container with id 0df33c820c10e12b24d38e8d4319533b7c9fd7ae73003b4ee9209b4d014aa9bd Nov 25 10:49:20 crc kubenswrapper[4821]: I1125 10:49:20.600402 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 10:49:20 crc kubenswrapper[4821]: W1125 10:49:20.640943 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2e08ffdf_0b33_475b_b040_c0ecda45ff57.slice/crio-8e7596ffeecbe1d869e74b9e5bd43193776fd6896b8cae38882cda50b189d643 WatchSource:0}: Error finding container 8e7596ffeecbe1d869e74b9e5bd43193776fd6896b8cae38882cda50b189d643: Status 404 returned error can't find the container with id 8e7596ffeecbe1d869e74b9e5bd43193776fd6896b8cae38882cda50b189d643 Nov 25 10:49:20 crc kubenswrapper[4821]: I1125 10:49:20.916534 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"16fdcebd-383a-4945-a5bf-a299b332bf8f","Type":"ContainerStarted","Data":"0df33c820c10e12b24d38e8d4319533b7c9fd7ae73003b4ee9209b4d014aa9bd"} Nov 25 10:49:20 crc kubenswrapper[4821]: I1125 10:49:20.918570 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-64dc894fb9-5xf8b" event={"ID":"f9398a81-673d-46b4-b560-91c6e2885532","Type":"ContainerStarted","Data":"3561021b89555fbb5e74f9d8972eb9d892f00a8f4362cad3f466a085cb92eab4"} Nov 25 10:49:20 crc kubenswrapper[4821]: I1125 10:49:20.919807 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-bkbhq" event={"ID":"5302114a-3896-429e-9693-3483c106730a","Type":"ContainerStarted","Data":"32d18923cd67022f9204c25d3e003a2fc5c3072c6ec6a445326c62f94b4735b7"} Nov 25 10:49:20 crc kubenswrapper[4821]: I1125 10:49:20.921245 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-869c6f5bfb-l785w" event={"ID":"e2327ad3-5d88-4f4b-91ea-91deaf7fc249","Type":"ContainerStarted","Data":"2b47a548244014b3290bfe8c4fd69cccaef99e1db0f5a41a31a5e9fc2a34c105"} Nov 25 10:49:20 crc kubenswrapper[4821]: I1125 10:49:20.922792 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-67f765d856-ncc6t" event={"ID":"93b4abfa-c14a-4df9-a6d1-9cdeaec918dd","Type":"ContainerStarted","Data":"9d40bb93938ec9df40ec41556147c2032da1e1d02519898ce6b61933c05277e3"} Nov 25 10:49:20 crc kubenswrapper[4821]: I1125 10:49:20.922820 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-67f765d856-ncc6t" event={"ID":"93b4abfa-c14a-4df9-a6d1-9cdeaec918dd","Type":"ContainerStarted","Data":"f4c921cc67f3de851ec1c3ae6499596c366218c1d486d8df9d406cdae014f798"} Nov 25 10:49:20 crc kubenswrapper[4821]: I1125 10:49:20.928801 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7558555969-g2z2s" event={"ID":"77e0b5ce-8867-4d7e-b89e-170e67f3a548","Type":"ContainerStarted","Data":"72ace1a8cf21a307bdc5ebcad22a3b55aefccfd5287aca26c9642a08991a80cc"} Nov 25 10:49:20 crc kubenswrapper[4821]: I1125 10:49:20.928874 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7558555969-g2z2s" podUID="77e0b5ce-8867-4d7e-b89e-170e67f3a548" containerName="horizon-log" containerID="cri-o://f3607e3b49353ca4be3512145e78ea3d93f1aaf59d3670587ad46301a2656aa3" gracePeriod=30 Nov 25 10:49:20 crc kubenswrapper[4821]: I1125 10:49:20.929084 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7558555969-g2z2s" event={"ID":"77e0b5ce-8867-4d7e-b89e-170e67f3a548","Type":"ContainerStarted","Data":"f3607e3b49353ca4be3512145e78ea3d93f1aaf59d3670587ad46301a2656aa3"} Nov 25 10:49:20 crc kubenswrapper[4821]: I1125 10:49:20.929132 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7558555969-g2z2s" podUID="77e0b5ce-8867-4d7e-b89e-170e67f3a548" containerName="horizon" containerID="cri-o://72ace1a8cf21a307bdc5ebcad22a3b55aefccfd5287aca26c9642a08991a80cc" gracePeriod=30 Nov 25 10:49:20 crc kubenswrapper[4821]: I1125 10:49:20.933897 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9cb63eff-3e7e-436e-aea6-ce8254c3db77","Type":"ContainerStarted","Data":"cf5c6db7d1dca33b423304d3462893904bd7f34d04492e4bdf0db6099734590c"} Nov 25 10:49:20 crc kubenswrapper[4821]: I1125 10:49:20.935283 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2e08ffdf-0b33-475b-b040-c0ecda45ff57","Type":"ContainerStarted","Data":"8e7596ffeecbe1d869e74b9e5bd43193776fd6896b8cae38882cda50b189d643"} Nov 25 10:49:20 crc kubenswrapper[4821]: I1125 10:49:20.938622 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-574c6969fc-sjxzg" event={"ID":"79a12e89-074b-48f2-bf44-b177fe2e84e2","Type":"ContainerStarted","Data":"704790aed35077ae5429fdc8a0d61310a824faa317eb4be8fe7009dd3293d6f7"} Nov 25 10:49:20 crc kubenswrapper[4821]: I1125 10:49:20.938662 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-574c6969fc-sjxzg" event={"ID":"79a12e89-074b-48f2-bf44-b177fe2e84e2","Type":"ContainerStarted","Data":"4006d528ae6d37112c7f67c5d9d59de50d95a411c30d1ffd8b665efe54524f95"} Nov 25 10:49:20 crc kubenswrapper[4821]: I1125 10:49:20.938788 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-574c6969fc-sjxzg" podUID="79a12e89-074b-48f2-bf44-b177fe2e84e2" containerName="horizon-log" containerID="cri-o://4006d528ae6d37112c7f67c5d9d59de50d95a411c30d1ffd8b665efe54524f95" gracePeriod=30 Nov 25 10:49:20 crc kubenswrapper[4821]: I1125 10:49:20.938828 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-574c6969fc-sjxzg" podUID="79a12e89-074b-48f2-bf44-b177fe2e84e2" containerName="horizon" containerID="cri-o://704790aed35077ae5429fdc8a0d61310a824faa317eb4be8fe7009dd3293d6f7" gracePeriod=30 Nov 25 10:49:20 crc kubenswrapper[4821]: I1125 10:49:20.959850 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7558555969-g2z2s" podStartSLOduration=3.165898395 podStartE2EDuration="29.959827905s" podCreationTimestamp="2025-11-25 10:48:51 +0000 UTC" firstStartedPulling="2025-11-25 10:48:53.054629265 +0000 UTC m=+1003.590949112" lastFinishedPulling="2025-11-25 10:49:19.848558775 +0000 UTC m=+1030.384878622" observedRunningTime="2025-11-25 10:49:20.950872478 +0000 UTC m=+1031.487192335" watchObservedRunningTime="2025-11-25 10:49:20.959827905 +0000 UTC m=+1031.496147752" Nov 25 10:49:20 crc kubenswrapper[4821]: I1125 10:49:20.981048 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-574c6969fc-sjxzg" podStartSLOduration=4.590204449 podStartE2EDuration="30.981023402s" podCreationTimestamp="2025-11-25 10:48:50 +0000 UTC" firstStartedPulling="2025-11-25 10:48:53.062884925 +0000 UTC m=+1003.599204782" lastFinishedPulling="2025-11-25 10:49:19.453703858 +0000 UTC m=+1029.990023735" observedRunningTime="2025-11-25 10:49:20.973310197 +0000 UTC m=+1031.509630054" watchObservedRunningTime="2025-11-25 10:49:20.981023402 +0000 UTC m=+1031.517343249" Nov 25 10:49:21 crc kubenswrapper[4821]: I1125 10:49:21.689143 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-574c6969fc-sjxzg" Nov 25 10:49:21 crc kubenswrapper[4821]: I1125 10:49:21.733426 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7558555969-g2z2s" Nov 25 10:49:21 crc kubenswrapper[4821]: I1125 10:49:21.958474 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-869c6f5bfb-l785w" event={"ID":"e2327ad3-5d88-4f4b-91ea-91deaf7fc249","Type":"ContainerStarted","Data":"0d6baf8077dd2e1ec034e0a15bb7d7957d591590cdfe3a7b244fa1f9081cdfcc"} Nov 25 10:49:21 crc kubenswrapper[4821]: I1125 10:49:21.958516 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-869c6f5bfb-l785w" event={"ID":"e2327ad3-5d88-4f4b-91ea-91deaf7fc249","Type":"ContainerStarted","Data":"6e8007ccc7348305c4f9b913d2dfa8dbf30b5f4417db7e25937dc070c2fa5264"} Nov 25 10:49:21 crc kubenswrapper[4821]: I1125 10:49:21.963398 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"16fdcebd-383a-4945-a5bf-a299b332bf8f","Type":"ContainerStarted","Data":"969ba4671a57151d49576d78ea3b5bfd6a8af49264c276fc5f8b40805e909f57"} Nov 25 10:49:21 crc kubenswrapper[4821]: I1125 10:49:21.965154 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2e08ffdf-0b33-475b-b040-c0ecda45ff57","Type":"ContainerStarted","Data":"ea6716a5af0656e1213bff87beb5471adc668c54ebd32c9e0f829ff8a91fdde6"} Nov 25 10:49:21 crc kubenswrapper[4821]: I1125 10:49:21.968281 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-64dc894fb9-5xf8b" event={"ID":"f9398a81-673d-46b4-b560-91c6e2885532","Type":"ContainerStarted","Data":"71773c54cfbe03e75458bf24e2a51cb38377a4ced5a4d0293161a4ebc1ea7207"} Nov 25 10:49:21 crc kubenswrapper[4821]: I1125 10:49:21.968411 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-64dc894fb9-5xf8b" podUID="f9398a81-673d-46b4-b560-91c6e2885532" containerName="horizon-log" containerID="cri-o://3561021b89555fbb5e74f9d8972eb9d892f00a8f4362cad3f466a085cb92eab4" gracePeriod=30 Nov 25 10:49:21 crc kubenswrapper[4821]: I1125 10:49:21.968511 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-64dc894fb9-5xf8b" podUID="f9398a81-673d-46b4-b560-91c6e2885532" containerName="horizon" containerID="cri-o://71773c54cfbe03e75458bf24e2a51cb38377a4ced5a4d0293161a4ebc1ea7207" gracePeriod=30 Nov 25 10:49:21 crc kubenswrapper[4821]: I1125 10:49:21.970365 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-bkbhq" event={"ID":"5302114a-3896-429e-9693-3483c106730a","Type":"ContainerStarted","Data":"391a9970a4c25e3326a56a91137910d12c345d3bc1f2e0e88b931ee47b675f82"} Nov 25 10:49:21 crc kubenswrapper[4821]: I1125 10:49:21.975254 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-67f765d856-ncc6t" event={"ID":"93b4abfa-c14a-4df9-a6d1-9cdeaec918dd","Type":"ContainerStarted","Data":"f3bd20d4dfc4b36ebce64cd7109802fde8245c576cc60107f7980dcae486774e"} Nov 25 10:49:21 crc kubenswrapper[4821]: I1125 10:49:21.980585 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-869c6f5bfb-l785w" podStartSLOduration=21.980564558 podStartE2EDuration="21.980564558s" podCreationTimestamp="2025-11-25 10:49:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:49:21.975586262 +0000 UTC m=+1032.511906109" watchObservedRunningTime="2025-11-25 10:49:21.980564558 +0000 UTC m=+1032.516884405" Nov 25 10:49:21 crc kubenswrapper[4821]: I1125 10:49:21.996650 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-64dc894fb9-5xf8b" podStartSLOduration=3.986243076 podStartE2EDuration="27.996626245s" podCreationTimestamp="2025-11-25 10:48:54 +0000 UTC" firstStartedPulling="2025-11-25 10:48:55.838006842 +0000 UTC m=+1006.374326689" lastFinishedPulling="2025-11-25 10:49:19.848390011 +0000 UTC m=+1030.384709858" observedRunningTime="2025-11-25 10:49:21.993476896 +0000 UTC m=+1032.529796743" watchObservedRunningTime="2025-11-25 10:49:21.996626245 +0000 UTC m=+1032.532946092" Nov 25 10:49:22 crc kubenswrapper[4821]: I1125 10:49:22.014514 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-bkbhq" podStartSLOduration=12.014497829 podStartE2EDuration="12.014497829s" podCreationTimestamp="2025-11-25 10:49:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:49:22.011570015 +0000 UTC m=+1032.547889862" watchObservedRunningTime="2025-11-25 10:49:22.014497829 +0000 UTC m=+1032.550817686" Nov 25 10:49:22 crc kubenswrapper[4821]: I1125 10:49:22.044858 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-67f765d856-ncc6t" podStartSLOduration=22.044838829 podStartE2EDuration="22.044838829s" podCreationTimestamp="2025-11-25 10:49:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:49:22.043494425 +0000 UTC m=+1032.579814272" watchObservedRunningTime="2025-11-25 10:49:22.044838829 +0000 UTC m=+1032.581158676" Nov 25 10:49:22 crc kubenswrapper[4821]: I1125 10:49:22.966739 4821 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-74dc88fc-bqbsk" podUID="ccd85fd6-b0a0-4184-8082-054d2de0adec" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.133:5353: i/o timeout" Nov 25 10:49:22 crc kubenswrapper[4821]: I1125 10:49:22.995905 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"16fdcebd-383a-4945-a5bf-a299b332bf8f","Type":"ContainerStarted","Data":"9c59a92b8bdcb2afaeb56232da0557ea8d992d2a693fb7bc91ebdcbb5d6e0940"} Nov 25 10:49:23 crc kubenswrapper[4821]: I1125 10:49:23.000575 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2e08ffdf-0b33-475b-b040-c0ecda45ff57","Type":"ContainerStarted","Data":"583283375d3838767a1436fa879285183d56c9e61af3523ef043f60584199559"} Nov 25 10:49:23 crc kubenswrapper[4821]: I1125 10:49:23.037773 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=18.037748046 podStartE2EDuration="18.037748046s" podCreationTimestamp="2025-11-25 10:49:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:49:23.019816031 +0000 UTC m=+1033.556135898" watchObservedRunningTime="2025-11-25 10:49:23.037748046 +0000 UTC m=+1033.574067893" Nov 25 10:49:23 crc kubenswrapper[4821]: I1125 10:49:23.047147 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.047126974 podStartE2EDuration="5.047126974s" podCreationTimestamp="2025-11-25 10:49:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:49:23.043444251 +0000 UTC m=+1033.579764108" watchObservedRunningTime="2025-11-25 10:49:23.047126974 +0000 UTC m=+1033.583446831" Nov 25 10:49:25 crc kubenswrapper[4821]: I1125 10:49:25.280006 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-64dc894fb9-5xf8b" Nov 25 10:49:26 crc kubenswrapper[4821]: I1125 10:49:26.032056 4821 generic.go:334] "Generic (PLEG): container finished" podID="5302114a-3896-429e-9693-3483c106730a" containerID="391a9970a4c25e3326a56a91137910d12c345d3bc1f2e0e88b931ee47b675f82" exitCode=0 Nov 25 10:49:26 crc kubenswrapper[4821]: I1125 10:49:26.032103 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-bkbhq" event={"ID":"5302114a-3896-429e-9693-3483c106730a","Type":"ContainerDied","Data":"391a9970a4c25e3326a56a91137910d12c345d3bc1f2e0e88b931ee47b675f82"} Nov 25 10:49:26 crc kubenswrapper[4821]: I1125 10:49:26.185662 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 25 10:49:26 crc kubenswrapper[4821]: I1125 10:49:26.185721 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 25 10:49:26 crc kubenswrapper[4821]: I1125 10:49:26.220702 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 25 10:49:26 crc kubenswrapper[4821]: I1125 10:49:26.231739 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 25 10:49:27 crc kubenswrapper[4821]: I1125 10:49:27.041396 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 25 10:49:27 crc kubenswrapper[4821]: I1125 10:49:27.041439 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 25 10:49:29 crc kubenswrapper[4821]: I1125 10:49:29.035207 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-bkbhq" Nov 25 10:49:29 crc kubenswrapper[4821]: I1125 10:49:29.067425 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-bkbhq" event={"ID":"5302114a-3896-429e-9693-3483c106730a","Type":"ContainerDied","Data":"32d18923cd67022f9204c25d3e003a2fc5c3072c6ec6a445326c62f94b4735b7"} Nov 25 10:49:29 crc kubenswrapper[4821]: I1125 10:49:29.067468 4821 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="32d18923cd67022f9204c25d3e003a2fc5c3072c6ec6a445326c62f94b4735b7" Nov 25 10:49:29 crc kubenswrapper[4821]: I1125 10:49:29.067526 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-bkbhq" Nov 25 10:49:29 crc kubenswrapper[4821]: I1125 10:49:29.112896 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-96bqm\" (UniqueName: \"kubernetes.io/projected/5302114a-3896-429e-9693-3483c106730a-kube-api-access-96bqm\") pod \"5302114a-3896-429e-9693-3483c106730a\" (UID: \"5302114a-3896-429e-9693-3483c106730a\") " Nov 25 10:49:29 crc kubenswrapper[4821]: I1125 10:49:29.112960 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5302114a-3896-429e-9693-3483c106730a-scripts\") pod \"5302114a-3896-429e-9693-3483c106730a\" (UID: \"5302114a-3896-429e-9693-3483c106730a\") " Nov 25 10:49:29 crc kubenswrapper[4821]: I1125 10:49:29.113003 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5302114a-3896-429e-9693-3483c106730a-credential-keys\") pod \"5302114a-3896-429e-9693-3483c106730a\" (UID: \"5302114a-3896-429e-9693-3483c106730a\") " Nov 25 10:49:29 crc kubenswrapper[4821]: I1125 10:49:29.113189 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5302114a-3896-429e-9693-3483c106730a-config-data\") pod \"5302114a-3896-429e-9693-3483c106730a\" (UID: \"5302114a-3896-429e-9693-3483c106730a\") " Nov 25 10:49:29 crc kubenswrapper[4821]: I1125 10:49:29.113295 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5302114a-3896-429e-9693-3483c106730a-fernet-keys\") pod \"5302114a-3896-429e-9693-3483c106730a\" (UID: \"5302114a-3896-429e-9693-3483c106730a\") " Nov 25 10:49:29 crc kubenswrapper[4821]: I1125 10:49:29.113363 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5302114a-3896-429e-9693-3483c106730a-combined-ca-bundle\") pod \"5302114a-3896-429e-9693-3483c106730a\" (UID: \"5302114a-3896-429e-9693-3483c106730a\") " Nov 25 10:49:29 crc kubenswrapper[4821]: I1125 10:49:29.118153 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5302114a-3896-429e-9693-3483c106730a-kube-api-access-96bqm" (OuterVolumeSpecName: "kube-api-access-96bqm") pod "5302114a-3896-429e-9693-3483c106730a" (UID: "5302114a-3896-429e-9693-3483c106730a"). InnerVolumeSpecName "kube-api-access-96bqm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:49:29 crc kubenswrapper[4821]: I1125 10:49:29.120793 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5302114a-3896-429e-9693-3483c106730a-scripts" (OuterVolumeSpecName: "scripts") pod "5302114a-3896-429e-9693-3483c106730a" (UID: "5302114a-3896-429e-9693-3483c106730a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:49:29 crc kubenswrapper[4821]: I1125 10:49:29.135331 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5302114a-3896-429e-9693-3483c106730a-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "5302114a-3896-429e-9693-3483c106730a" (UID: "5302114a-3896-429e-9693-3483c106730a"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:49:29 crc kubenswrapper[4821]: I1125 10:49:29.151394 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5302114a-3896-429e-9693-3483c106730a-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "5302114a-3896-429e-9693-3483c106730a" (UID: "5302114a-3896-429e-9693-3483c106730a"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:49:29 crc kubenswrapper[4821]: I1125 10:49:29.168295 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5302114a-3896-429e-9693-3483c106730a-config-data" (OuterVolumeSpecName: "config-data") pod "5302114a-3896-429e-9693-3483c106730a" (UID: "5302114a-3896-429e-9693-3483c106730a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:49:29 crc kubenswrapper[4821]: I1125 10:49:29.182675 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5302114a-3896-429e-9693-3483c106730a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5302114a-3896-429e-9693-3483c106730a" (UID: "5302114a-3896-429e-9693-3483c106730a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:49:29 crc kubenswrapper[4821]: I1125 10:49:29.217869 4821 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5302114a-3896-429e-9693-3483c106730a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:29 crc kubenswrapper[4821]: I1125 10:49:29.217946 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-96bqm\" (UniqueName: \"kubernetes.io/projected/5302114a-3896-429e-9693-3483c106730a-kube-api-access-96bqm\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:29 crc kubenswrapper[4821]: I1125 10:49:29.217964 4821 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5302114a-3896-429e-9693-3483c106730a-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:29 crc kubenswrapper[4821]: I1125 10:49:29.217979 4821 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5302114a-3896-429e-9693-3483c106730a-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:29 crc kubenswrapper[4821]: I1125 10:49:29.217990 4821 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5302114a-3896-429e-9693-3483c106730a-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:29 crc kubenswrapper[4821]: I1125 10:49:29.218003 4821 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5302114a-3896-429e-9693-3483c106730a-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:29 crc kubenswrapper[4821]: I1125 10:49:29.278220 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 25 10:49:29 crc kubenswrapper[4821]: I1125 10:49:29.278524 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 25 10:49:29 crc kubenswrapper[4821]: I1125 10:49:29.329496 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 25 10:49:29 crc kubenswrapper[4821]: I1125 10:49:29.354655 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 25 10:49:29 crc kubenswrapper[4821]: I1125 10:49:29.535512 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 25 10:49:29 crc kubenswrapper[4821]: I1125 10:49:29.535627 4821 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 10:49:29 crc kubenswrapper[4821]: I1125 10:49:29.753207 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 25 10:49:30 crc kubenswrapper[4821]: I1125 10:49:30.082231 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-xxdmc" event={"ID":"bee9a405-f60c-449e-beba-71cd802daa79","Type":"ContainerStarted","Data":"789f7caa916d8d2c62126a6a258d74071bb67e90d27c88e23ea85583c8c12ef0"} Nov 25 10:49:30 crc kubenswrapper[4821]: I1125 10:49:30.084454 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 25 10:49:30 crc kubenswrapper[4821]: I1125 10:49:30.084483 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 25 10:49:30 crc kubenswrapper[4821]: I1125 10:49:30.134046 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-xxdmc" podStartSLOduration=3.313026451 podStartE2EDuration="39.134019694s" podCreationTimestamp="2025-11-25 10:48:51 +0000 UTC" firstStartedPulling="2025-11-25 10:48:53.057281113 +0000 UTC m=+1003.593600960" lastFinishedPulling="2025-11-25 10:49:28.878274356 +0000 UTC m=+1039.414594203" observedRunningTime="2025-11-25 10:49:30.09874953 +0000 UTC m=+1040.635069377" watchObservedRunningTime="2025-11-25 10:49:30.134019694 +0000 UTC m=+1040.670339551" Nov 25 10:49:30 crc kubenswrapper[4821]: I1125 10:49:30.179686 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-5fb467464c-hlrl9"] Nov 25 10:49:30 crc kubenswrapper[4821]: E1125 10:49:30.180397 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5302114a-3896-429e-9693-3483c106730a" containerName="keystone-bootstrap" Nov 25 10:49:30 crc kubenswrapper[4821]: I1125 10:49:30.180412 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="5302114a-3896-429e-9693-3483c106730a" containerName="keystone-bootstrap" Nov 25 10:49:30 crc kubenswrapper[4821]: I1125 10:49:30.180602 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="5302114a-3896-429e-9693-3483c106730a" containerName="keystone-bootstrap" Nov 25 10:49:30 crc kubenswrapper[4821]: I1125 10:49:30.181155 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5fb467464c-hlrl9" Nov 25 10:49:30 crc kubenswrapper[4821]: I1125 10:49:30.201353 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5fb467464c-hlrl9"] Nov 25 10:49:30 crc kubenswrapper[4821]: I1125 10:49:30.213536 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 25 10:49:30 crc kubenswrapper[4821]: I1125 10:49:30.213895 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 25 10:49:30 crc kubenswrapper[4821]: I1125 10:49:30.214016 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-rzxc4" Nov 25 10:49:30 crc kubenswrapper[4821]: I1125 10:49:30.214121 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 25 10:49:30 crc kubenswrapper[4821]: I1125 10:49:30.214273 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Nov 25 10:49:30 crc kubenswrapper[4821]: I1125 10:49:30.214383 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Nov 25 10:49:30 crc kubenswrapper[4821]: I1125 10:49:30.460249 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/eabb0cb5-e9c3-474f-80ba-df6ecf373c3f-fernet-keys\") pod \"keystone-5fb467464c-hlrl9\" (UID: \"eabb0cb5-e9c3-474f-80ba-df6ecf373c3f\") " pod="openstack/keystone-5fb467464c-hlrl9" Nov 25 10:49:30 crc kubenswrapper[4821]: I1125 10:49:30.460288 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eabb0cb5-e9c3-474f-80ba-df6ecf373c3f-combined-ca-bundle\") pod \"keystone-5fb467464c-hlrl9\" (UID: \"eabb0cb5-e9c3-474f-80ba-df6ecf373c3f\") " pod="openstack/keystone-5fb467464c-hlrl9" Nov 25 10:49:30 crc kubenswrapper[4821]: I1125 10:49:30.460340 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8frg8\" (UniqueName: \"kubernetes.io/projected/eabb0cb5-e9c3-474f-80ba-df6ecf373c3f-kube-api-access-8frg8\") pod \"keystone-5fb467464c-hlrl9\" (UID: \"eabb0cb5-e9c3-474f-80ba-df6ecf373c3f\") " pod="openstack/keystone-5fb467464c-hlrl9" Nov 25 10:49:30 crc kubenswrapper[4821]: I1125 10:49:30.460393 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/eabb0cb5-e9c3-474f-80ba-df6ecf373c3f-public-tls-certs\") pod \"keystone-5fb467464c-hlrl9\" (UID: \"eabb0cb5-e9c3-474f-80ba-df6ecf373c3f\") " pod="openstack/keystone-5fb467464c-hlrl9" Nov 25 10:49:30 crc kubenswrapper[4821]: I1125 10:49:30.460426 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eabb0cb5-e9c3-474f-80ba-df6ecf373c3f-scripts\") pod \"keystone-5fb467464c-hlrl9\" (UID: \"eabb0cb5-e9c3-474f-80ba-df6ecf373c3f\") " pod="openstack/keystone-5fb467464c-hlrl9" Nov 25 10:49:30 crc kubenswrapper[4821]: I1125 10:49:30.460464 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eabb0cb5-e9c3-474f-80ba-df6ecf373c3f-config-data\") pod \"keystone-5fb467464c-hlrl9\" (UID: \"eabb0cb5-e9c3-474f-80ba-df6ecf373c3f\") " pod="openstack/keystone-5fb467464c-hlrl9" Nov 25 10:49:30 crc kubenswrapper[4821]: I1125 10:49:30.460504 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/eabb0cb5-e9c3-474f-80ba-df6ecf373c3f-internal-tls-certs\") pod \"keystone-5fb467464c-hlrl9\" (UID: \"eabb0cb5-e9c3-474f-80ba-df6ecf373c3f\") " pod="openstack/keystone-5fb467464c-hlrl9" Nov 25 10:49:30 crc kubenswrapper[4821]: I1125 10:49:30.460527 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/eabb0cb5-e9c3-474f-80ba-df6ecf373c3f-credential-keys\") pod \"keystone-5fb467464c-hlrl9\" (UID: \"eabb0cb5-e9c3-474f-80ba-df6ecf373c3f\") " pod="openstack/keystone-5fb467464c-hlrl9" Nov 25 10:49:30 crc kubenswrapper[4821]: I1125 10:49:30.579274 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8frg8\" (UniqueName: \"kubernetes.io/projected/eabb0cb5-e9c3-474f-80ba-df6ecf373c3f-kube-api-access-8frg8\") pod \"keystone-5fb467464c-hlrl9\" (UID: \"eabb0cb5-e9c3-474f-80ba-df6ecf373c3f\") " pod="openstack/keystone-5fb467464c-hlrl9" Nov 25 10:49:30 crc kubenswrapper[4821]: I1125 10:49:30.579605 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/eabb0cb5-e9c3-474f-80ba-df6ecf373c3f-public-tls-certs\") pod \"keystone-5fb467464c-hlrl9\" (UID: \"eabb0cb5-e9c3-474f-80ba-df6ecf373c3f\") " pod="openstack/keystone-5fb467464c-hlrl9" Nov 25 10:49:30 crc kubenswrapper[4821]: I1125 10:49:30.579758 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eabb0cb5-e9c3-474f-80ba-df6ecf373c3f-scripts\") pod \"keystone-5fb467464c-hlrl9\" (UID: \"eabb0cb5-e9c3-474f-80ba-df6ecf373c3f\") " pod="openstack/keystone-5fb467464c-hlrl9" Nov 25 10:49:30 crc kubenswrapper[4821]: I1125 10:49:30.579888 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eabb0cb5-e9c3-474f-80ba-df6ecf373c3f-config-data\") pod \"keystone-5fb467464c-hlrl9\" (UID: \"eabb0cb5-e9c3-474f-80ba-df6ecf373c3f\") " pod="openstack/keystone-5fb467464c-hlrl9" Nov 25 10:49:30 crc kubenswrapper[4821]: I1125 10:49:30.580037 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/eabb0cb5-e9c3-474f-80ba-df6ecf373c3f-internal-tls-certs\") pod \"keystone-5fb467464c-hlrl9\" (UID: \"eabb0cb5-e9c3-474f-80ba-df6ecf373c3f\") " pod="openstack/keystone-5fb467464c-hlrl9" Nov 25 10:49:30 crc kubenswrapper[4821]: I1125 10:49:30.580150 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/eabb0cb5-e9c3-474f-80ba-df6ecf373c3f-credential-keys\") pod \"keystone-5fb467464c-hlrl9\" (UID: \"eabb0cb5-e9c3-474f-80ba-df6ecf373c3f\") " pod="openstack/keystone-5fb467464c-hlrl9" Nov 25 10:49:30 crc kubenswrapper[4821]: I1125 10:49:30.580313 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/eabb0cb5-e9c3-474f-80ba-df6ecf373c3f-fernet-keys\") pod \"keystone-5fb467464c-hlrl9\" (UID: \"eabb0cb5-e9c3-474f-80ba-df6ecf373c3f\") " pod="openstack/keystone-5fb467464c-hlrl9" Nov 25 10:49:30 crc kubenswrapper[4821]: I1125 10:49:30.580416 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eabb0cb5-e9c3-474f-80ba-df6ecf373c3f-combined-ca-bundle\") pod \"keystone-5fb467464c-hlrl9\" (UID: \"eabb0cb5-e9c3-474f-80ba-df6ecf373c3f\") " pod="openstack/keystone-5fb467464c-hlrl9" Nov 25 10:49:30 crc kubenswrapper[4821]: I1125 10:49:30.592316 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/eabb0cb5-e9c3-474f-80ba-df6ecf373c3f-credential-keys\") pod \"keystone-5fb467464c-hlrl9\" (UID: \"eabb0cb5-e9c3-474f-80ba-df6ecf373c3f\") " pod="openstack/keystone-5fb467464c-hlrl9" Nov 25 10:49:30 crc kubenswrapper[4821]: I1125 10:49:30.592397 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eabb0cb5-e9c3-474f-80ba-df6ecf373c3f-config-data\") pod \"keystone-5fb467464c-hlrl9\" (UID: \"eabb0cb5-e9c3-474f-80ba-df6ecf373c3f\") " pod="openstack/keystone-5fb467464c-hlrl9" Nov 25 10:49:30 crc kubenswrapper[4821]: I1125 10:49:30.593034 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eabb0cb5-e9c3-474f-80ba-df6ecf373c3f-combined-ca-bundle\") pod \"keystone-5fb467464c-hlrl9\" (UID: \"eabb0cb5-e9c3-474f-80ba-df6ecf373c3f\") " pod="openstack/keystone-5fb467464c-hlrl9" Nov 25 10:49:30 crc kubenswrapper[4821]: I1125 10:49:30.593301 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/eabb0cb5-e9c3-474f-80ba-df6ecf373c3f-internal-tls-certs\") pod \"keystone-5fb467464c-hlrl9\" (UID: \"eabb0cb5-e9c3-474f-80ba-df6ecf373c3f\") " pod="openstack/keystone-5fb467464c-hlrl9" Nov 25 10:49:30 crc kubenswrapper[4821]: I1125 10:49:30.596536 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/eabb0cb5-e9c3-474f-80ba-df6ecf373c3f-fernet-keys\") pod \"keystone-5fb467464c-hlrl9\" (UID: \"eabb0cb5-e9c3-474f-80ba-df6ecf373c3f\") " pod="openstack/keystone-5fb467464c-hlrl9" Nov 25 10:49:30 crc kubenswrapper[4821]: I1125 10:49:30.597641 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eabb0cb5-e9c3-474f-80ba-df6ecf373c3f-scripts\") pod \"keystone-5fb467464c-hlrl9\" (UID: \"eabb0cb5-e9c3-474f-80ba-df6ecf373c3f\") " pod="openstack/keystone-5fb467464c-hlrl9" Nov 25 10:49:30 crc kubenswrapper[4821]: I1125 10:49:30.602798 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/eabb0cb5-e9c3-474f-80ba-df6ecf373c3f-public-tls-certs\") pod \"keystone-5fb467464c-hlrl9\" (UID: \"eabb0cb5-e9c3-474f-80ba-df6ecf373c3f\") " pod="openstack/keystone-5fb467464c-hlrl9" Nov 25 10:49:30 crc kubenswrapper[4821]: I1125 10:49:30.651869 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8frg8\" (UniqueName: \"kubernetes.io/projected/eabb0cb5-e9c3-474f-80ba-df6ecf373c3f-kube-api-access-8frg8\") pod \"keystone-5fb467464c-hlrl9\" (UID: \"eabb0cb5-e9c3-474f-80ba-df6ecf373c3f\") " pod="openstack/keystone-5fb467464c-hlrl9" Nov 25 10:49:30 crc kubenswrapper[4821]: I1125 10:49:30.939797 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5fb467464c-hlrl9" Nov 25 10:49:30 crc kubenswrapper[4821]: I1125 10:49:30.992700 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-869c6f5bfb-l785w" Nov 25 10:49:30 crc kubenswrapper[4821]: I1125 10:49:30.992755 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-869c6f5bfb-l785w" Nov 25 10:49:31 crc kubenswrapper[4821]: I1125 10:49:31.067604 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-67f765d856-ncc6t" Nov 25 10:49:31 crc kubenswrapper[4821]: I1125 10:49:31.068057 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-67f765d856-ncc6t" Nov 25 10:49:31 crc kubenswrapper[4821]: I1125 10:49:31.087364 4821 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-67f765d856-ncc6t" podUID="93b4abfa-c14a-4df9-a6d1-9cdeaec918dd" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.149:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.149:8443: connect: connection refused" Nov 25 10:49:31 crc kubenswrapper[4821]: I1125 10:49:31.156624 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9cb63eff-3e7e-436e-aea6-ce8254c3db77","Type":"ContainerStarted","Data":"666f50101704e276ddca6f07195913aeb8365df95b6639c5717f6c54b431976e"} Nov 25 10:49:31 crc kubenswrapper[4821]: I1125 10:49:31.430514 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5fb467464c-hlrl9"] Nov 25 10:49:32 crc kubenswrapper[4821]: I1125 10:49:32.168799 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5fb467464c-hlrl9" event={"ID":"eabb0cb5-e9c3-474f-80ba-df6ecf373c3f","Type":"ContainerStarted","Data":"fdaa88f75c120dbf88a02325a89b37b0483209beb827290d6dea437ac4318af3"} Nov 25 10:49:32 crc kubenswrapper[4821]: I1125 10:49:32.168853 4821 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 10:49:32 crc kubenswrapper[4821]: I1125 10:49:32.169190 4821 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 10:49:32 crc kubenswrapper[4821]: I1125 10:49:32.767781 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 25 10:49:33 crc kubenswrapper[4821]: I1125 10:49:33.025607 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 25 10:49:33 crc kubenswrapper[4821]: I1125 10:49:33.199884 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5fb467464c-hlrl9" event={"ID":"eabb0cb5-e9c3-474f-80ba-df6ecf373c3f","Type":"ContainerStarted","Data":"d4ee3ad06f84b52bf987b55356cf0594e9872ddaf7c7c79546d15b101cb80dec"} Nov 25 10:49:33 crc kubenswrapper[4821]: I1125 10:49:33.200310 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-5fb467464c-hlrl9" Nov 25 10:49:33 crc kubenswrapper[4821]: I1125 10:49:33.241459 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-5fb467464c-hlrl9" podStartSLOduration=3.241439825 podStartE2EDuration="3.241439825s" podCreationTimestamp="2025-11-25 10:49:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:49:33.236698132 +0000 UTC m=+1043.773017979" watchObservedRunningTime="2025-11-25 10:49:33.241439825 +0000 UTC m=+1043.777759662" Nov 25 10:49:34 crc kubenswrapper[4821]: I1125 10:49:34.211724 4821 generic.go:334] "Generic (PLEG): container finished" podID="325653d2-264c-4c6c-851f-6e38fe3e61a5" containerID="c9151e75ad24f3fd80b6828468427f0883ab2f705ae36476b0fd3c2fa4619697" exitCode=0 Nov 25 10:49:34 crc kubenswrapper[4821]: I1125 10:49:34.212349 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-tq9mv" event={"ID":"325653d2-264c-4c6c-851f-6e38fe3e61a5","Type":"ContainerDied","Data":"c9151e75ad24f3fd80b6828468427f0883ab2f705ae36476b0fd3c2fa4619697"} Nov 25 10:49:35 crc kubenswrapper[4821]: I1125 10:49:35.224896 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-7znwh" event={"ID":"2eb6852b-84a1-4d4a-8402-77a7e2252ab2","Type":"ContainerStarted","Data":"8f3e63e3b217fdcee82a9e2153ed5efa0a9ca89caff4b8c351478c6944df32c4"} Nov 25 10:49:35 crc kubenswrapper[4821]: I1125 10:49:35.228610 4821 generic.go:334] "Generic (PLEG): container finished" podID="bee9a405-f60c-449e-beba-71cd802daa79" containerID="789f7caa916d8d2c62126a6a258d74071bb67e90d27c88e23ea85583c8c12ef0" exitCode=0 Nov 25 10:49:35 crc kubenswrapper[4821]: I1125 10:49:35.228692 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-xxdmc" event={"ID":"bee9a405-f60c-449e-beba-71cd802daa79","Type":"ContainerDied","Data":"789f7caa916d8d2c62126a6a258d74071bb67e90d27c88e23ea85583c8c12ef0"} Nov 25 10:49:35 crc kubenswrapper[4821]: I1125 10:49:35.234113 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-8bp9n" event={"ID":"eb6c9acf-d69c-4788-a85d-b72135de1ef3","Type":"ContainerStarted","Data":"068886c87f6a67c94577ca2971d3df071fe62df65a20f17589ce4f0a9d9008c5"} Nov 25 10:49:35 crc kubenswrapper[4821]: I1125 10:49:35.269039 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-7znwh" podStartSLOduration=3.120261193 podStartE2EDuration="44.269019603s" podCreationTimestamp="2025-11-25 10:48:51 +0000 UTC" firstStartedPulling="2025-11-25 10:48:53.133896656 +0000 UTC m=+1003.670216503" lastFinishedPulling="2025-11-25 10:49:34.282655066 +0000 UTC m=+1044.818974913" observedRunningTime="2025-11-25 10:49:35.242371021 +0000 UTC m=+1045.778690868" watchObservedRunningTime="2025-11-25 10:49:35.269019603 +0000 UTC m=+1045.805339460" Nov 25 10:49:35 crc kubenswrapper[4821]: I1125 10:49:35.280933 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-8bp9n" podStartSLOduration=4.845358126 podStartE2EDuration="45.280909907s" podCreationTimestamp="2025-11-25 10:48:50 +0000 UTC" firstStartedPulling="2025-11-25 10:48:52.735626713 +0000 UTC m=+1003.271946560" lastFinishedPulling="2025-11-25 10:49:33.171178494 +0000 UTC m=+1043.707498341" observedRunningTime="2025-11-25 10:49:35.275998882 +0000 UTC m=+1045.812318739" watchObservedRunningTime="2025-11-25 10:49:35.280909907 +0000 UTC m=+1045.817229754" Nov 25 10:49:35 crc kubenswrapper[4821]: I1125 10:49:35.642763 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-tq9mv" Nov 25 10:49:35 crc kubenswrapper[4821]: I1125 10:49:35.722845 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/325653d2-264c-4c6c-851f-6e38fe3e61a5-combined-ca-bundle\") pod \"325653d2-264c-4c6c-851f-6e38fe3e61a5\" (UID: \"325653d2-264c-4c6c-851f-6e38fe3e61a5\") " Nov 25 10:49:35 crc kubenswrapper[4821]: I1125 10:49:35.723003 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/325653d2-264c-4c6c-851f-6e38fe3e61a5-config\") pod \"325653d2-264c-4c6c-851f-6e38fe3e61a5\" (UID: \"325653d2-264c-4c6c-851f-6e38fe3e61a5\") " Nov 25 10:49:35 crc kubenswrapper[4821]: I1125 10:49:35.723131 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hvdv9\" (UniqueName: \"kubernetes.io/projected/325653d2-264c-4c6c-851f-6e38fe3e61a5-kube-api-access-hvdv9\") pod \"325653d2-264c-4c6c-851f-6e38fe3e61a5\" (UID: \"325653d2-264c-4c6c-851f-6e38fe3e61a5\") " Nov 25 10:49:35 crc kubenswrapper[4821]: I1125 10:49:35.734500 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/325653d2-264c-4c6c-851f-6e38fe3e61a5-kube-api-access-hvdv9" (OuterVolumeSpecName: "kube-api-access-hvdv9") pod "325653d2-264c-4c6c-851f-6e38fe3e61a5" (UID: "325653d2-264c-4c6c-851f-6e38fe3e61a5"). InnerVolumeSpecName "kube-api-access-hvdv9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:49:35 crc kubenswrapper[4821]: I1125 10:49:35.756290 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/325653d2-264c-4c6c-851f-6e38fe3e61a5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "325653d2-264c-4c6c-851f-6e38fe3e61a5" (UID: "325653d2-264c-4c6c-851f-6e38fe3e61a5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:49:35 crc kubenswrapper[4821]: I1125 10:49:35.761466 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/325653d2-264c-4c6c-851f-6e38fe3e61a5-config" (OuterVolumeSpecName: "config") pod "325653d2-264c-4c6c-851f-6e38fe3e61a5" (UID: "325653d2-264c-4c6c-851f-6e38fe3e61a5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:49:35 crc kubenswrapper[4821]: I1125 10:49:35.828342 4821 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/325653d2-264c-4c6c-851f-6e38fe3e61a5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:35 crc kubenswrapper[4821]: I1125 10:49:35.828386 4821 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/325653d2-264c-4c6c-851f-6e38fe3e61a5-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:35 crc kubenswrapper[4821]: I1125 10:49:35.828396 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hvdv9\" (UniqueName: \"kubernetes.io/projected/325653d2-264c-4c6c-851f-6e38fe3e61a5-kube-api-access-hvdv9\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:36 crc kubenswrapper[4821]: I1125 10:49:36.273502 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-tq9mv" Nov 25 10:49:36 crc kubenswrapper[4821]: I1125 10:49:36.274555 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-tq9mv" event={"ID":"325653d2-264c-4c6c-851f-6e38fe3e61a5","Type":"ContainerDied","Data":"d051074367e8fb3a3803c8649e596398c8187a81d4c341c26e7ff844f3526c4c"} Nov 25 10:49:36 crc kubenswrapper[4821]: I1125 10:49:36.274581 4821 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d051074367e8fb3a3803c8649e596398c8187a81d4c341c26e7ff844f3526c4c" Nov 25 10:49:36 crc kubenswrapper[4821]: I1125 10:49:36.471768 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-dpjr6"] Nov 25 10:49:36 crc kubenswrapper[4821]: E1125 10:49:36.472151 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="325653d2-264c-4c6c-851f-6e38fe3e61a5" containerName="neutron-db-sync" Nov 25 10:49:36 crc kubenswrapper[4821]: I1125 10:49:36.472167 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="325653d2-264c-4c6c-851f-6e38fe3e61a5" containerName="neutron-db-sync" Nov 25 10:49:36 crc kubenswrapper[4821]: I1125 10:49:36.472390 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="325653d2-264c-4c6c-851f-6e38fe3e61a5" containerName="neutron-db-sync" Nov 25 10:49:36 crc kubenswrapper[4821]: I1125 10:49:36.473439 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-dpjr6" Nov 25 10:49:36 crc kubenswrapper[4821]: I1125 10:49:36.491652 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-dpjr6"] Nov 25 10:49:36 crc kubenswrapper[4821]: I1125 10:49:36.545857 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/53f9f168-1063-4bb0-8d1c-03482761ae7f-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-dpjr6\" (UID: \"53f9f168-1063-4bb0-8d1c-03482761ae7f\") " pod="openstack/dnsmasq-dns-6b7b667979-dpjr6" Nov 25 10:49:36 crc kubenswrapper[4821]: I1125 10:49:36.545907 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/53f9f168-1063-4bb0-8d1c-03482761ae7f-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-dpjr6\" (UID: \"53f9f168-1063-4bb0-8d1c-03482761ae7f\") " pod="openstack/dnsmasq-dns-6b7b667979-dpjr6" Nov 25 10:49:36 crc kubenswrapper[4821]: I1125 10:49:36.545927 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cbgd9\" (UniqueName: \"kubernetes.io/projected/53f9f168-1063-4bb0-8d1c-03482761ae7f-kube-api-access-cbgd9\") pod \"dnsmasq-dns-6b7b667979-dpjr6\" (UID: \"53f9f168-1063-4bb0-8d1c-03482761ae7f\") " pod="openstack/dnsmasq-dns-6b7b667979-dpjr6" Nov 25 10:49:36 crc kubenswrapper[4821]: I1125 10:49:36.545961 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/53f9f168-1063-4bb0-8d1c-03482761ae7f-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-dpjr6\" (UID: \"53f9f168-1063-4bb0-8d1c-03482761ae7f\") " pod="openstack/dnsmasq-dns-6b7b667979-dpjr6" Nov 25 10:49:36 crc kubenswrapper[4821]: I1125 10:49:36.545981 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53f9f168-1063-4bb0-8d1c-03482761ae7f-config\") pod \"dnsmasq-dns-6b7b667979-dpjr6\" (UID: \"53f9f168-1063-4bb0-8d1c-03482761ae7f\") " pod="openstack/dnsmasq-dns-6b7b667979-dpjr6" Nov 25 10:49:36 crc kubenswrapper[4821]: I1125 10:49:36.546030 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/53f9f168-1063-4bb0-8d1c-03482761ae7f-dns-svc\") pod \"dnsmasq-dns-6b7b667979-dpjr6\" (UID: \"53f9f168-1063-4bb0-8d1c-03482761ae7f\") " pod="openstack/dnsmasq-dns-6b7b667979-dpjr6" Nov 25 10:49:36 crc kubenswrapper[4821]: I1125 10:49:36.638611 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-84cbbf9b54-k7shj"] Nov 25 10:49:36 crc kubenswrapper[4821]: I1125 10:49:36.640177 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-84cbbf9b54-k7shj" Nov 25 10:49:36 crc kubenswrapper[4821]: I1125 10:49:36.646045 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-fv5gx" Nov 25 10:49:36 crc kubenswrapper[4821]: I1125 10:49:36.646369 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 25 10:49:36 crc kubenswrapper[4821]: I1125 10:49:36.646492 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Nov 25 10:49:36 crc kubenswrapper[4821]: I1125 10:49:36.646614 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 25 10:49:36 crc kubenswrapper[4821]: I1125 10:49:36.647788 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/53f9f168-1063-4bb0-8d1c-03482761ae7f-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-dpjr6\" (UID: \"53f9f168-1063-4bb0-8d1c-03482761ae7f\") " pod="openstack/dnsmasq-dns-6b7b667979-dpjr6" Nov 25 10:49:36 crc kubenswrapper[4821]: I1125 10:49:36.647825 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/53f9f168-1063-4bb0-8d1c-03482761ae7f-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-dpjr6\" (UID: \"53f9f168-1063-4bb0-8d1c-03482761ae7f\") " pod="openstack/dnsmasq-dns-6b7b667979-dpjr6" Nov 25 10:49:36 crc kubenswrapper[4821]: I1125 10:49:36.647860 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cbgd9\" (UniqueName: \"kubernetes.io/projected/53f9f168-1063-4bb0-8d1c-03482761ae7f-kube-api-access-cbgd9\") pod \"dnsmasq-dns-6b7b667979-dpjr6\" (UID: \"53f9f168-1063-4bb0-8d1c-03482761ae7f\") " pod="openstack/dnsmasq-dns-6b7b667979-dpjr6" Nov 25 10:49:36 crc kubenswrapper[4821]: I1125 10:49:36.647908 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/53f9f168-1063-4bb0-8d1c-03482761ae7f-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-dpjr6\" (UID: \"53f9f168-1063-4bb0-8d1c-03482761ae7f\") " pod="openstack/dnsmasq-dns-6b7b667979-dpjr6" Nov 25 10:49:36 crc kubenswrapper[4821]: I1125 10:49:36.647931 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53f9f168-1063-4bb0-8d1c-03482761ae7f-config\") pod \"dnsmasq-dns-6b7b667979-dpjr6\" (UID: \"53f9f168-1063-4bb0-8d1c-03482761ae7f\") " pod="openstack/dnsmasq-dns-6b7b667979-dpjr6" Nov 25 10:49:36 crc kubenswrapper[4821]: I1125 10:49:36.648000 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/53f9f168-1063-4bb0-8d1c-03482761ae7f-dns-svc\") pod \"dnsmasq-dns-6b7b667979-dpjr6\" (UID: \"53f9f168-1063-4bb0-8d1c-03482761ae7f\") " pod="openstack/dnsmasq-dns-6b7b667979-dpjr6" Nov 25 10:49:36 crc kubenswrapper[4821]: I1125 10:49:36.649044 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/53f9f168-1063-4bb0-8d1c-03482761ae7f-dns-svc\") pod \"dnsmasq-dns-6b7b667979-dpjr6\" (UID: \"53f9f168-1063-4bb0-8d1c-03482761ae7f\") " pod="openstack/dnsmasq-dns-6b7b667979-dpjr6" Nov 25 10:49:36 crc kubenswrapper[4821]: I1125 10:49:36.649640 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/53f9f168-1063-4bb0-8d1c-03482761ae7f-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-dpjr6\" (UID: \"53f9f168-1063-4bb0-8d1c-03482761ae7f\") " pod="openstack/dnsmasq-dns-6b7b667979-dpjr6" Nov 25 10:49:36 crc kubenswrapper[4821]: I1125 10:49:36.650285 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/53f9f168-1063-4bb0-8d1c-03482761ae7f-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-dpjr6\" (UID: \"53f9f168-1063-4bb0-8d1c-03482761ae7f\") " pod="openstack/dnsmasq-dns-6b7b667979-dpjr6" Nov 25 10:49:36 crc kubenswrapper[4821]: I1125 10:49:36.651101 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53f9f168-1063-4bb0-8d1c-03482761ae7f-config\") pod \"dnsmasq-dns-6b7b667979-dpjr6\" (UID: \"53f9f168-1063-4bb0-8d1c-03482761ae7f\") " pod="openstack/dnsmasq-dns-6b7b667979-dpjr6" Nov 25 10:49:36 crc kubenswrapper[4821]: I1125 10:49:36.651536 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/53f9f168-1063-4bb0-8d1c-03482761ae7f-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-dpjr6\" (UID: \"53f9f168-1063-4bb0-8d1c-03482761ae7f\") " pod="openstack/dnsmasq-dns-6b7b667979-dpjr6" Nov 25 10:49:36 crc kubenswrapper[4821]: I1125 10:49:36.655888 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-84cbbf9b54-k7shj"] Nov 25 10:49:36 crc kubenswrapper[4821]: I1125 10:49:36.683967 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cbgd9\" (UniqueName: \"kubernetes.io/projected/53f9f168-1063-4bb0-8d1c-03482761ae7f-kube-api-access-cbgd9\") pod \"dnsmasq-dns-6b7b667979-dpjr6\" (UID: \"53f9f168-1063-4bb0-8d1c-03482761ae7f\") " pod="openstack/dnsmasq-dns-6b7b667979-dpjr6" Nov 25 10:49:36 crc kubenswrapper[4821]: I1125 10:49:36.750704 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/690339d6-9d90-46b3-b471-afabc2da9e7c-httpd-config\") pod \"neutron-84cbbf9b54-k7shj\" (UID: \"690339d6-9d90-46b3-b471-afabc2da9e7c\") " pod="openstack/neutron-84cbbf9b54-k7shj" Nov 25 10:49:36 crc kubenswrapper[4821]: I1125 10:49:36.750922 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/690339d6-9d90-46b3-b471-afabc2da9e7c-ovndb-tls-certs\") pod \"neutron-84cbbf9b54-k7shj\" (UID: \"690339d6-9d90-46b3-b471-afabc2da9e7c\") " pod="openstack/neutron-84cbbf9b54-k7shj" Nov 25 10:49:36 crc kubenswrapper[4821]: I1125 10:49:36.750997 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/690339d6-9d90-46b3-b471-afabc2da9e7c-config\") pod \"neutron-84cbbf9b54-k7shj\" (UID: \"690339d6-9d90-46b3-b471-afabc2da9e7c\") " pod="openstack/neutron-84cbbf9b54-k7shj" Nov 25 10:49:36 crc kubenswrapper[4821]: I1125 10:49:36.751046 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/690339d6-9d90-46b3-b471-afabc2da9e7c-combined-ca-bundle\") pod \"neutron-84cbbf9b54-k7shj\" (UID: \"690339d6-9d90-46b3-b471-afabc2da9e7c\") " pod="openstack/neutron-84cbbf9b54-k7shj" Nov 25 10:49:36 crc kubenswrapper[4821]: I1125 10:49:36.751219 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bmnv7\" (UniqueName: \"kubernetes.io/projected/690339d6-9d90-46b3-b471-afabc2da9e7c-kube-api-access-bmnv7\") pod \"neutron-84cbbf9b54-k7shj\" (UID: \"690339d6-9d90-46b3-b471-afabc2da9e7c\") " pod="openstack/neutron-84cbbf9b54-k7shj" Nov 25 10:49:36 crc kubenswrapper[4821]: I1125 10:49:36.833119 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-dpjr6" Nov 25 10:49:36 crc kubenswrapper[4821]: I1125 10:49:36.853857 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/690339d6-9d90-46b3-b471-afabc2da9e7c-ovndb-tls-certs\") pod \"neutron-84cbbf9b54-k7shj\" (UID: \"690339d6-9d90-46b3-b471-afabc2da9e7c\") " pod="openstack/neutron-84cbbf9b54-k7shj" Nov 25 10:49:36 crc kubenswrapper[4821]: I1125 10:49:36.853909 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/690339d6-9d90-46b3-b471-afabc2da9e7c-config\") pod \"neutron-84cbbf9b54-k7shj\" (UID: \"690339d6-9d90-46b3-b471-afabc2da9e7c\") " pod="openstack/neutron-84cbbf9b54-k7shj" Nov 25 10:49:36 crc kubenswrapper[4821]: I1125 10:49:36.853935 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/690339d6-9d90-46b3-b471-afabc2da9e7c-combined-ca-bundle\") pod \"neutron-84cbbf9b54-k7shj\" (UID: \"690339d6-9d90-46b3-b471-afabc2da9e7c\") " pod="openstack/neutron-84cbbf9b54-k7shj" Nov 25 10:49:36 crc kubenswrapper[4821]: I1125 10:49:36.854424 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bmnv7\" (UniqueName: \"kubernetes.io/projected/690339d6-9d90-46b3-b471-afabc2da9e7c-kube-api-access-bmnv7\") pod \"neutron-84cbbf9b54-k7shj\" (UID: \"690339d6-9d90-46b3-b471-afabc2da9e7c\") " pod="openstack/neutron-84cbbf9b54-k7shj" Nov 25 10:49:36 crc kubenswrapper[4821]: I1125 10:49:36.854506 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/690339d6-9d90-46b3-b471-afabc2da9e7c-httpd-config\") pod \"neutron-84cbbf9b54-k7shj\" (UID: \"690339d6-9d90-46b3-b471-afabc2da9e7c\") " pod="openstack/neutron-84cbbf9b54-k7shj" Nov 25 10:49:36 crc kubenswrapper[4821]: I1125 10:49:36.862305 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/690339d6-9d90-46b3-b471-afabc2da9e7c-combined-ca-bundle\") pod \"neutron-84cbbf9b54-k7shj\" (UID: \"690339d6-9d90-46b3-b471-afabc2da9e7c\") " pod="openstack/neutron-84cbbf9b54-k7shj" Nov 25 10:49:36 crc kubenswrapper[4821]: I1125 10:49:36.863142 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/690339d6-9d90-46b3-b471-afabc2da9e7c-ovndb-tls-certs\") pod \"neutron-84cbbf9b54-k7shj\" (UID: \"690339d6-9d90-46b3-b471-afabc2da9e7c\") " pod="openstack/neutron-84cbbf9b54-k7shj" Nov 25 10:49:36 crc kubenswrapper[4821]: I1125 10:49:36.878971 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/690339d6-9d90-46b3-b471-afabc2da9e7c-config\") pod \"neutron-84cbbf9b54-k7shj\" (UID: \"690339d6-9d90-46b3-b471-afabc2da9e7c\") " pod="openstack/neutron-84cbbf9b54-k7shj" Nov 25 10:49:36 crc kubenswrapper[4821]: I1125 10:49:36.881641 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/690339d6-9d90-46b3-b471-afabc2da9e7c-httpd-config\") pod \"neutron-84cbbf9b54-k7shj\" (UID: \"690339d6-9d90-46b3-b471-afabc2da9e7c\") " pod="openstack/neutron-84cbbf9b54-k7shj" Nov 25 10:49:36 crc kubenswrapper[4821]: I1125 10:49:36.890287 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bmnv7\" (UniqueName: \"kubernetes.io/projected/690339d6-9d90-46b3-b471-afabc2da9e7c-kube-api-access-bmnv7\") pod \"neutron-84cbbf9b54-k7shj\" (UID: \"690339d6-9d90-46b3-b471-afabc2da9e7c\") " pod="openstack/neutron-84cbbf9b54-k7shj" Nov 25 10:49:37 crc kubenswrapper[4821]: I1125 10:49:37.025895 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-xxdmc" Nov 25 10:49:37 crc kubenswrapper[4821]: I1125 10:49:37.029532 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-84cbbf9b54-k7shj" Nov 25 10:49:37 crc kubenswrapper[4821]: I1125 10:49:37.060563 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bee9a405-f60c-449e-beba-71cd802daa79-config-data\") pod \"bee9a405-f60c-449e-beba-71cd802daa79\" (UID: \"bee9a405-f60c-449e-beba-71cd802daa79\") " Nov 25 10:49:37 crc kubenswrapper[4821]: I1125 10:49:37.060599 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bee9a405-f60c-449e-beba-71cd802daa79-combined-ca-bundle\") pod \"bee9a405-f60c-449e-beba-71cd802daa79\" (UID: \"bee9a405-f60c-449e-beba-71cd802daa79\") " Nov 25 10:49:37 crc kubenswrapper[4821]: I1125 10:49:37.060660 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bee9a405-f60c-449e-beba-71cd802daa79-logs\") pod \"bee9a405-f60c-449e-beba-71cd802daa79\" (UID: \"bee9a405-f60c-449e-beba-71cd802daa79\") " Nov 25 10:49:37 crc kubenswrapper[4821]: I1125 10:49:37.060722 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g4h77\" (UniqueName: \"kubernetes.io/projected/bee9a405-f60c-449e-beba-71cd802daa79-kube-api-access-g4h77\") pod \"bee9a405-f60c-449e-beba-71cd802daa79\" (UID: \"bee9a405-f60c-449e-beba-71cd802daa79\") " Nov 25 10:49:37 crc kubenswrapper[4821]: I1125 10:49:37.060792 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bee9a405-f60c-449e-beba-71cd802daa79-scripts\") pod \"bee9a405-f60c-449e-beba-71cd802daa79\" (UID: \"bee9a405-f60c-449e-beba-71cd802daa79\") " Nov 25 10:49:37 crc kubenswrapper[4821]: I1125 10:49:37.063035 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bee9a405-f60c-449e-beba-71cd802daa79-logs" (OuterVolumeSpecName: "logs") pod "bee9a405-f60c-449e-beba-71cd802daa79" (UID: "bee9a405-f60c-449e-beba-71cd802daa79"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:49:37 crc kubenswrapper[4821]: I1125 10:49:37.075155 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bee9a405-f60c-449e-beba-71cd802daa79-kube-api-access-g4h77" (OuterVolumeSpecName: "kube-api-access-g4h77") pod "bee9a405-f60c-449e-beba-71cd802daa79" (UID: "bee9a405-f60c-449e-beba-71cd802daa79"). InnerVolumeSpecName "kube-api-access-g4h77". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:49:37 crc kubenswrapper[4821]: I1125 10:49:37.095539 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bee9a405-f60c-449e-beba-71cd802daa79-scripts" (OuterVolumeSpecName: "scripts") pod "bee9a405-f60c-449e-beba-71cd802daa79" (UID: "bee9a405-f60c-449e-beba-71cd802daa79"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:49:37 crc kubenswrapper[4821]: I1125 10:49:37.162100 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bee9a405-f60c-449e-beba-71cd802daa79-config-data" (OuterVolumeSpecName: "config-data") pod "bee9a405-f60c-449e-beba-71cd802daa79" (UID: "bee9a405-f60c-449e-beba-71cd802daa79"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:49:37 crc kubenswrapper[4821]: I1125 10:49:37.162423 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bee9a405-f60c-449e-beba-71cd802daa79-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bee9a405-f60c-449e-beba-71cd802daa79" (UID: "bee9a405-f60c-449e-beba-71cd802daa79"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:49:37 crc kubenswrapper[4821]: I1125 10:49:37.163345 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bee9a405-f60c-449e-beba-71cd802daa79-config-data\") pod \"bee9a405-f60c-449e-beba-71cd802daa79\" (UID: \"bee9a405-f60c-449e-beba-71cd802daa79\") " Nov 25 10:49:37 crc kubenswrapper[4821]: I1125 10:49:37.163457 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bee9a405-f60c-449e-beba-71cd802daa79-combined-ca-bundle\") pod \"bee9a405-f60c-449e-beba-71cd802daa79\" (UID: \"bee9a405-f60c-449e-beba-71cd802daa79\") " Nov 25 10:49:37 crc kubenswrapper[4821]: W1125 10:49:37.163562 4821 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/bee9a405-f60c-449e-beba-71cd802daa79/volumes/kubernetes.io~secret/config-data Nov 25 10:49:37 crc kubenswrapper[4821]: I1125 10:49:37.163589 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bee9a405-f60c-449e-beba-71cd802daa79-config-data" (OuterVolumeSpecName: "config-data") pod "bee9a405-f60c-449e-beba-71cd802daa79" (UID: "bee9a405-f60c-449e-beba-71cd802daa79"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:49:37 crc kubenswrapper[4821]: W1125 10:49:37.163731 4821 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/bee9a405-f60c-449e-beba-71cd802daa79/volumes/kubernetes.io~secret/combined-ca-bundle Nov 25 10:49:37 crc kubenswrapper[4821]: I1125 10:49:37.163741 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bee9a405-f60c-449e-beba-71cd802daa79-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bee9a405-f60c-449e-beba-71cd802daa79" (UID: "bee9a405-f60c-449e-beba-71cd802daa79"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:49:37 crc kubenswrapper[4821]: I1125 10:49:37.164011 4821 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bee9a405-f60c-449e-beba-71cd802daa79-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:37 crc kubenswrapper[4821]: I1125 10:49:37.164028 4821 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bee9a405-f60c-449e-beba-71cd802daa79-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:37 crc kubenswrapper[4821]: I1125 10:49:37.164081 4821 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bee9a405-f60c-449e-beba-71cd802daa79-logs\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:37 crc kubenswrapper[4821]: I1125 10:49:37.164093 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g4h77\" (UniqueName: \"kubernetes.io/projected/bee9a405-f60c-449e-beba-71cd802daa79-kube-api-access-g4h77\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:37 crc kubenswrapper[4821]: I1125 10:49:37.164104 4821 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bee9a405-f60c-449e-beba-71cd802daa79-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:37 crc kubenswrapper[4821]: I1125 10:49:37.303583 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-xxdmc" event={"ID":"bee9a405-f60c-449e-beba-71cd802daa79","Type":"ContainerDied","Data":"14e27cff5971cdc47d84904e5d760f78db7c0348d4da8f82991561d3d2924af3"} Nov 25 10:49:37 crc kubenswrapper[4821]: I1125 10:49:37.303942 4821 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="14e27cff5971cdc47d84904e5d760f78db7c0348d4da8f82991561d3d2924af3" Nov 25 10:49:37 crc kubenswrapper[4821]: I1125 10:49:37.304007 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-xxdmc" Nov 25 10:49:37 crc kubenswrapper[4821]: I1125 10:49:37.463264 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-f8667547d-9v5xh"] Nov 25 10:49:37 crc kubenswrapper[4821]: E1125 10:49:37.463668 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bee9a405-f60c-449e-beba-71cd802daa79" containerName="placement-db-sync" Nov 25 10:49:37 crc kubenswrapper[4821]: I1125 10:49:37.463684 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="bee9a405-f60c-449e-beba-71cd802daa79" containerName="placement-db-sync" Nov 25 10:49:37 crc kubenswrapper[4821]: I1125 10:49:37.463920 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="bee9a405-f60c-449e-beba-71cd802daa79" containerName="placement-db-sync" Nov 25 10:49:37 crc kubenswrapper[4821]: I1125 10:49:37.464869 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-f8667547d-9v5xh" Nov 25 10:49:37 crc kubenswrapper[4821]: I1125 10:49:37.471276 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Nov 25 10:49:37 crc kubenswrapper[4821]: I1125 10:49:37.471522 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 25 10:49:37 crc kubenswrapper[4821]: I1125 10:49:37.471566 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Nov 25 10:49:37 crc kubenswrapper[4821]: I1125 10:49:37.471763 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-ndsbm" Nov 25 10:49:37 crc kubenswrapper[4821]: I1125 10:49:37.489396 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b755bd74-3de7-49ed-857a-a5f23cf84f30-scripts\") pod \"placement-f8667547d-9v5xh\" (UID: \"b755bd74-3de7-49ed-857a-a5f23cf84f30\") " pod="openstack/placement-f8667547d-9v5xh" Nov 25 10:49:37 crc kubenswrapper[4821]: I1125 10:49:37.489482 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b755bd74-3de7-49ed-857a-a5f23cf84f30-logs\") pod \"placement-f8667547d-9v5xh\" (UID: \"b755bd74-3de7-49ed-857a-a5f23cf84f30\") " pod="openstack/placement-f8667547d-9v5xh" Nov 25 10:49:37 crc kubenswrapper[4821]: I1125 10:49:37.489551 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b755bd74-3de7-49ed-857a-a5f23cf84f30-combined-ca-bundle\") pod \"placement-f8667547d-9v5xh\" (UID: \"b755bd74-3de7-49ed-857a-a5f23cf84f30\") " pod="openstack/placement-f8667547d-9v5xh" Nov 25 10:49:37 crc kubenswrapper[4821]: I1125 10:49:37.489620 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b755bd74-3de7-49ed-857a-a5f23cf84f30-public-tls-certs\") pod \"placement-f8667547d-9v5xh\" (UID: \"b755bd74-3de7-49ed-857a-a5f23cf84f30\") " pod="openstack/placement-f8667547d-9v5xh" Nov 25 10:49:37 crc kubenswrapper[4821]: I1125 10:49:37.489893 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b755bd74-3de7-49ed-857a-a5f23cf84f30-config-data\") pod \"placement-f8667547d-9v5xh\" (UID: \"b755bd74-3de7-49ed-857a-a5f23cf84f30\") " pod="openstack/placement-f8667547d-9v5xh" Nov 25 10:49:37 crc kubenswrapper[4821]: I1125 10:49:37.491003 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l68sc\" (UniqueName: \"kubernetes.io/projected/b755bd74-3de7-49ed-857a-a5f23cf84f30-kube-api-access-l68sc\") pod \"placement-f8667547d-9v5xh\" (UID: \"b755bd74-3de7-49ed-857a-a5f23cf84f30\") " pod="openstack/placement-f8667547d-9v5xh" Nov 25 10:49:37 crc kubenswrapper[4821]: I1125 10:49:37.491058 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b755bd74-3de7-49ed-857a-a5f23cf84f30-internal-tls-certs\") pod \"placement-f8667547d-9v5xh\" (UID: \"b755bd74-3de7-49ed-857a-a5f23cf84f30\") " pod="openstack/placement-f8667547d-9v5xh" Nov 25 10:49:37 crc kubenswrapper[4821]: I1125 10:49:37.531028 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 25 10:49:37 crc kubenswrapper[4821]: I1125 10:49:37.542536 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-f8667547d-9v5xh"] Nov 25 10:49:37 crc kubenswrapper[4821]: I1125 10:49:37.596763 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b755bd74-3de7-49ed-857a-a5f23cf84f30-scripts\") pod \"placement-f8667547d-9v5xh\" (UID: \"b755bd74-3de7-49ed-857a-a5f23cf84f30\") " pod="openstack/placement-f8667547d-9v5xh" Nov 25 10:49:37 crc kubenswrapper[4821]: I1125 10:49:37.596822 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b755bd74-3de7-49ed-857a-a5f23cf84f30-logs\") pod \"placement-f8667547d-9v5xh\" (UID: \"b755bd74-3de7-49ed-857a-a5f23cf84f30\") " pod="openstack/placement-f8667547d-9v5xh" Nov 25 10:49:37 crc kubenswrapper[4821]: I1125 10:49:37.596854 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b755bd74-3de7-49ed-857a-a5f23cf84f30-combined-ca-bundle\") pod \"placement-f8667547d-9v5xh\" (UID: \"b755bd74-3de7-49ed-857a-a5f23cf84f30\") " pod="openstack/placement-f8667547d-9v5xh" Nov 25 10:49:37 crc kubenswrapper[4821]: I1125 10:49:37.596883 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b755bd74-3de7-49ed-857a-a5f23cf84f30-public-tls-certs\") pod \"placement-f8667547d-9v5xh\" (UID: \"b755bd74-3de7-49ed-857a-a5f23cf84f30\") " pod="openstack/placement-f8667547d-9v5xh" Nov 25 10:49:37 crc kubenswrapper[4821]: I1125 10:49:37.596899 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b755bd74-3de7-49ed-857a-a5f23cf84f30-config-data\") pod \"placement-f8667547d-9v5xh\" (UID: \"b755bd74-3de7-49ed-857a-a5f23cf84f30\") " pod="openstack/placement-f8667547d-9v5xh" Nov 25 10:49:37 crc kubenswrapper[4821]: I1125 10:49:37.596946 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l68sc\" (UniqueName: \"kubernetes.io/projected/b755bd74-3de7-49ed-857a-a5f23cf84f30-kube-api-access-l68sc\") pod \"placement-f8667547d-9v5xh\" (UID: \"b755bd74-3de7-49ed-857a-a5f23cf84f30\") " pod="openstack/placement-f8667547d-9v5xh" Nov 25 10:49:37 crc kubenswrapper[4821]: I1125 10:49:37.596975 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b755bd74-3de7-49ed-857a-a5f23cf84f30-internal-tls-certs\") pod \"placement-f8667547d-9v5xh\" (UID: \"b755bd74-3de7-49ed-857a-a5f23cf84f30\") " pod="openstack/placement-f8667547d-9v5xh" Nov 25 10:49:37 crc kubenswrapper[4821]: I1125 10:49:37.601078 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b755bd74-3de7-49ed-857a-a5f23cf84f30-logs\") pod \"placement-f8667547d-9v5xh\" (UID: \"b755bd74-3de7-49ed-857a-a5f23cf84f30\") " pod="openstack/placement-f8667547d-9v5xh" Nov 25 10:49:37 crc kubenswrapper[4821]: I1125 10:49:37.603911 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b755bd74-3de7-49ed-857a-a5f23cf84f30-public-tls-certs\") pod \"placement-f8667547d-9v5xh\" (UID: \"b755bd74-3de7-49ed-857a-a5f23cf84f30\") " pod="openstack/placement-f8667547d-9v5xh" Nov 25 10:49:37 crc kubenswrapper[4821]: I1125 10:49:37.604373 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b755bd74-3de7-49ed-857a-a5f23cf84f30-internal-tls-certs\") pod \"placement-f8667547d-9v5xh\" (UID: \"b755bd74-3de7-49ed-857a-a5f23cf84f30\") " pod="openstack/placement-f8667547d-9v5xh" Nov 25 10:49:37 crc kubenswrapper[4821]: I1125 10:49:37.605916 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-dpjr6"] Nov 25 10:49:37 crc kubenswrapper[4821]: I1125 10:49:37.608112 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b755bd74-3de7-49ed-857a-a5f23cf84f30-combined-ca-bundle\") pod \"placement-f8667547d-9v5xh\" (UID: \"b755bd74-3de7-49ed-857a-a5f23cf84f30\") " pod="openstack/placement-f8667547d-9v5xh" Nov 25 10:49:37 crc kubenswrapper[4821]: I1125 10:49:37.609387 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b755bd74-3de7-49ed-857a-a5f23cf84f30-config-data\") pod \"placement-f8667547d-9v5xh\" (UID: \"b755bd74-3de7-49ed-857a-a5f23cf84f30\") " pod="openstack/placement-f8667547d-9v5xh" Nov 25 10:49:37 crc kubenswrapper[4821]: I1125 10:49:37.610841 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b755bd74-3de7-49ed-857a-a5f23cf84f30-scripts\") pod \"placement-f8667547d-9v5xh\" (UID: \"b755bd74-3de7-49ed-857a-a5f23cf84f30\") " pod="openstack/placement-f8667547d-9v5xh" Nov 25 10:49:37 crc kubenswrapper[4821]: I1125 10:49:37.623922 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l68sc\" (UniqueName: \"kubernetes.io/projected/b755bd74-3de7-49ed-857a-a5f23cf84f30-kube-api-access-l68sc\") pod \"placement-f8667547d-9v5xh\" (UID: \"b755bd74-3de7-49ed-857a-a5f23cf84f30\") " pod="openstack/placement-f8667547d-9v5xh" Nov 25 10:49:37 crc kubenswrapper[4821]: I1125 10:49:37.817642 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-f8667547d-9v5xh" Nov 25 10:49:37 crc kubenswrapper[4821]: I1125 10:49:37.950399 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-84cbbf9b54-k7shj"] Nov 25 10:49:37 crc kubenswrapper[4821]: W1125 10:49:37.988386 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod690339d6_9d90_46b3_b471_afabc2da9e7c.slice/crio-7c8f465e6e511bedd5549cf2ae1805de33731586756bc5ef47c073dce9b38b0c WatchSource:0}: Error finding container 7c8f465e6e511bedd5549cf2ae1805de33731586756bc5ef47c073dce9b38b0c: Status 404 returned error can't find the container with id 7c8f465e6e511bedd5549cf2ae1805de33731586756bc5ef47c073dce9b38b0c Nov 25 10:49:38 crc kubenswrapper[4821]: I1125 10:49:38.320602 4821 generic.go:334] "Generic (PLEG): container finished" podID="53f9f168-1063-4bb0-8d1c-03482761ae7f" containerID="8e12b5f21ce0e86d4aebb00d928fcd06fd64a985b3d050b09d26a3e16cea1647" exitCode=0 Nov 25 10:49:38 crc kubenswrapper[4821]: I1125 10:49:38.320734 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-dpjr6" event={"ID":"53f9f168-1063-4bb0-8d1c-03482761ae7f","Type":"ContainerDied","Data":"8e12b5f21ce0e86d4aebb00d928fcd06fd64a985b3d050b09d26a3e16cea1647"} Nov 25 10:49:38 crc kubenswrapper[4821]: I1125 10:49:38.321057 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-dpjr6" event={"ID":"53f9f168-1063-4bb0-8d1c-03482761ae7f","Type":"ContainerStarted","Data":"958ca843b2ab4ec81ead689869cb03397b72daa4519d47bf3f5386e2894b2290"} Nov 25 10:49:38 crc kubenswrapper[4821]: I1125 10:49:38.328031 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-84cbbf9b54-k7shj" event={"ID":"690339d6-9d90-46b3-b471-afabc2da9e7c","Type":"ContainerStarted","Data":"7c8f465e6e511bedd5549cf2ae1805de33731586756bc5ef47c073dce9b38b0c"} Nov 25 10:49:38 crc kubenswrapper[4821]: I1125 10:49:38.462199 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-f8667547d-9v5xh"] Nov 25 10:49:38 crc kubenswrapper[4821]: W1125 10:49:38.475403 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb755bd74_3de7_49ed_857a_a5f23cf84f30.slice/crio-0d4226e2723641f0e2a157f21ba51e7afdf49571648e030fb31b7b0ac2887043 WatchSource:0}: Error finding container 0d4226e2723641f0e2a157f21ba51e7afdf49571648e030fb31b7b0ac2887043: Status 404 returned error can't find the container with id 0d4226e2723641f0e2a157f21ba51e7afdf49571648e030fb31b7b0ac2887043 Nov 25 10:49:39 crc kubenswrapper[4821]: I1125 10:49:39.345013 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-dpjr6" event={"ID":"53f9f168-1063-4bb0-8d1c-03482761ae7f","Type":"ContainerStarted","Data":"065de96b781ff5e2a331bea1a4e20284824bef6044fdf9fc5450f3ff83be103c"} Nov 25 10:49:39 crc kubenswrapper[4821]: I1125 10:49:39.345615 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6b7b667979-dpjr6" Nov 25 10:49:39 crc kubenswrapper[4821]: I1125 10:49:39.348653 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-84cbbf9b54-k7shj" event={"ID":"690339d6-9d90-46b3-b471-afabc2da9e7c","Type":"ContainerStarted","Data":"9fc413e8b8971197f05d3f9d1675fd56f8c25a1a5c97327431248c71d7982d6f"} Nov 25 10:49:39 crc kubenswrapper[4821]: I1125 10:49:39.348701 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-84cbbf9b54-k7shj" event={"ID":"690339d6-9d90-46b3-b471-afabc2da9e7c","Type":"ContainerStarted","Data":"520722ded67872ccc942a42008a87b2908384fdc9b316a93acadb80181fde2e9"} Nov 25 10:49:39 crc kubenswrapper[4821]: I1125 10:49:39.348767 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-84cbbf9b54-k7shj" Nov 25 10:49:39 crc kubenswrapper[4821]: I1125 10:49:39.352933 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-f8667547d-9v5xh" event={"ID":"b755bd74-3de7-49ed-857a-a5f23cf84f30","Type":"ContainerStarted","Data":"612dc30b5bf11bbfd3bee9e769423ff50d1393b54017f03d9d68b30a5699ff0d"} Nov 25 10:49:39 crc kubenswrapper[4821]: I1125 10:49:39.352985 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-f8667547d-9v5xh" event={"ID":"b755bd74-3de7-49ed-857a-a5f23cf84f30","Type":"ContainerStarted","Data":"3e6f4abf53d825388e54ec6db980d8e1bc68d52cc35d74c53427166b484dc6ae"} Nov 25 10:49:39 crc kubenswrapper[4821]: I1125 10:49:39.353007 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-f8667547d-9v5xh" Nov 25 10:49:39 crc kubenswrapper[4821]: I1125 10:49:39.353018 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-f8667547d-9v5xh" event={"ID":"b755bd74-3de7-49ed-857a-a5f23cf84f30","Type":"ContainerStarted","Data":"0d4226e2723641f0e2a157f21ba51e7afdf49571648e030fb31b7b0ac2887043"} Nov 25 10:49:39 crc kubenswrapper[4821]: I1125 10:49:39.353032 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-f8667547d-9v5xh" Nov 25 10:49:39 crc kubenswrapper[4821]: I1125 10:49:39.355414 4821 generic.go:334] "Generic (PLEG): container finished" podID="2eb6852b-84a1-4d4a-8402-77a7e2252ab2" containerID="8f3e63e3b217fdcee82a9e2153ed5efa0a9ca89caff4b8c351478c6944df32c4" exitCode=0 Nov 25 10:49:39 crc kubenswrapper[4821]: I1125 10:49:39.355449 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-7znwh" event={"ID":"2eb6852b-84a1-4d4a-8402-77a7e2252ab2","Type":"ContainerDied","Data":"8f3e63e3b217fdcee82a9e2153ed5efa0a9ca89caff4b8c351478c6944df32c4"} Nov 25 10:49:39 crc kubenswrapper[4821]: I1125 10:49:39.375034 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6b7b667979-dpjr6" podStartSLOduration=3.375014883 podStartE2EDuration="3.375014883s" podCreationTimestamp="2025-11-25 10:49:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:49:39.366143966 +0000 UTC m=+1049.902463823" watchObservedRunningTime="2025-11-25 10:49:39.375014883 +0000 UTC m=+1049.911334730" Nov 25 10:49:39 crc kubenswrapper[4821]: I1125 10:49:39.409714 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-f8667547d-9v5xh" podStartSLOduration=2.409694521 podStartE2EDuration="2.409694521s" podCreationTimestamp="2025-11-25 10:49:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:49:39.399073819 +0000 UTC m=+1049.935393676" watchObservedRunningTime="2025-11-25 10:49:39.409694521 +0000 UTC m=+1049.946014368" Nov 25 10:49:39 crc kubenswrapper[4821]: I1125 10:49:39.427768 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-84cbbf9b54-k7shj" podStartSLOduration=3.427748624 podStartE2EDuration="3.427748624s" podCreationTimestamp="2025-11-25 10:49:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:49:39.418923678 +0000 UTC m=+1049.955243525" watchObservedRunningTime="2025-11-25 10:49:39.427748624 +0000 UTC m=+1049.964068471" Nov 25 10:49:39 crc kubenswrapper[4821]: I1125 10:49:39.899767 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-55575b4d57-6t4kz"] Nov 25 10:49:39 crc kubenswrapper[4821]: I1125 10:49:39.901567 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-55575b4d57-6t4kz" Nov 25 10:49:39 crc kubenswrapper[4821]: I1125 10:49:39.903928 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Nov 25 10:49:39 crc kubenswrapper[4821]: I1125 10:49:39.904019 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Nov 25 10:49:39 crc kubenswrapper[4821]: I1125 10:49:39.916301 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-55575b4d57-6t4kz"] Nov 25 10:49:39 crc kubenswrapper[4821]: I1125 10:49:39.966154 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8419a57-35ff-4340-b2d7-fd11724948fe-combined-ca-bundle\") pod \"neutron-55575b4d57-6t4kz\" (UID: \"a8419a57-35ff-4340-b2d7-fd11724948fe\") " pod="openstack/neutron-55575b4d57-6t4kz" Nov 25 10:49:39 crc kubenswrapper[4821]: I1125 10:49:39.966232 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a8419a57-35ff-4340-b2d7-fd11724948fe-public-tls-certs\") pod \"neutron-55575b4d57-6t4kz\" (UID: \"a8419a57-35ff-4340-b2d7-fd11724948fe\") " pod="openstack/neutron-55575b4d57-6t4kz" Nov 25 10:49:39 crc kubenswrapper[4821]: I1125 10:49:39.966272 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a8419a57-35ff-4340-b2d7-fd11724948fe-ovndb-tls-certs\") pod \"neutron-55575b4d57-6t4kz\" (UID: \"a8419a57-35ff-4340-b2d7-fd11724948fe\") " pod="openstack/neutron-55575b4d57-6t4kz" Nov 25 10:49:39 crc kubenswrapper[4821]: I1125 10:49:39.966912 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a8419a57-35ff-4340-b2d7-fd11724948fe-httpd-config\") pod \"neutron-55575b4d57-6t4kz\" (UID: \"a8419a57-35ff-4340-b2d7-fd11724948fe\") " pod="openstack/neutron-55575b4d57-6t4kz" Nov 25 10:49:39 crc kubenswrapper[4821]: I1125 10:49:39.966971 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/a8419a57-35ff-4340-b2d7-fd11724948fe-config\") pod \"neutron-55575b4d57-6t4kz\" (UID: \"a8419a57-35ff-4340-b2d7-fd11724948fe\") " pod="openstack/neutron-55575b4d57-6t4kz" Nov 25 10:49:39 crc kubenswrapper[4821]: I1125 10:49:39.967147 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a8419a57-35ff-4340-b2d7-fd11724948fe-internal-tls-certs\") pod \"neutron-55575b4d57-6t4kz\" (UID: \"a8419a57-35ff-4340-b2d7-fd11724948fe\") " pod="openstack/neutron-55575b4d57-6t4kz" Nov 25 10:49:39 crc kubenswrapper[4821]: I1125 10:49:39.967210 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7q8q\" (UniqueName: \"kubernetes.io/projected/a8419a57-35ff-4340-b2d7-fd11724948fe-kube-api-access-m7q8q\") pod \"neutron-55575b4d57-6t4kz\" (UID: \"a8419a57-35ff-4340-b2d7-fd11724948fe\") " pod="openstack/neutron-55575b4d57-6t4kz" Nov 25 10:49:40 crc kubenswrapper[4821]: I1125 10:49:40.068789 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a8419a57-35ff-4340-b2d7-fd11724948fe-internal-tls-certs\") pod \"neutron-55575b4d57-6t4kz\" (UID: \"a8419a57-35ff-4340-b2d7-fd11724948fe\") " pod="openstack/neutron-55575b4d57-6t4kz" Nov 25 10:49:40 crc kubenswrapper[4821]: I1125 10:49:40.068835 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7q8q\" (UniqueName: \"kubernetes.io/projected/a8419a57-35ff-4340-b2d7-fd11724948fe-kube-api-access-m7q8q\") pod \"neutron-55575b4d57-6t4kz\" (UID: \"a8419a57-35ff-4340-b2d7-fd11724948fe\") " pod="openstack/neutron-55575b4d57-6t4kz" Nov 25 10:49:40 crc kubenswrapper[4821]: I1125 10:49:40.068889 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8419a57-35ff-4340-b2d7-fd11724948fe-combined-ca-bundle\") pod \"neutron-55575b4d57-6t4kz\" (UID: \"a8419a57-35ff-4340-b2d7-fd11724948fe\") " pod="openstack/neutron-55575b4d57-6t4kz" Nov 25 10:49:40 crc kubenswrapper[4821]: I1125 10:49:40.068905 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a8419a57-35ff-4340-b2d7-fd11724948fe-public-tls-certs\") pod \"neutron-55575b4d57-6t4kz\" (UID: \"a8419a57-35ff-4340-b2d7-fd11724948fe\") " pod="openstack/neutron-55575b4d57-6t4kz" Nov 25 10:49:40 crc kubenswrapper[4821]: I1125 10:49:40.068928 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a8419a57-35ff-4340-b2d7-fd11724948fe-ovndb-tls-certs\") pod \"neutron-55575b4d57-6t4kz\" (UID: \"a8419a57-35ff-4340-b2d7-fd11724948fe\") " pod="openstack/neutron-55575b4d57-6t4kz" Nov 25 10:49:40 crc kubenswrapper[4821]: I1125 10:49:40.068993 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a8419a57-35ff-4340-b2d7-fd11724948fe-httpd-config\") pod \"neutron-55575b4d57-6t4kz\" (UID: \"a8419a57-35ff-4340-b2d7-fd11724948fe\") " pod="openstack/neutron-55575b4d57-6t4kz" Nov 25 10:49:40 crc kubenswrapper[4821]: I1125 10:49:40.069011 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/a8419a57-35ff-4340-b2d7-fd11724948fe-config\") pod \"neutron-55575b4d57-6t4kz\" (UID: \"a8419a57-35ff-4340-b2d7-fd11724948fe\") " pod="openstack/neutron-55575b4d57-6t4kz" Nov 25 10:49:40 crc kubenswrapper[4821]: I1125 10:49:40.084520 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8419a57-35ff-4340-b2d7-fd11724948fe-combined-ca-bundle\") pod \"neutron-55575b4d57-6t4kz\" (UID: \"a8419a57-35ff-4340-b2d7-fd11724948fe\") " pod="openstack/neutron-55575b4d57-6t4kz" Nov 25 10:49:40 crc kubenswrapper[4821]: I1125 10:49:40.085484 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a8419a57-35ff-4340-b2d7-fd11724948fe-internal-tls-certs\") pod \"neutron-55575b4d57-6t4kz\" (UID: \"a8419a57-35ff-4340-b2d7-fd11724948fe\") " pod="openstack/neutron-55575b4d57-6t4kz" Nov 25 10:49:40 crc kubenswrapper[4821]: I1125 10:49:40.085703 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a8419a57-35ff-4340-b2d7-fd11724948fe-ovndb-tls-certs\") pod \"neutron-55575b4d57-6t4kz\" (UID: \"a8419a57-35ff-4340-b2d7-fd11724948fe\") " pod="openstack/neutron-55575b4d57-6t4kz" Nov 25 10:49:40 crc kubenswrapper[4821]: I1125 10:49:40.087998 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7q8q\" (UniqueName: \"kubernetes.io/projected/a8419a57-35ff-4340-b2d7-fd11724948fe-kube-api-access-m7q8q\") pod \"neutron-55575b4d57-6t4kz\" (UID: \"a8419a57-35ff-4340-b2d7-fd11724948fe\") " pod="openstack/neutron-55575b4d57-6t4kz" Nov 25 10:49:40 crc kubenswrapper[4821]: I1125 10:49:40.088299 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/a8419a57-35ff-4340-b2d7-fd11724948fe-config\") pod \"neutron-55575b4d57-6t4kz\" (UID: \"a8419a57-35ff-4340-b2d7-fd11724948fe\") " pod="openstack/neutron-55575b4d57-6t4kz" Nov 25 10:49:40 crc kubenswrapper[4821]: I1125 10:49:40.089520 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a8419a57-35ff-4340-b2d7-fd11724948fe-public-tls-certs\") pod \"neutron-55575b4d57-6t4kz\" (UID: \"a8419a57-35ff-4340-b2d7-fd11724948fe\") " pod="openstack/neutron-55575b4d57-6t4kz" Nov 25 10:49:40 crc kubenswrapper[4821]: I1125 10:49:40.111047 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a8419a57-35ff-4340-b2d7-fd11724948fe-httpd-config\") pod \"neutron-55575b4d57-6t4kz\" (UID: \"a8419a57-35ff-4340-b2d7-fd11724948fe\") " pod="openstack/neutron-55575b4d57-6t4kz" Nov 25 10:49:40 crc kubenswrapper[4821]: I1125 10:49:40.218609 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-55575b4d57-6t4kz" Nov 25 10:49:40 crc kubenswrapper[4821]: I1125 10:49:40.993947 4821 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-869c6f5bfb-l785w" podUID="e2327ad3-5d88-4f4b-91ea-91deaf7fc249" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Nov 25 10:49:41 crc kubenswrapper[4821]: I1125 10:49:41.068130 4821 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-67f765d856-ncc6t" podUID="93b4abfa-c14a-4df9-a6d1-9cdeaec918dd" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.149:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.149:8443: connect: connection refused" Nov 25 10:49:43 crc kubenswrapper[4821]: I1125 10:49:43.436216 4821 generic.go:334] "Generic (PLEG): container finished" podID="eb6c9acf-d69c-4788-a85d-b72135de1ef3" containerID="068886c87f6a67c94577ca2971d3df071fe62df65a20f17589ce4f0a9d9008c5" exitCode=0 Nov 25 10:49:43 crc kubenswrapper[4821]: I1125 10:49:43.436660 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-8bp9n" event={"ID":"eb6c9acf-d69c-4788-a85d-b72135de1ef3","Type":"ContainerDied","Data":"068886c87f6a67c94577ca2971d3df071fe62df65a20f17589ce4f0a9d9008c5"} Nov 25 10:49:46 crc kubenswrapper[4821]: I1125 10:49:46.283021 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-8bp9n" Nov 25 10:49:46 crc kubenswrapper[4821]: I1125 10:49:46.297278 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-7znwh" Nov 25 10:49:46 crc kubenswrapper[4821]: I1125 10:49:46.425619 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb6c9acf-d69c-4788-a85d-b72135de1ef3-combined-ca-bundle\") pod \"eb6c9acf-d69c-4788-a85d-b72135de1ef3\" (UID: \"eb6c9acf-d69c-4788-a85d-b72135de1ef3\") " Nov 25 10:49:46 crc kubenswrapper[4821]: I1125 10:49:46.425840 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2eb6852b-84a1-4d4a-8402-77a7e2252ab2-db-sync-config-data\") pod \"2eb6852b-84a1-4d4a-8402-77a7e2252ab2\" (UID: \"2eb6852b-84a1-4d4a-8402-77a7e2252ab2\") " Nov 25 10:49:46 crc kubenswrapper[4821]: I1125 10:49:46.426138 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rfcbt\" (UniqueName: \"kubernetes.io/projected/eb6c9acf-d69c-4788-a85d-b72135de1ef3-kube-api-access-rfcbt\") pod \"eb6c9acf-d69c-4788-a85d-b72135de1ef3\" (UID: \"eb6c9acf-d69c-4788-a85d-b72135de1ef3\") " Nov 25 10:49:46 crc kubenswrapper[4821]: I1125 10:49:46.426660 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/eb6c9acf-d69c-4788-a85d-b72135de1ef3-db-sync-config-data\") pod \"eb6c9acf-d69c-4788-a85d-b72135de1ef3\" (UID: \"eb6c9acf-d69c-4788-a85d-b72135de1ef3\") " Nov 25 10:49:46 crc kubenswrapper[4821]: I1125 10:49:46.427203 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/eb6c9acf-d69c-4788-a85d-b72135de1ef3-etc-machine-id\") pod \"eb6c9acf-d69c-4788-a85d-b72135de1ef3\" (UID: \"eb6c9acf-d69c-4788-a85d-b72135de1ef3\") " Nov 25 10:49:46 crc kubenswrapper[4821]: I1125 10:49:46.427338 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eb6c9acf-d69c-4788-a85d-b72135de1ef3-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "eb6c9acf-d69c-4788-a85d-b72135de1ef3" (UID: "eb6c9acf-d69c-4788-a85d-b72135de1ef3"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:49:46 crc kubenswrapper[4821]: I1125 10:49:46.427621 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l58nm\" (UniqueName: \"kubernetes.io/projected/2eb6852b-84a1-4d4a-8402-77a7e2252ab2-kube-api-access-l58nm\") pod \"2eb6852b-84a1-4d4a-8402-77a7e2252ab2\" (UID: \"2eb6852b-84a1-4d4a-8402-77a7e2252ab2\") " Nov 25 10:49:46 crc kubenswrapper[4821]: I1125 10:49:46.427749 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb6c9acf-d69c-4788-a85d-b72135de1ef3-scripts\") pod \"eb6c9acf-d69c-4788-a85d-b72135de1ef3\" (UID: \"eb6c9acf-d69c-4788-a85d-b72135de1ef3\") " Nov 25 10:49:46 crc kubenswrapper[4821]: I1125 10:49:46.427789 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb6c9acf-d69c-4788-a85d-b72135de1ef3-config-data\") pod \"eb6c9acf-d69c-4788-a85d-b72135de1ef3\" (UID: \"eb6c9acf-d69c-4788-a85d-b72135de1ef3\") " Nov 25 10:49:46 crc kubenswrapper[4821]: I1125 10:49:46.427848 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2eb6852b-84a1-4d4a-8402-77a7e2252ab2-combined-ca-bundle\") pod \"2eb6852b-84a1-4d4a-8402-77a7e2252ab2\" (UID: \"2eb6852b-84a1-4d4a-8402-77a7e2252ab2\") " Nov 25 10:49:46 crc kubenswrapper[4821]: I1125 10:49:46.428618 4821 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/eb6c9acf-d69c-4788-a85d-b72135de1ef3-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:46 crc kubenswrapper[4821]: I1125 10:49:46.433906 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb6c9acf-d69c-4788-a85d-b72135de1ef3-kube-api-access-rfcbt" (OuterVolumeSpecName: "kube-api-access-rfcbt") pod "eb6c9acf-d69c-4788-a85d-b72135de1ef3" (UID: "eb6c9acf-d69c-4788-a85d-b72135de1ef3"). InnerVolumeSpecName "kube-api-access-rfcbt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:49:46 crc kubenswrapper[4821]: I1125 10:49:46.435389 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2eb6852b-84a1-4d4a-8402-77a7e2252ab2-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "2eb6852b-84a1-4d4a-8402-77a7e2252ab2" (UID: "2eb6852b-84a1-4d4a-8402-77a7e2252ab2"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:49:46 crc kubenswrapper[4821]: I1125 10:49:46.436512 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb6c9acf-d69c-4788-a85d-b72135de1ef3-scripts" (OuterVolumeSpecName: "scripts") pod "eb6c9acf-d69c-4788-a85d-b72135de1ef3" (UID: "eb6c9acf-d69c-4788-a85d-b72135de1ef3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:49:46 crc kubenswrapper[4821]: I1125 10:49:46.438310 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb6c9acf-d69c-4788-a85d-b72135de1ef3-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "eb6c9acf-d69c-4788-a85d-b72135de1ef3" (UID: "eb6c9acf-d69c-4788-a85d-b72135de1ef3"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:49:46 crc kubenswrapper[4821]: I1125 10:49:46.446360 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2eb6852b-84a1-4d4a-8402-77a7e2252ab2-kube-api-access-l58nm" (OuterVolumeSpecName: "kube-api-access-l58nm") pod "2eb6852b-84a1-4d4a-8402-77a7e2252ab2" (UID: "2eb6852b-84a1-4d4a-8402-77a7e2252ab2"). InnerVolumeSpecName "kube-api-access-l58nm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:49:46 crc kubenswrapper[4821]: I1125 10:49:46.468640 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-7znwh" event={"ID":"2eb6852b-84a1-4d4a-8402-77a7e2252ab2","Type":"ContainerDied","Data":"f72d78a2fdc159af1f6b1a9f9d141b637201e9ebc83f8e48b6f743759e1b04fd"} Nov 25 10:49:46 crc kubenswrapper[4821]: I1125 10:49:46.468870 4821 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f72d78a2fdc159af1f6b1a9f9d141b637201e9ebc83f8e48b6f743759e1b04fd" Nov 25 10:49:46 crc kubenswrapper[4821]: I1125 10:49:46.468996 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-7znwh" Nov 25 10:49:46 crc kubenswrapper[4821]: I1125 10:49:46.469880 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2eb6852b-84a1-4d4a-8402-77a7e2252ab2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2eb6852b-84a1-4d4a-8402-77a7e2252ab2" (UID: "2eb6852b-84a1-4d4a-8402-77a7e2252ab2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:49:46 crc kubenswrapper[4821]: I1125 10:49:46.481570 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-8bp9n" event={"ID":"eb6c9acf-d69c-4788-a85d-b72135de1ef3","Type":"ContainerDied","Data":"9c4305acf12d8d1e49951f88e49be96bacbd6239479548174634d1401b0bf93f"} Nov 25 10:49:46 crc kubenswrapper[4821]: I1125 10:49:46.481784 4821 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9c4305acf12d8d1e49951f88e49be96bacbd6239479548174634d1401b0bf93f" Nov 25 10:49:46 crc kubenswrapper[4821]: I1125 10:49:46.481908 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-8bp9n" Nov 25 10:49:46 crc kubenswrapper[4821]: I1125 10:49:46.489589 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb6c9acf-d69c-4788-a85d-b72135de1ef3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "eb6c9acf-d69c-4788-a85d-b72135de1ef3" (UID: "eb6c9acf-d69c-4788-a85d-b72135de1ef3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:49:46 crc kubenswrapper[4821]: I1125 10:49:46.511974 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb6c9acf-d69c-4788-a85d-b72135de1ef3-config-data" (OuterVolumeSpecName: "config-data") pod "eb6c9acf-d69c-4788-a85d-b72135de1ef3" (UID: "eb6c9acf-d69c-4788-a85d-b72135de1ef3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:49:46 crc kubenswrapper[4821]: I1125 10:49:46.530433 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rfcbt\" (UniqueName: \"kubernetes.io/projected/eb6c9acf-d69c-4788-a85d-b72135de1ef3-kube-api-access-rfcbt\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:46 crc kubenswrapper[4821]: I1125 10:49:46.530465 4821 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/eb6c9acf-d69c-4788-a85d-b72135de1ef3-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:46 crc kubenswrapper[4821]: I1125 10:49:46.530475 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l58nm\" (UniqueName: \"kubernetes.io/projected/2eb6852b-84a1-4d4a-8402-77a7e2252ab2-kube-api-access-l58nm\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:46 crc kubenswrapper[4821]: I1125 10:49:46.530484 4821 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb6c9acf-d69c-4788-a85d-b72135de1ef3-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:46 crc kubenswrapper[4821]: I1125 10:49:46.530493 4821 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb6c9acf-d69c-4788-a85d-b72135de1ef3-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:46 crc kubenswrapper[4821]: I1125 10:49:46.530502 4821 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2eb6852b-84a1-4d4a-8402-77a7e2252ab2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:46 crc kubenswrapper[4821]: I1125 10:49:46.530509 4821 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb6c9acf-d69c-4788-a85d-b72135de1ef3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:46 crc kubenswrapper[4821]: I1125 10:49:46.530521 4821 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2eb6852b-84a1-4d4a-8402-77a7e2252ab2-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:46 crc kubenswrapper[4821]: I1125 10:49:46.808762 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-55575b4d57-6t4kz"] Nov 25 10:49:46 crc kubenswrapper[4821]: I1125 10:49:46.835454 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6b7b667979-dpjr6" Nov 25 10:49:46 crc kubenswrapper[4821]: I1125 10:49:46.918180 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-zbmpp"] Nov 25 10:49:46 crc kubenswrapper[4821]: I1125 10:49:46.918523 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-56df8fb6b7-zbmpp" podUID="d7c7075f-28ad-46d5-b301-98f3868873db" containerName="dnsmasq-dns" containerID="cri-o://369a7f9ccf800c63e11bd5c783755f2d86cc3c245075d4e9988f5bfb73f07b61" gracePeriod=10 Nov 25 10:49:47 crc kubenswrapper[4821]: I1125 10:49:47.494307 4821 generic.go:334] "Generic (PLEG): container finished" podID="d7c7075f-28ad-46d5-b301-98f3868873db" containerID="369a7f9ccf800c63e11bd5c783755f2d86cc3c245075d4e9988f5bfb73f07b61" exitCode=0 Nov 25 10:49:47 crc kubenswrapper[4821]: I1125 10:49:47.494358 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-zbmpp" event={"ID":"d7c7075f-28ad-46d5-b301-98f3868873db","Type":"ContainerDied","Data":"369a7f9ccf800c63e11bd5c783755f2d86cc3c245075d4e9988f5bfb73f07b61"} Nov 25 10:49:47 crc kubenswrapper[4821]: I1125 10:49:47.605106 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 25 10:49:47 crc kubenswrapper[4821]: E1125 10:49:47.608991 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb6c9acf-d69c-4788-a85d-b72135de1ef3" containerName="cinder-db-sync" Nov 25 10:49:47 crc kubenswrapper[4821]: I1125 10:49:47.609315 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb6c9acf-d69c-4788-a85d-b72135de1ef3" containerName="cinder-db-sync" Nov 25 10:49:47 crc kubenswrapper[4821]: E1125 10:49:47.609347 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2eb6852b-84a1-4d4a-8402-77a7e2252ab2" containerName="barbican-db-sync" Nov 25 10:49:47 crc kubenswrapper[4821]: I1125 10:49:47.609356 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="2eb6852b-84a1-4d4a-8402-77a7e2252ab2" containerName="barbican-db-sync" Nov 25 10:49:47 crc kubenswrapper[4821]: I1125 10:49:47.609610 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="2eb6852b-84a1-4d4a-8402-77a7e2252ab2" containerName="barbican-db-sync" Nov 25 10:49:47 crc kubenswrapper[4821]: I1125 10:49:47.609626 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb6c9acf-d69c-4788-a85d-b72135de1ef3" containerName="cinder-db-sync" Nov 25 10:49:47 crc kubenswrapper[4821]: I1125 10:49:47.612603 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 25 10:49:47 crc kubenswrapper[4821]: I1125 10:49:47.649339 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 25 10:49:47 crc kubenswrapper[4821]: I1125 10:49:47.649577 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-zf45b" Nov 25 10:49:47 crc kubenswrapper[4821]: I1125 10:49:47.649722 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 25 10:49:47 crc kubenswrapper[4821]: I1125 10:49:47.650008 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 25 10:49:47 crc kubenswrapper[4821]: I1125 10:49:47.651714 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4f5bff28-70e9-4b56-ab38-49511f85852d-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"4f5bff28-70e9-4b56-ab38-49511f85852d\") " pod="openstack/cinder-scheduler-0" Nov 25 10:49:47 crc kubenswrapper[4821]: I1125 10:49:47.651784 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4f5bff28-70e9-4b56-ab38-49511f85852d-scripts\") pod \"cinder-scheduler-0\" (UID: \"4f5bff28-70e9-4b56-ab38-49511f85852d\") " pod="openstack/cinder-scheduler-0" Nov 25 10:49:47 crc kubenswrapper[4821]: I1125 10:49:47.651849 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4f5bff28-70e9-4b56-ab38-49511f85852d-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"4f5bff28-70e9-4b56-ab38-49511f85852d\") " pod="openstack/cinder-scheduler-0" Nov 25 10:49:47 crc kubenswrapper[4821]: I1125 10:49:47.651953 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvqhz\" (UniqueName: \"kubernetes.io/projected/4f5bff28-70e9-4b56-ab38-49511f85852d-kube-api-access-vvqhz\") pod \"cinder-scheduler-0\" (UID: \"4f5bff28-70e9-4b56-ab38-49511f85852d\") " pod="openstack/cinder-scheduler-0" Nov 25 10:49:47 crc kubenswrapper[4821]: I1125 10:49:47.651995 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f5bff28-70e9-4b56-ab38-49511f85852d-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"4f5bff28-70e9-4b56-ab38-49511f85852d\") " pod="openstack/cinder-scheduler-0" Nov 25 10:49:47 crc kubenswrapper[4821]: I1125 10:49:47.652046 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f5bff28-70e9-4b56-ab38-49511f85852d-config-data\") pod \"cinder-scheduler-0\" (UID: \"4f5bff28-70e9-4b56-ab38-49511f85852d\") " pod="openstack/cinder-scheduler-0" Nov 25 10:49:47 crc kubenswrapper[4821]: I1125 10:49:47.686528 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-f75f66468-2k6lr"] Nov 25 10:49:47 crc kubenswrapper[4821]: I1125 10:49:47.688346 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-f75f66468-2k6lr" Nov 25 10:49:47 crc kubenswrapper[4821]: I1125 10:49:47.709501 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 25 10:49:47 crc kubenswrapper[4821]: I1125 10:49:47.709801 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Nov 25 10:49:47 crc kubenswrapper[4821]: I1125 10:49:47.709942 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-r9x5r" Nov 25 10:49:47 crc kubenswrapper[4821]: I1125 10:49:47.738318 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-5c9c8c87dc-mqk8l"] Nov 25 10:49:47 crc kubenswrapper[4821]: I1125 10:49:47.745884 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5c9c8c87dc-mqk8l" Nov 25 10:49:47 crc kubenswrapper[4821]: I1125 10:49:47.751476 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Nov 25 10:49:47 crc kubenswrapper[4821]: I1125 10:49:47.755737 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4f5bff28-70e9-4b56-ab38-49511f85852d-scripts\") pod \"cinder-scheduler-0\" (UID: \"4f5bff28-70e9-4b56-ab38-49511f85852d\") " pod="openstack/cinder-scheduler-0" Nov 25 10:49:47 crc kubenswrapper[4821]: I1125 10:49:47.755789 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1e6fee50-9d84-4965-a222-a55f6dcfc4cc-config-data-custom\") pod \"barbican-keystone-listener-f75f66468-2k6lr\" (UID: \"1e6fee50-9d84-4965-a222-a55f6dcfc4cc\") " pod="openstack/barbican-keystone-listener-f75f66468-2k6lr" Nov 25 10:49:47 crc kubenswrapper[4821]: I1125 10:49:47.755836 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e6fee50-9d84-4965-a222-a55f6dcfc4cc-combined-ca-bundle\") pod \"barbican-keystone-listener-f75f66468-2k6lr\" (UID: \"1e6fee50-9d84-4965-a222-a55f6dcfc4cc\") " pod="openstack/barbican-keystone-listener-f75f66468-2k6lr" Nov 25 10:49:47 crc kubenswrapper[4821]: I1125 10:49:47.755876 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w6t9n\" (UniqueName: \"kubernetes.io/projected/1e6fee50-9d84-4965-a222-a55f6dcfc4cc-kube-api-access-w6t9n\") pod \"barbican-keystone-listener-f75f66468-2k6lr\" (UID: \"1e6fee50-9d84-4965-a222-a55f6dcfc4cc\") " pod="openstack/barbican-keystone-listener-f75f66468-2k6lr" Nov 25 10:49:47 crc kubenswrapper[4821]: I1125 10:49:47.755910 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4f5bff28-70e9-4b56-ab38-49511f85852d-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"4f5bff28-70e9-4b56-ab38-49511f85852d\") " pod="openstack/cinder-scheduler-0" Nov 25 10:49:47 crc kubenswrapper[4821]: I1125 10:49:47.764195 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e6fee50-9d84-4965-a222-a55f6dcfc4cc-config-data\") pod \"barbican-keystone-listener-f75f66468-2k6lr\" (UID: \"1e6fee50-9d84-4965-a222-a55f6dcfc4cc\") " pod="openstack/barbican-keystone-listener-f75f66468-2k6lr" Nov 25 10:49:47 crc kubenswrapper[4821]: I1125 10:49:47.764332 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1e6fee50-9d84-4965-a222-a55f6dcfc4cc-logs\") pod \"barbican-keystone-listener-f75f66468-2k6lr\" (UID: \"1e6fee50-9d84-4965-a222-a55f6dcfc4cc\") " pod="openstack/barbican-keystone-listener-f75f66468-2k6lr" Nov 25 10:49:47 crc kubenswrapper[4821]: I1125 10:49:47.764442 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvqhz\" (UniqueName: \"kubernetes.io/projected/4f5bff28-70e9-4b56-ab38-49511f85852d-kube-api-access-vvqhz\") pod \"cinder-scheduler-0\" (UID: \"4f5bff28-70e9-4b56-ab38-49511f85852d\") " pod="openstack/cinder-scheduler-0" Nov 25 10:49:47 crc kubenswrapper[4821]: I1125 10:49:47.764496 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f5bff28-70e9-4b56-ab38-49511f85852d-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"4f5bff28-70e9-4b56-ab38-49511f85852d\") " pod="openstack/cinder-scheduler-0" Nov 25 10:49:47 crc kubenswrapper[4821]: I1125 10:49:47.764568 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f5bff28-70e9-4b56-ab38-49511f85852d-config-data\") pod \"cinder-scheduler-0\" (UID: \"4f5bff28-70e9-4b56-ab38-49511f85852d\") " pod="openstack/cinder-scheduler-0" Nov 25 10:49:47 crc kubenswrapper[4821]: I1125 10:49:47.764633 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4f5bff28-70e9-4b56-ab38-49511f85852d-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"4f5bff28-70e9-4b56-ab38-49511f85852d\") " pod="openstack/cinder-scheduler-0" Nov 25 10:49:47 crc kubenswrapper[4821]: I1125 10:49:47.767552 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4f5bff28-70e9-4b56-ab38-49511f85852d-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"4f5bff28-70e9-4b56-ab38-49511f85852d\") " pod="openstack/cinder-scheduler-0" Nov 25 10:49:47 crc kubenswrapper[4821]: I1125 10:49:47.778351 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f5bff28-70e9-4b56-ab38-49511f85852d-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"4f5bff28-70e9-4b56-ab38-49511f85852d\") " pod="openstack/cinder-scheduler-0" Nov 25 10:49:47 crc kubenswrapper[4821]: I1125 10:49:47.778382 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4f5bff28-70e9-4b56-ab38-49511f85852d-scripts\") pod \"cinder-scheduler-0\" (UID: \"4f5bff28-70e9-4b56-ab38-49511f85852d\") " pod="openstack/cinder-scheduler-0" Nov 25 10:49:47 crc kubenswrapper[4821]: I1125 10:49:47.779389 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f5bff28-70e9-4b56-ab38-49511f85852d-config-data\") pod \"cinder-scheduler-0\" (UID: \"4f5bff28-70e9-4b56-ab38-49511f85852d\") " pod="openstack/cinder-scheduler-0" Nov 25 10:49:47 crc kubenswrapper[4821]: I1125 10:49:47.782812 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4f5bff28-70e9-4b56-ab38-49511f85852d-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"4f5bff28-70e9-4b56-ab38-49511f85852d\") " pod="openstack/cinder-scheduler-0" Nov 25 10:49:47 crc kubenswrapper[4821]: I1125 10:49:47.794190 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 25 10:49:47 crc kubenswrapper[4821]: I1125 10:49:47.820754 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-f75f66468-2k6lr"] Nov 25 10:49:47 crc kubenswrapper[4821]: I1125 10:49:47.821872 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvqhz\" (UniqueName: \"kubernetes.io/projected/4f5bff28-70e9-4b56-ab38-49511f85852d-kube-api-access-vvqhz\") pod \"cinder-scheduler-0\" (UID: \"4f5bff28-70e9-4b56-ab38-49511f85852d\") " pod="openstack/cinder-scheduler-0" Nov 25 10:49:47 crc kubenswrapper[4821]: I1125 10:49:47.827054 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 25 10:49:47 crc kubenswrapper[4821]: I1125 10:49:47.863486 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5c9c8c87dc-mqk8l"] Nov 25 10:49:47 crc kubenswrapper[4821]: I1125 10:49:47.872344 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5b7f\" (UniqueName: \"kubernetes.io/projected/62fc74a6-526c-4953-8439-b4af8ce5f39b-kube-api-access-p5b7f\") pod \"barbican-worker-5c9c8c87dc-mqk8l\" (UID: \"62fc74a6-526c-4953-8439-b4af8ce5f39b\") " pod="openstack/barbican-worker-5c9c8c87dc-mqk8l" Nov 25 10:49:47 crc kubenswrapper[4821]: I1125 10:49:47.880898 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/62fc74a6-526c-4953-8439-b4af8ce5f39b-config-data-custom\") pod \"barbican-worker-5c9c8c87dc-mqk8l\" (UID: \"62fc74a6-526c-4953-8439-b4af8ce5f39b\") " pod="openstack/barbican-worker-5c9c8c87dc-mqk8l" Nov 25 10:49:47 crc kubenswrapper[4821]: I1125 10:49:47.881349 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1e6fee50-9d84-4965-a222-a55f6dcfc4cc-config-data-custom\") pod \"barbican-keystone-listener-f75f66468-2k6lr\" (UID: \"1e6fee50-9d84-4965-a222-a55f6dcfc4cc\") " pod="openstack/barbican-keystone-listener-f75f66468-2k6lr" Nov 25 10:49:47 crc kubenswrapper[4821]: I1125 10:49:47.881660 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e6fee50-9d84-4965-a222-a55f6dcfc4cc-combined-ca-bundle\") pod \"barbican-keystone-listener-f75f66468-2k6lr\" (UID: \"1e6fee50-9d84-4965-a222-a55f6dcfc4cc\") " pod="openstack/barbican-keystone-listener-f75f66468-2k6lr" Nov 25 10:49:47 crc kubenswrapper[4821]: I1125 10:49:47.881852 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w6t9n\" (UniqueName: \"kubernetes.io/projected/1e6fee50-9d84-4965-a222-a55f6dcfc4cc-kube-api-access-w6t9n\") pod \"barbican-keystone-listener-f75f66468-2k6lr\" (UID: \"1e6fee50-9d84-4965-a222-a55f6dcfc4cc\") " pod="openstack/barbican-keystone-listener-f75f66468-2k6lr" Nov 25 10:49:47 crc kubenswrapper[4821]: I1125 10:49:47.881985 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e6fee50-9d84-4965-a222-a55f6dcfc4cc-config-data\") pod \"barbican-keystone-listener-f75f66468-2k6lr\" (UID: \"1e6fee50-9d84-4965-a222-a55f6dcfc4cc\") " pod="openstack/barbican-keystone-listener-f75f66468-2k6lr" Nov 25 10:49:47 crc kubenswrapper[4821]: I1125 10:49:47.882025 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62fc74a6-526c-4953-8439-b4af8ce5f39b-combined-ca-bundle\") pod \"barbican-worker-5c9c8c87dc-mqk8l\" (UID: \"62fc74a6-526c-4953-8439-b4af8ce5f39b\") " pod="openstack/barbican-worker-5c9c8c87dc-mqk8l" Nov 25 10:49:47 crc kubenswrapper[4821]: I1125 10:49:47.882085 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62fc74a6-526c-4953-8439-b4af8ce5f39b-config-data\") pod \"barbican-worker-5c9c8c87dc-mqk8l\" (UID: \"62fc74a6-526c-4953-8439-b4af8ce5f39b\") " pod="openstack/barbican-worker-5c9c8c87dc-mqk8l" Nov 25 10:49:47 crc kubenswrapper[4821]: I1125 10:49:47.882217 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1e6fee50-9d84-4965-a222-a55f6dcfc4cc-logs\") pod \"barbican-keystone-listener-f75f66468-2k6lr\" (UID: \"1e6fee50-9d84-4965-a222-a55f6dcfc4cc\") " pod="openstack/barbican-keystone-listener-f75f66468-2k6lr" Nov 25 10:49:47 crc kubenswrapper[4821]: I1125 10:49:47.882314 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62fc74a6-526c-4953-8439-b4af8ce5f39b-logs\") pod \"barbican-worker-5c9c8c87dc-mqk8l\" (UID: \"62fc74a6-526c-4953-8439-b4af8ce5f39b\") " pod="openstack/barbican-worker-5c9c8c87dc-mqk8l" Nov 25 10:49:47 crc kubenswrapper[4821]: I1125 10:49:47.889887 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1e6fee50-9d84-4965-a222-a55f6dcfc4cc-config-data-custom\") pod \"barbican-keystone-listener-f75f66468-2k6lr\" (UID: \"1e6fee50-9d84-4965-a222-a55f6dcfc4cc\") " pod="openstack/barbican-keystone-listener-f75f66468-2k6lr" Nov 25 10:49:47 crc kubenswrapper[4821]: I1125 10:49:47.893956 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1e6fee50-9d84-4965-a222-a55f6dcfc4cc-logs\") pod \"barbican-keystone-listener-f75f66468-2k6lr\" (UID: \"1e6fee50-9d84-4965-a222-a55f6dcfc4cc\") " pod="openstack/barbican-keystone-listener-f75f66468-2k6lr" Nov 25 10:49:47 crc kubenswrapper[4821]: I1125 10:49:47.931463 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e6fee50-9d84-4965-a222-a55f6dcfc4cc-config-data\") pod \"barbican-keystone-listener-f75f66468-2k6lr\" (UID: \"1e6fee50-9d84-4965-a222-a55f6dcfc4cc\") " pod="openstack/barbican-keystone-listener-f75f66468-2k6lr" Nov 25 10:49:47 crc kubenswrapper[4821]: I1125 10:49:47.933304 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w6t9n\" (UniqueName: \"kubernetes.io/projected/1e6fee50-9d84-4965-a222-a55f6dcfc4cc-kube-api-access-w6t9n\") pod \"barbican-keystone-listener-f75f66468-2k6lr\" (UID: \"1e6fee50-9d84-4965-a222-a55f6dcfc4cc\") " pod="openstack/barbican-keystone-listener-f75f66468-2k6lr" Nov 25 10:49:47 crc kubenswrapper[4821]: I1125 10:49:47.938916 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e6fee50-9d84-4965-a222-a55f6dcfc4cc-combined-ca-bundle\") pod \"barbican-keystone-listener-f75f66468-2k6lr\" (UID: \"1e6fee50-9d84-4965-a222-a55f6dcfc4cc\") " pod="openstack/barbican-keystone-listener-f75f66468-2k6lr" Nov 25 10:49:47 crc kubenswrapper[4821]: I1125 10:49:47.938998 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-bmmvh"] Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.023950 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-bmmvh" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.027109 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62fc74a6-526c-4953-8439-b4af8ce5f39b-combined-ca-bundle\") pod \"barbican-worker-5c9c8c87dc-mqk8l\" (UID: \"62fc74a6-526c-4953-8439-b4af8ce5f39b\") " pod="openstack/barbican-worker-5c9c8c87dc-mqk8l" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.027181 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62fc74a6-526c-4953-8439-b4af8ce5f39b-config-data\") pod \"barbican-worker-5c9c8c87dc-mqk8l\" (UID: \"62fc74a6-526c-4953-8439-b4af8ce5f39b\") " pod="openstack/barbican-worker-5c9c8c87dc-mqk8l" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.027285 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62fc74a6-526c-4953-8439-b4af8ce5f39b-logs\") pod \"barbican-worker-5c9c8c87dc-mqk8l\" (UID: \"62fc74a6-526c-4953-8439-b4af8ce5f39b\") " pod="openstack/barbican-worker-5c9c8c87dc-mqk8l" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.027337 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p5b7f\" (UniqueName: \"kubernetes.io/projected/62fc74a6-526c-4953-8439-b4af8ce5f39b-kube-api-access-p5b7f\") pod \"barbican-worker-5c9c8c87dc-mqk8l\" (UID: \"62fc74a6-526c-4953-8439-b4af8ce5f39b\") " pod="openstack/barbican-worker-5c9c8c87dc-mqk8l" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.027383 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/62fc74a6-526c-4953-8439-b4af8ce5f39b-config-data-custom\") pod \"barbican-worker-5c9c8c87dc-mqk8l\" (UID: \"62fc74a6-526c-4953-8439-b4af8ce5f39b\") " pod="openstack/barbican-worker-5c9c8c87dc-mqk8l" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.033884 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62fc74a6-526c-4953-8439-b4af8ce5f39b-logs\") pod \"barbican-worker-5c9c8c87dc-mqk8l\" (UID: \"62fc74a6-526c-4953-8439-b4af8ce5f39b\") " pod="openstack/barbican-worker-5c9c8c87dc-mqk8l" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.061955 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/62fc74a6-526c-4953-8439-b4af8ce5f39b-config-data-custom\") pod \"barbican-worker-5c9c8c87dc-mqk8l\" (UID: \"62fc74a6-526c-4953-8439-b4af8ce5f39b\") " pod="openstack/barbican-worker-5c9c8c87dc-mqk8l" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.097251 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62fc74a6-526c-4953-8439-b4af8ce5f39b-config-data\") pod \"barbican-worker-5c9c8c87dc-mqk8l\" (UID: \"62fc74a6-526c-4953-8439-b4af8ce5f39b\") " pod="openstack/barbican-worker-5c9c8c87dc-mqk8l" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.098585 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62fc74a6-526c-4953-8439-b4af8ce5f39b-combined-ca-bundle\") pod \"barbican-worker-5c9c8c87dc-mqk8l\" (UID: \"62fc74a6-526c-4953-8439-b4af8ce5f39b\") " pod="openstack/barbican-worker-5c9c8c87dc-mqk8l" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.101379 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5b7f\" (UniqueName: \"kubernetes.io/projected/62fc74a6-526c-4953-8439-b4af8ce5f39b-kube-api-access-p5b7f\") pod \"barbican-worker-5c9c8c87dc-mqk8l\" (UID: \"62fc74a6-526c-4953-8439-b4af8ce5f39b\") " pod="openstack/barbican-worker-5c9c8c87dc-mqk8l" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.144955 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2fda2637-8b50-43ee-a1dc-6d6c3340030c-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-bmmvh\" (UID: \"2fda2637-8b50-43ee-a1dc-6d6c3340030c\") " pod="openstack/dnsmasq-dns-848cf88cfc-bmmvh" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.145054 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2fda2637-8b50-43ee-a1dc-6d6c3340030c-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-bmmvh\" (UID: \"2fda2637-8b50-43ee-a1dc-6d6c3340030c\") " pod="openstack/dnsmasq-dns-848cf88cfc-bmmvh" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.145145 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2fda2637-8b50-43ee-a1dc-6d6c3340030c-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-bmmvh\" (UID: \"2fda2637-8b50-43ee-a1dc-6d6c3340030c\") " pod="openstack/dnsmasq-dns-848cf88cfc-bmmvh" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.145238 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4l5q7\" (UniqueName: \"kubernetes.io/projected/2fda2637-8b50-43ee-a1dc-6d6c3340030c-kube-api-access-4l5q7\") pod \"dnsmasq-dns-848cf88cfc-bmmvh\" (UID: \"2fda2637-8b50-43ee-a1dc-6d6c3340030c\") " pod="openstack/dnsmasq-dns-848cf88cfc-bmmvh" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.145288 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2fda2637-8b50-43ee-a1dc-6d6c3340030c-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-bmmvh\" (UID: \"2fda2637-8b50-43ee-a1dc-6d6c3340030c\") " pod="openstack/dnsmasq-dns-848cf88cfc-bmmvh" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.145330 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fda2637-8b50-43ee-a1dc-6d6c3340030c-config\") pod \"dnsmasq-dns-848cf88cfc-bmmvh\" (UID: \"2fda2637-8b50-43ee-a1dc-6d6c3340030c\") " pod="openstack/dnsmasq-dns-848cf88cfc-bmmvh" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.158044 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-bmmvh"] Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.185363 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-bmmvh"] Nov 25 10:49:48 crc kubenswrapper[4821]: E1125 10:49:48.186018 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc dns-swift-storage-0 kube-api-access-4l5q7 ovsdbserver-nb ovsdbserver-sb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-848cf88cfc-bmmvh" podUID="2fda2637-8b50-43ee-a1dc-6d6c3340030c" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.201839 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-f75f66468-2k6lr" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.205250 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-55c965854d-xpf4q"] Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.207544 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-55c965854d-xpf4q" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.212915 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.218705 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-zbmpp" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.240131 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-55c965854d-xpf4q"] Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.245906 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d7c7075f-28ad-46d5-b301-98f3868873db-dns-svc\") pod \"d7c7075f-28ad-46d5-b301-98f3868873db\" (UID: \"d7c7075f-28ad-46d5-b301-98f3868873db\") " Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.245961 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d7c7075f-28ad-46d5-b301-98f3868873db-ovsdbserver-sb\") pod \"d7c7075f-28ad-46d5-b301-98f3868873db\" (UID: \"d7c7075f-28ad-46d5-b301-98f3868873db\") " Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.246047 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7c7075f-28ad-46d5-b301-98f3868873db-config\") pod \"d7c7075f-28ad-46d5-b301-98f3868873db\" (UID: \"d7c7075f-28ad-46d5-b301-98f3868873db\") " Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.246078 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d7c7075f-28ad-46d5-b301-98f3868873db-dns-swift-storage-0\") pod \"d7c7075f-28ad-46d5-b301-98f3868873db\" (UID: \"d7c7075f-28ad-46d5-b301-98f3868873db\") " Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.246115 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vq2kd\" (UniqueName: \"kubernetes.io/projected/d7c7075f-28ad-46d5-b301-98f3868873db-kube-api-access-vq2kd\") pod \"d7c7075f-28ad-46d5-b301-98f3868873db\" (UID: \"d7c7075f-28ad-46d5-b301-98f3868873db\") " Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.246141 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d7c7075f-28ad-46d5-b301-98f3868873db-ovsdbserver-nb\") pod \"d7c7075f-28ad-46d5-b301-98f3868873db\" (UID: \"d7c7075f-28ad-46d5-b301-98f3868873db\") " Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.246323 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2fda2637-8b50-43ee-a1dc-6d6c3340030c-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-bmmvh\" (UID: \"2fda2637-8b50-43ee-a1dc-6d6c3340030c\") " pod="openstack/dnsmasq-dns-848cf88cfc-bmmvh" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.246351 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7-logs\") pod \"barbican-api-55c965854d-xpf4q\" (UID: \"73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7\") " pod="openstack/barbican-api-55c965854d-xpf4q" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.246396 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4l5q7\" (UniqueName: \"kubernetes.io/projected/2fda2637-8b50-43ee-a1dc-6d6c3340030c-kube-api-access-4l5q7\") pod \"dnsmasq-dns-848cf88cfc-bmmvh\" (UID: \"2fda2637-8b50-43ee-a1dc-6d6c3340030c\") " pod="openstack/dnsmasq-dns-848cf88cfc-bmmvh" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.246428 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2fda2637-8b50-43ee-a1dc-6d6c3340030c-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-bmmvh\" (UID: \"2fda2637-8b50-43ee-a1dc-6d6c3340030c\") " pod="openstack/dnsmasq-dns-848cf88cfc-bmmvh" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.246463 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7-config-data-custom\") pod \"barbican-api-55c965854d-xpf4q\" (UID: \"73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7\") " pod="openstack/barbican-api-55c965854d-xpf4q" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.246486 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fda2637-8b50-43ee-a1dc-6d6c3340030c-config\") pod \"dnsmasq-dns-848cf88cfc-bmmvh\" (UID: \"2fda2637-8b50-43ee-a1dc-6d6c3340030c\") " pod="openstack/dnsmasq-dns-848cf88cfc-bmmvh" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.249041 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2fda2637-8b50-43ee-a1dc-6d6c3340030c-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-bmmvh\" (UID: \"2fda2637-8b50-43ee-a1dc-6d6c3340030c\") " pod="openstack/dnsmasq-dns-848cf88cfc-bmmvh" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.250278 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7-combined-ca-bundle\") pod \"barbican-api-55c965854d-xpf4q\" (UID: \"73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7\") " pod="openstack/barbican-api-55c965854d-xpf4q" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.250923 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m64k7\" (UniqueName: \"kubernetes.io/projected/73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7-kube-api-access-m64k7\") pod \"barbican-api-55c965854d-xpf4q\" (UID: \"73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7\") " pod="openstack/barbican-api-55c965854d-xpf4q" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.252256 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fda2637-8b50-43ee-a1dc-6d6c3340030c-config\") pod \"dnsmasq-dns-848cf88cfc-bmmvh\" (UID: \"2fda2637-8b50-43ee-a1dc-6d6c3340030c\") " pod="openstack/dnsmasq-dns-848cf88cfc-bmmvh" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.252966 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2fda2637-8b50-43ee-a1dc-6d6c3340030c-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-bmmvh\" (UID: \"2fda2637-8b50-43ee-a1dc-6d6c3340030c\") " pod="openstack/dnsmasq-dns-848cf88cfc-bmmvh" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.271626 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2fda2637-8b50-43ee-a1dc-6d6c3340030c-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-bmmvh\" (UID: \"2fda2637-8b50-43ee-a1dc-6d6c3340030c\") " pod="openstack/dnsmasq-dns-848cf88cfc-bmmvh" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.271990 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7-config-data\") pod \"barbican-api-55c965854d-xpf4q\" (UID: \"73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7\") " pod="openstack/barbican-api-55c965854d-xpf4q" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.275141 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2fda2637-8b50-43ee-a1dc-6d6c3340030c-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-bmmvh\" (UID: \"2fda2637-8b50-43ee-a1dc-6d6c3340030c\") " pod="openstack/dnsmasq-dns-848cf88cfc-bmmvh" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.276689 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2fda2637-8b50-43ee-a1dc-6d6c3340030c-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-bmmvh\" (UID: \"2fda2637-8b50-43ee-a1dc-6d6c3340030c\") " pod="openstack/dnsmasq-dns-848cf88cfc-bmmvh" Nov 25 10:49:48 crc kubenswrapper[4821]: E1125 10:49:48.283645 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="9cb63eff-3e7e-436e-aea6-ce8254c3db77" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.284147 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7c7075f-28ad-46d5-b301-98f3868873db-kube-api-access-vq2kd" (OuterVolumeSpecName: "kube-api-access-vq2kd") pod "d7c7075f-28ad-46d5-b301-98f3868873db" (UID: "d7c7075f-28ad-46d5-b301-98f3868873db"). InnerVolumeSpecName "kube-api-access-vq2kd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.284586 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5c9c8c87dc-mqk8l" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.296605 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4l5q7\" (UniqueName: \"kubernetes.io/projected/2fda2637-8b50-43ee-a1dc-6d6c3340030c-kube-api-access-4l5q7\") pod \"dnsmasq-dns-848cf88cfc-bmmvh\" (UID: \"2fda2637-8b50-43ee-a1dc-6d6c3340030c\") " pod="openstack/dnsmasq-dns-848cf88cfc-bmmvh" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.306733 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2fda2637-8b50-43ee-a1dc-6d6c3340030c-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-bmmvh\" (UID: \"2fda2637-8b50-43ee-a1dc-6d6c3340030c\") " pod="openstack/dnsmasq-dns-848cf88cfc-bmmvh" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.338943 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7c7075f-28ad-46d5-b301-98f3868873db-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "d7c7075f-28ad-46d5-b301-98f3868873db" (UID: "d7c7075f-28ad-46d5-b301-98f3868873db"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.352024 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-svh8j"] Nov 25 10:49:48 crc kubenswrapper[4821]: E1125 10:49:48.353327 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7c7075f-28ad-46d5-b301-98f3868873db" containerName="init" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.353372 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7c7075f-28ad-46d5-b301-98f3868873db" containerName="init" Nov 25 10:49:48 crc kubenswrapper[4821]: E1125 10:49:48.353393 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7c7075f-28ad-46d5-b301-98f3868873db" containerName="dnsmasq-dns" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.353401 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7c7075f-28ad-46d5-b301-98f3868873db" containerName="dnsmasq-dns" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.353727 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7c7075f-28ad-46d5-b301-98f3868873db" containerName="dnsmasq-dns" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.358104 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-svh8j" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.377747 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7-combined-ca-bundle\") pod \"barbican-api-55c965854d-xpf4q\" (UID: \"73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7\") " pod="openstack/barbican-api-55c965854d-xpf4q" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.377808 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m64k7\" (UniqueName: \"kubernetes.io/projected/73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7-kube-api-access-m64k7\") pod \"barbican-api-55c965854d-xpf4q\" (UID: \"73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7\") " pod="openstack/barbican-api-55c965854d-xpf4q" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.389610 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7-config-data\") pod \"barbican-api-55c965854d-xpf4q\" (UID: \"73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7\") " pod="openstack/barbican-api-55c965854d-xpf4q" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.390241 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7-logs\") pod \"barbican-api-55c965854d-xpf4q\" (UID: \"73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7\") " pod="openstack/barbican-api-55c965854d-xpf4q" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.390707 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7-config-data-custom\") pod \"barbican-api-55c965854d-xpf4q\" (UID: \"73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7\") " pod="openstack/barbican-api-55c965854d-xpf4q" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.390898 4821 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d7c7075f-28ad-46d5-b301-98f3868873db-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.390916 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vq2kd\" (UniqueName: \"kubernetes.io/projected/d7c7075f-28ad-46d5-b301-98f3868873db-kube-api-access-vq2kd\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.392944 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7-logs\") pod \"barbican-api-55c965854d-xpf4q\" (UID: \"73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7\") " pod="openstack/barbican-api-55c965854d-xpf4q" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.400237 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-svh8j"] Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.428923 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7-config-data-custom\") pod \"barbican-api-55c965854d-xpf4q\" (UID: \"73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7\") " pod="openstack/barbican-api-55c965854d-xpf4q" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.429344 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7c7075f-28ad-46d5-b301-98f3868873db-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d7c7075f-28ad-46d5-b301-98f3868873db" (UID: "d7c7075f-28ad-46d5-b301-98f3868873db"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.436219 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.438070 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.439215 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m64k7\" (UniqueName: \"kubernetes.io/projected/73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7-kube-api-access-m64k7\") pod \"barbican-api-55c965854d-xpf4q\" (UID: \"73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7\") " pod="openstack/barbican-api-55c965854d-xpf4q" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.446761 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7c7075f-28ad-46d5-b301-98f3868873db-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d7c7075f-28ad-46d5-b301-98f3868873db" (UID: "d7c7075f-28ad-46d5-b301-98f3868873db"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.446963 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.451113 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7-combined-ca-bundle\") pod \"barbican-api-55c965854d-xpf4q\" (UID: \"73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7\") " pod="openstack/barbican-api-55c965854d-xpf4q" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.452989 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7-config-data\") pod \"barbican-api-55c965854d-xpf4q\" (UID: \"73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7\") " pod="openstack/barbican-api-55c965854d-xpf4q" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.461536 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.492556 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/51b2c969-f09a-4c90-a897-8b6291716ee7-scripts\") pod \"cinder-api-0\" (UID: \"51b2c969-f09a-4c90-a897-8b6291716ee7\") " pod="openstack/cinder-api-0" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.492618 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v86b8\" (UniqueName: \"kubernetes.io/projected/3d48a46a-3ae2-433a-a70b-cdda6d764e53-kube-api-access-v86b8\") pod \"dnsmasq-dns-6578955fd5-svh8j\" (UID: \"3d48a46a-3ae2-433a-a70b-cdda6d764e53\") " pod="openstack/dnsmasq-dns-6578955fd5-svh8j" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.492647 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3d48a46a-3ae2-433a-a70b-cdda6d764e53-dns-svc\") pod \"dnsmasq-dns-6578955fd5-svh8j\" (UID: \"3d48a46a-3ae2-433a-a70b-cdda6d764e53\") " pod="openstack/dnsmasq-dns-6578955fd5-svh8j" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.492714 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51b2c969-f09a-4c90-a897-8b6291716ee7-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"51b2c969-f09a-4c90-a897-8b6291716ee7\") " pod="openstack/cinder-api-0" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.492745 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/51b2c969-f09a-4c90-a897-8b6291716ee7-etc-machine-id\") pod \"cinder-api-0\" (UID: \"51b2c969-f09a-4c90-a897-8b6291716ee7\") " pod="openstack/cinder-api-0" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.492771 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3d48a46a-3ae2-433a-a70b-cdda6d764e53-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-svh8j\" (UID: \"3d48a46a-3ae2-433a-a70b-cdda6d764e53\") " pod="openstack/dnsmasq-dns-6578955fd5-svh8j" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.492841 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51b2c969-f09a-4c90-a897-8b6291716ee7-config-data\") pod \"cinder-api-0\" (UID: \"51b2c969-f09a-4c90-a897-8b6291716ee7\") " pod="openstack/cinder-api-0" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.492884 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mwxk5\" (UniqueName: \"kubernetes.io/projected/51b2c969-f09a-4c90-a897-8b6291716ee7-kube-api-access-mwxk5\") pod \"cinder-api-0\" (UID: \"51b2c969-f09a-4c90-a897-8b6291716ee7\") " pod="openstack/cinder-api-0" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.492911 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/51b2c969-f09a-4c90-a897-8b6291716ee7-logs\") pod \"cinder-api-0\" (UID: \"51b2c969-f09a-4c90-a897-8b6291716ee7\") " pod="openstack/cinder-api-0" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.492943 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d48a46a-3ae2-433a-a70b-cdda6d764e53-config\") pod \"dnsmasq-dns-6578955fd5-svh8j\" (UID: \"3d48a46a-3ae2-433a-a70b-cdda6d764e53\") " pod="openstack/dnsmasq-dns-6578955fd5-svh8j" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.492968 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3d48a46a-3ae2-433a-a70b-cdda6d764e53-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-svh8j\" (UID: \"3d48a46a-3ae2-433a-a70b-cdda6d764e53\") " pod="openstack/dnsmasq-dns-6578955fd5-svh8j" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.492995 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/51b2c969-f09a-4c90-a897-8b6291716ee7-config-data-custom\") pod \"cinder-api-0\" (UID: \"51b2c969-f09a-4c90-a897-8b6291716ee7\") " pod="openstack/cinder-api-0" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.493029 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3d48a46a-3ae2-433a-a70b-cdda6d764e53-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-svh8j\" (UID: \"3d48a46a-3ae2-433a-a70b-cdda6d764e53\") " pod="openstack/dnsmasq-dns-6578955fd5-svh8j" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.493119 4821 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d7c7075f-28ad-46d5-b301-98f3868873db-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.493134 4821 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d7c7075f-28ad-46d5-b301-98f3868873db-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.493178 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7c7075f-28ad-46d5-b301-98f3868873db-config" (OuterVolumeSpecName: "config") pod "d7c7075f-28ad-46d5-b301-98f3868873db" (UID: "d7c7075f-28ad-46d5-b301-98f3868873db"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.510449 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-55575b4d57-6t4kz" event={"ID":"a8419a57-35ff-4340-b2d7-fd11724948fe","Type":"ContainerStarted","Data":"e55e386a3f98c42efc38239d2014686b26451669cce2f0b545387745411a95dd"} Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.510508 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-55575b4d57-6t4kz" event={"ID":"a8419a57-35ff-4340-b2d7-fd11724948fe","Type":"ContainerStarted","Data":"e3bddb537b83e7c402b9ff077be5f86390e55b58e12ec74c1a3e4dafcebf39d9"} Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.523063 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7c7075f-28ad-46d5-b301-98f3868873db-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d7c7075f-28ad-46d5-b301-98f3868873db" (UID: "d7c7075f-28ad-46d5-b301-98f3868873db"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.555557 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-zbmpp" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.555554 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-zbmpp" event={"ID":"d7c7075f-28ad-46d5-b301-98f3868873db","Type":"ContainerDied","Data":"f6cd48561cb54a03ba2af584e88ca05798dac2fdd347d3343872d689cfcdce5a"} Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.556097 4821 scope.go:117] "RemoveContainer" containerID="369a7f9ccf800c63e11bd5c783755f2d86cc3c245075d4e9988f5bfb73f07b61" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.585383 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-bmmvh" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.585383 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9cb63eff-3e7e-436e-aea6-ce8254c3db77","Type":"ContainerStarted","Data":"0764f0b45f4c9a4cc4d1c4342d5f4f2d448117f396e64fdd31e641a7ca96dd03"} Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.586414 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9cb63eff-3e7e-436e-aea6-ce8254c3db77" containerName="ceilometer-notification-agent" containerID="cri-o://cf5c6db7d1dca33b423304d3462893904bd7f34d04492e4bdf0db6099734590c" gracePeriod=30 Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.586574 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9cb63eff-3e7e-436e-aea6-ce8254c3db77" containerName="proxy-httpd" containerID="cri-o://0764f0b45f4c9a4cc4d1c4342d5f4f2d448117f396e64fdd31e641a7ca96dd03" gracePeriod=30 Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.586636 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9cb63eff-3e7e-436e-aea6-ce8254c3db77" containerName="sg-core" containerID="cri-o://666f50101704e276ddca6f07195913aeb8365df95b6639c5717f6c54b431976e" gracePeriod=30 Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.595897 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/51b2c969-f09a-4c90-a897-8b6291716ee7-scripts\") pod \"cinder-api-0\" (UID: \"51b2c969-f09a-4c90-a897-8b6291716ee7\") " pod="openstack/cinder-api-0" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.595967 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v86b8\" (UniqueName: \"kubernetes.io/projected/3d48a46a-3ae2-433a-a70b-cdda6d764e53-kube-api-access-v86b8\") pod \"dnsmasq-dns-6578955fd5-svh8j\" (UID: \"3d48a46a-3ae2-433a-a70b-cdda6d764e53\") " pod="openstack/dnsmasq-dns-6578955fd5-svh8j" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.595999 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3d48a46a-3ae2-433a-a70b-cdda6d764e53-dns-svc\") pod \"dnsmasq-dns-6578955fd5-svh8j\" (UID: \"3d48a46a-3ae2-433a-a70b-cdda6d764e53\") " pod="openstack/dnsmasq-dns-6578955fd5-svh8j" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.596037 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51b2c969-f09a-4c90-a897-8b6291716ee7-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"51b2c969-f09a-4c90-a897-8b6291716ee7\") " pod="openstack/cinder-api-0" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.596076 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/51b2c969-f09a-4c90-a897-8b6291716ee7-etc-machine-id\") pod \"cinder-api-0\" (UID: \"51b2c969-f09a-4c90-a897-8b6291716ee7\") " pod="openstack/cinder-api-0" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.596107 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3d48a46a-3ae2-433a-a70b-cdda6d764e53-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-svh8j\" (UID: \"3d48a46a-3ae2-433a-a70b-cdda6d764e53\") " pod="openstack/dnsmasq-dns-6578955fd5-svh8j" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.596208 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51b2c969-f09a-4c90-a897-8b6291716ee7-config-data\") pod \"cinder-api-0\" (UID: \"51b2c969-f09a-4c90-a897-8b6291716ee7\") " pod="openstack/cinder-api-0" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.596316 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mwxk5\" (UniqueName: \"kubernetes.io/projected/51b2c969-f09a-4c90-a897-8b6291716ee7-kube-api-access-mwxk5\") pod \"cinder-api-0\" (UID: \"51b2c969-f09a-4c90-a897-8b6291716ee7\") " pod="openstack/cinder-api-0" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.596382 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/51b2c969-f09a-4c90-a897-8b6291716ee7-logs\") pod \"cinder-api-0\" (UID: \"51b2c969-f09a-4c90-a897-8b6291716ee7\") " pod="openstack/cinder-api-0" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.596453 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d48a46a-3ae2-433a-a70b-cdda6d764e53-config\") pod \"dnsmasq-dns-6578955fd5-svh8j\" (UID: \"3d48a46a-3ae2-433a-a70b-cdda6d764e53\") " pod="openstack/dnsmasq-dns-6578955fd5-svh8j" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.596487 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3d48a46a-3ae2-433a-a70b-cdda6d764e53-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-svh8j\" (UID: \"3d48a46a-3ae2-433a-a70b-cdda6d764e53\") " pod="openstack/dnsmasq-dns-6578955fd5-svh8j" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.596524 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/51b2c969-f09a-4c90-a897-8b6291716ee7-config-data-custom\") pod \"cinder-api-0\" (UID: \"51b2c969-f09a-4c90-a897-8b6291716ee7\") " pod="openstack/cinder-api-0" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.596572 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3d48a46a-3ae2-433a-a70b-cdda6d764e53-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-svh8j\" (UID: \"3d48a46a-3ae2-433a-a70b-cdda6d764e53\") " pod="openstack/dnsmasq-dns-6578955fd5-svh8j" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.596641 4821 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7c7075f-28ad-46d5-b301-98f3868873db-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.596658 4821 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d7c7075f-28ad-46d5-b301-98f3868873db-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.596954 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/51b2c969-f09a-4c90-a897-8b6291716ee7-etc-machine-id\") pod \"cinder-api-0\" (UID: \"51b2c969-f09a-4c90-a897-8b6291716ee7\") " pod="openstack/cinder-api-0" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.600526 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/51b2c969-f09a-4c90-a897-8b6291716ee7-logs\") pod \"cinder-api-0\" (UID: \"51b2c969-f09a-4c90-a897-8b6291716ee7\") " pod="openstack/cinder-api-0" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.602275 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3d48a46a-3ae2-433a-a70b-cdda6d764e53-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-svh8j\" (UID: \"3d48a46a-3ae2-433a-a70b-cdda6d764e53\") " pod="openstack/dnsmasq-dns-6578955fd5-svh8j" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.603028 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3d48a46a-3ae2-433a-a70b-cdda6d764e53-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-svh8j\" (UID: \"3d48a46a-3ae2-433a-a70b-cdda6d764e53\") " pod="openstack/dnsmasq-dns-6578955fd5-svh8j" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.604472 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3d48a46a-3ae2-433a-a70b-cdda6d764e53-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-svh8j\" (UID: \"3d48a46a-3ae2-433a-a70b-cdda6d764e53\") " pod="openstack/dnsmasq-dns-6578955fd5-svh8j" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.607576 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/51b2c969-f09a-4c90-a897-8b6291716ee7-config-data-custom\") pod \"cinder-api-0\" (UID: \"51b2c969-f09a-4c90-a897-8b6291716ee7\") " pod="openstack/cinder-api-0" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.609746 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-55c965854d-xpf4q" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.613220 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51b2c969-f09a-4c90-a897-8b6291716ee7-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"51b2c969-f09a-4c90-a897-8b6291716ee7\") " pod="openstack/cinder-api-0" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.613842 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3d48a46a-3ae2-433a-a70b-cdda6d764e53-dns-svc\") pod \"dnsmasq-dns-6578955fd5-svh8j\" (UID: \"3d48a46a-3ae2-433a-a70b-cdda6d764e53\") " pod="openstack/dnsmasq-dns-6578955fd5-svh8j" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.617501 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d48a46a-3ae2-433a-a70b-cdda6d764e53-config\") pod \"dnsmasq-dns-6578955fd5-svh8j\" (UID: \"3d48a46a-3ae2-433a-a70b-cdda6d764e53\") " pod="openstack/dnsmasq-dns-6578955fd5-svh8j" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.621958 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51b2c969-f09a-4c90-a897-8b6291716ee7-config-data\") pod \"cinder-api-0\" (UID: \"51b2c969-f09a-4c90-a897-8b6291716ee7\") " pod="openstack/cinder-api-0" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.622149 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/51b2c969-f09a-4c90-a897-8b6291716ee7-scripts\") pod \"cinder-api-0\" (UID: \"51b2c969-f09a-4c90-a897-8b6291716ee7\") " pod="openstack/cinder-api-0" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.628423 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mwxk5\" (UniqueName: \"kubernetes.io/projected/51b2c969-f09a-4c90-a897-8b6291716ee7-kube-api-access-mwxk5\") pod \"cinder-api-0\" (UID: \"51b2c969-f09a-4c90-a897-8b6291716ee7\") " pod="openstack/cinder-api-0" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.632803 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v86b8\" (UniqueName: \"kubernetes.io/projected/3d48a46a-3ae2-433a-a70b-cdda6d764e53-kube-api-access-v86b8\") pod \"dnsmasq-dns-6578955fd5-svh8j\" (UID: \"3d48a46a-3ae2-433a-a70b-cdda6d764e53\") " pod="openstack/dnsmasq-dns-6578955fd5-svh8j" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.728728 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.729445 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-bmmvh" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.739764 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-svh8j" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.747903 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-zbmpp"] Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.756310 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-zbmpp"] Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.793801 4821 scope.go:117] "RemoveContainer" containerID="0a2ed13b5371cbf233dfa5f8b5c412fad3488bad7b3d4bc5787335ea66f93a7a" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.804027 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2fda2637-8b50-43ee-a1dc-6d6c3340030c-dns-svc\") pod \"2fda2637-8b50-43ee-a1dc-6d6c3340030c\" (UID: \"2fda2637-8b50-43ee-a1dc-6d6c3340030c\") " Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.804215 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fda2637-8b50-43ee-a1dc-6d6c3340030c-config\") pod \"2fda2637-8b50-43ee-a1dc-6d6c3340030c\" (UID: \"2fda2637-8b50-43ee-a1dc-6d6c3340030c\") " Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.804239 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2fda2637-8b50-43ee-a1dc-6d6c3340030c-ovsdbserver-sb\") pod \"2fda2637-8b50-43ee-a1dc-6d6c3340030c\" (UID: \"2fda2637-8b50-43ee-a1dc-6d6c3340030c\") " Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.804739 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2fda2637-8b50-43ee-a1dc-6d6c3340030c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2fda2637-8b50-43ee-a1dc-6d6c3340030c" (UID: "2fda2637-8b50-43ee-a1dc-6d6c3340030c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.804785 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2fda2637-8b50-43ee-a1dc-6d6c3340030c-dns-swift-storage-0\") pod \"2fda2637-8b50-43ee-a1dc-6d6c3340030c\" (UID: \"2fda2637-8b50-43ee-a1dc-6d6c3340030c\") " Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.804870 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2fda2637-8b50-43ee-a1dc-6d6c3340030c-ovsdbserver-nb\") pod \"2fda2637-8b50-43ee-a1dc-6d6c3340030c\" (UID: \"2fda2637-8b50-43ee-a1dc-6d6c3340030c\") " Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.804953 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4l5q7\" (UniqueName: \"kubernetes.io/projected/2fda2637-8b50-43ee-a1dc-6d6c3340030c-kube-api-access-4l5q7\") pod \"2fda2637-8b50-43ee-a1dc-6d6c3340030c\" (UID: \"2fda2637-8b50-43ee-a1dc-6d6c3340030c\") " Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.805205 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2fda2637-8b50-43ee-a1dc-6d6c3340030c-config" (OuterVolumeSpecName: "config") pod "2fda2637-8b50-43ee-a1dc-6d6c3340030c" (UID: "2fda2637-8b50-43ee-a1dc-6d6c3340030c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.805704 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2fda2637-8b50-43ee-a1dc-6d6c3340030c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "2fda2637-8b50-43ee-a1dc-6d6c3340030c" (UID: "2fda2637-8b50-43ee-a1dc-6d6c3340030c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.805774 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2fda2637-8b50-43ee-a1dc-6d6c3340030c-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "2fda2637-8b50-43ee-a1dc-6d6c3340030c" (UID: "2fda2637-8b50-43ee-a1dc-6d6c3340030c"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.806546 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2fda2637-8b50-43ee-a1dc-6d6c3340030c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "2fda2637-8b50-43ee-a1dc-6d6c3340030c" (UID: "2fda2637-8b50-43ee-a1dc-6d6c3340030c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.807630 4821 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2fda2637-8b50-43ee-a1dc-6d6c3340030c-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.807655 4821 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fda2637-8b50-43ee-a1dc-6d6c3340030c-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.807666 4821 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2fda2637-8b50-43ee-a1dc-6d6c3340030c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.807677 4821 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2fda2637-8b50-43ee-a1dc-6d6c3340030c-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.807689 4821 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2fda2637-8b50-43ee-a1dc-6d6c3340030c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.809628 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.809868 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2fda2637-8b50-43ee-a1dc-6d6c3340030c-kube-api-access-4l5q7" (OuterVolumeSpecName: "kube-api-access-4l5q7") pod "2fda2637-8b50-43ee-a1dc-6d6c3340030c" (UID: "2fda2637-8b50-43ee-a1dc-6d6c3340030c"). InnerVolumeSpecName "kube-api-access-4l5q7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:49:48 crc kubenswrapper[4821]: I1125 10:49:48.910638 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4l5q7\" (UniqueName: \"kubernetes.io/projected/2fda2637-8b50-43ee-a1dc-6d6c3340030c-kube-api-access-4l5q7\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:49 crc kubenswrapper[4821]: I1125 10:49:49.084863 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-f75f66468-2k6lr"] Nov 25 10:49:49 crc kubenswrapper[4821]: I1125 10:49:49.250465 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5c9c8c87dc-mqk8l"] Nov 25 10:49:49 crc kubenswrapper[4821]: I1125 10:49:49.364349 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-55c965854d-xpf4q"] Nov 25 10:49:49 crc kubenswrapper[4821]: I1125 10:49:49.393281 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-svh8j"] Nov 25 10:49:49 crc kubenswrapper[4821]: I1125 10:49:49.556548 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 25 10:49:49 crc kubenswrapper[4821]: I1125 10:49:49.602759 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"4f5bff28-70e9-4b56-ab38-49511f85852d","Type":"ContainerStarted","Data":"3da9a016ff98c68c6d2cd152abf9e95deba6ba01d9f4f20698258cfea411b238"} Nov 25 10:49:49 crc kubenswrapper[4821]: I1125 10:49:49.608939 4821 generic.go:334] "Generic (PLEG): container finished" podID="9cb63eff-3e7e-436e-aea6-ce8254c3db77" containerID="0764f0b45f4c9a4cc4d1c4342d5f4f2d448117f396e64fdd31e641a7ca96dd03" exitCode=0 Nov 25 10:49:49 crc kubenswrapper[4821]: I1125 10:49:49.608969 4821 generic.go:334] "Generic (PLEG): container finished" podID="9cb63eff-3e7e-436e-aea6-ce8254c3db77" containerID="666f50101704e276ddca6f07195913aeb8365df95b6639c5717f6c54b431976e" exitCode=2 Nov 25 10:49:49 crc kubenswrapper[4821]: I1125 10:49:49.609005 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9cb63eff-3e7e-436e-aea6-ce8254c3db77","Type":"ContainerDied","Data":"0764f0b45f4c9a4cc4d1c4342d5f4f2d448117f396e64fdd31e641a7ca96dd03"} Nov 25 10:49:49 crc kubenswrapper[4821]: I1125 10:49:49.609027 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9cb63eff-3e7e-436e-aea6-ce8254c3db77","Type":"ContainerDied","Data":"666f50101704e276ddca6f07195913aeb8365df95b6639c5717f6c54b431976e"} Nov 25 10:49:49 crc kubenswrapper[4821]: I1125 10:49:49.613032 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-55c965854d-xpf4q" event={"ID":"73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7","Type":"ContainerStarted","Data":"7563ef010033402324871f9d93411bb4645bbf2d61d1a16b84199972670d5566"} Nov 25 10:49:49 crc kubenswrapper[4821]: I1125 10:49:49.618735 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-55575b4d57-6t4kz" event={"ID":"a8419a57-35ff-4340-b2d7-fd11724948fe","Type":"ContainerStarted","Data":"cf3ddc470a9683f17b2c14a914337e8b66622d6b19f678e6b44531500a250722"} Nov 25 10:49:49 crc kubenswrapper[4821]: I1125 10:49:49.618945 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-55575b4d57-6t4kz" Nov 25 10:49:49 crc kubenswrapper[4821]: I1125 10:49:49.621857 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5c9c8c87dc-mqk8l" event={"ID":"62fc74a6-526c-4953-8439-b4af8ce5f39b","Type":"ContainerStarted","Data":"9bda208075afae12f7e16fe767b8e41db4669ef1a4a51c661e8f15eb238ad76f"} Nov 25 10:49:49 crc kubenswrapper[4821]: I1125 10:49:49.629205 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-f75f66468-2k6lr" event={"ID":"1e6fee50-9d84-4965-a222-a55f6dcfc4cc","Type":"ContainerStarted","Data":"8e57e0ba9451ae922c54db96e0f5048e66069f92946c701ef770f251fe727011"} Nov 25 10:49:49 crc kubenswrapper[4821]: W1125 10:49:49.636810 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod51b2c969_f09a_4c90_a897_8b6291716ee7.slice/crio-3489affce0bd0bc4dc75c1b7ce2eb1a6d4226cf0b2524407e17eaf765dd7adb4 WatchSource:0}: Error finding container 3489affce0bd0bc4dc75c1b7ce2eb1a6d4226cf0b2524407e17eaf765dd7adb4: Status 404 returned error can't find the container with id 3489affce0bd0bc4dc75c1b7ce2eb1a6d4226cf0b2524407e17eaf765dd7adb4 Nov 25 10:49:49 crc kubenswrapper[4821]: I1125 10:49:49.640145 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-bmmvh" Nov 25 10:49:49 crc kubenswrapper[4821]: I1125 10:49:49.647212 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-svh8j" event={"ID":"3d48a46a-3ae2-433a-a70b-cdda6d764e53","Type":"ContainerStarted","Data":"7db3ab42634c39bef4a1bdab6529d7c05eeee67cf31996620083613a8667997a"} Nov 25 10:49:49 crc kubenswrapper[4821]: I1125 10:49:49.662135 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-55575b4d57-6t4kz" podStartSLOduration=10.662111598 podStartE2EDuration="10.662111598s" podCreationTimestamp="2025-11-25 10:49:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:49:49.646816567 +0000 UTC m=+1060.183136434" watchObservedRunningTime="2025-11-25 10:49:49.662111598 +0000 UTC m=+1060.198431435" Nov 25 10:49:49 crc kubenswrapper[4821]: I1125 10:49:49.791086 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-bmmvh"] Nov 25 10:49:49 crc kubenswrapper[4821]: I1125 10:49:49.798723 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-bmmvh"] Nov 25 10:49:50 crc kubenswrapper[4821]: I1125 10:49:50.161632 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2fda2637-8b50-43ee-a1dc-6d6c3340030c" path="/var/lib/kubelet/pods/2fda2637-8b50-43ee-a1dc-6d6c3340030c/volumes" Nov 25 10:49:50 crc kubenswrapper[4821]: I1125 10:49:50.162153 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7c7075f-28ad-46d5-b301-98f3868873db" path="/var/lib/kubelet/pods/d7c7075f-28ad-46d5-b301-98f3868873db/volumes" Nov 25 10:49:50 crc kubenswrapper[4821]: I1125 10:49:50.519852 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 25 10:49:50 crc kubenswrapper[4821]: I1125 10:49:50.672614 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-55c965854d-xpf4q" event={"ID":"73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7","Type":"ContainerStarted","Data":"1b4b4e05d8669cea149cb416589c262aeee2db1f26dafb11f5ec8795a3a85047"} Nov 25 10:49:50 crc kubenswrapper[4821]: I1125 10:49:50.672687 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-55c965854d-xpf4q" event={"ID":"73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7","Type":"ContainerStarted","Data":"8bee610994bbfbd0e4f395db69e2e44e857f94bfbafa8c30ddba075694222c63"} Nov 25 10:49:50 crc kubenswrapper[4821]: I1125 10:49:50.675686 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-55c965854d-xpf4q" Nov 25 10:49:50 crc kubenswrapper[4821]: I1125 10:49:50.675748 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-55c965854d-xpf4q" Nov 25 10:49:50 crc kubenswrapper[4821]: I1125 10:49:50.684122 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"51b2c969-f09a-4c90-a897-8b6291716ee7","Type":"ContainerStarted","Data":"3489affce0bd0bc4dc75c1b7ce2eb1a6d4226cf0b2524407e17eaf765dd7adb4"} Nov 25 10:49:50 crc kubenswrapper[4821]: I1125 10:49:50.687472 4821 generic.go:334] "Generic (PLEG): container finished" podID="3d48a46a-3ae2-433a-a70b-cdda6d764e53" containerID="9ff6a635f643fcc567f969ef5dcb1db96e5cdbfa8ccc746ae9f443de342fa116" exitCode=0 Nov 25 10:49:50 crc kubenswrapper[4821]: I1125 10:49:50.687570 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-svh8j" event={"ID":"3d48a46a-3ae2-433a-a70b-cdda6d764e53","Type":"ContainerDied","Data":"9ff6a635f643fcc567f969ef5dcb1db96e5cdbfa8ccc746ae9f443de342fa116"} Nov 25 10:49:50 crc kubenswrapper[4821]: I1125 10:49:50.724839 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-55c965854d-xpf4q" podStartSLOduration=3.724816661 podStartE2EDuration="3.724816661s" podCreationTimestamp="2025-11-25 10:49:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:49:50.710540465 +0000 UTC m=+1061.246860322" watchObservedRunningTime="2025-11-25 10:49:50.724816661 +0000 UTC m=+1061.261136518" Nov 25 10:49:51 crc kubenswrapper[4821]: I1125 10:49:51.412358 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 25 10:49:51 crc kubenswrapper[4821]: I1125 10:49:51.709133 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"51b2c969-f09a-4c90-a897-8b6291716ee7","Type":"ContainerStarted","Data":"bfcd58ab3bfcac6d66961f3ccea2866bf6fd3ae05317283d51b75ec0f336edee"} Nov 25 10:49:51 crc kubenswrapper[4821]: I1125 10:49:51.711095 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-svh8j" event={"ID":"3d48a46a-3ae2-433a-a70b-cdda6d764e53","Type":"ContainerStarted","Data":"66efde77ebbf8d3d6233842f84c3ca3fba12aabb2793cf543e112fd2340ca7a2"} Nov 25 10:49:51 crc kubenswrapper[4821]: I1125 10:49:51.711830 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6578955fd5-svh8j" Nov 25 10:49:51 crc kubenswrapper[4821]: I1125 10:49:51.722710 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"4f5bff28-70e9-4b56-ab38-49511f85852d","Type":"ContainerStarted","Data":"a4e878765149277342071df06b9a3d4de0e2483bb18a27e40f85f6f2947fde9b"} Nov 25 10:49:51 crc kubenswrapper[4821]: I1125 10:49:51.731537 4821 generic.go:334] "Generic (PLEG): container finished" podID="79a12e89-074b-48f2-bf44-b177fe2e84e2" containerID="704790aed35077ae5429fdc8a0d61310a824faa317eb4be8fe7009dd3293d6f7" exitCode=137 Nov 25 10:49:51 crc kubenswrapper[4821]: I1125 10:49:51.731577 4821 generic.go:334] "Generic (PLEG): container finished" podID="79a12e89-074b-48f2-bf44-b177fe2e84e2" containerID="4006d528ae6d37112c7f67c5d9d59de50d95a411c30d1ffd8b665efe54524f95" exitCode=137 Nov 25 10:49:51 crc kubenswrapper[4821]: I1125 10:49:51.731648 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-574c6969fc-sjxzg" event={"ID":"79a12e89-074b-48f2-bf44-b177fe2e84e2","Type":"ContainerDied","Data":"704790aed35077ae5429fdc8a0d61310a824faa317eb4be8fe7009dd3293d6f7"} Nov 25 10:49:51 crc kubenswrapper[4821]: I1125 10:49:51.731725 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-574c6969fc-sjxzg" event={"ID":"79a12e89-074b-48f2-bf44-b177fe2e84e2","Type":"ContainerDied","Data":"4006d528ae6d37112c7f67c5d9d59de50d95a411c30d1ffd8b665efe54524f95"} Nov 25 10:49:51 crc kubenswrapper[4821]: I1125 10:49:51.732113 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6578955fd5-svh8j" podStartSLOduration=4.732096694 podStartE2EDuration="4.732096694s" podCreationTimestamp="2025-11-25 10:49:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:49:51.729533538 +0000 UTC m=+1062.265853385" watchObservedRunningTime="2025-11-25 10:49:51.732096694 +0000 UTC m=+1062.268416551" Nov 25 10:49:51 crc kubenswrapper[4821]: I1125 10:49:51.736402 4821 generic.go:334] "Generic (PLEG): container finished" podID="77e0b5ce-8867-4d7e-b89e-170e67f3a548" containerID="72ace1a8cf21a307bdc5ebcad22a3b55aefccfd5287aca26c9642a08991a80cc" exitCode=137 Nov 25 10:49:51 crc kubenswrapper[4821]: I1125 10:49:51.736428 4821 generic.go:334] "Generic (PLEG): container finished" podID="77e0b5ce-8867-4d7e-b89e-170e67f3a548" containerID="f3607e3b49353ca4be3512145e78ea3d93f1aaf59d3670587ad46301a2656aa3" exitCode=137 Nov 25 10:49:51 crc kubenswrapper[4821]: I1125 10:49:51.737216 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7558555969-g2z2s" event={"ID":"77e0b5ce-8867-4d7e-b89e-170e67f3a548","Type":"ContainerDied","Data":"72ace1a8cf21a307bdc5ebcad22a3b55aefccfd5287aca26c9642a08991a80cc"} Nov 25 10:49:51 crc kubenswrapper[4821]: I1125 10:49:51.737245 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7558555969-g2z2s" event={"ID":"77e0b5ce-8867-4d7e-b89e-170e67f3a548","Type":"ContainerDied","Data":"f3607e3b49353ca4be3512145e78ea3d93f1aaf59d3670587ad46301a2656aa3"} Nov 25 10:49:52 crc kubenswrapper[4821]: I1125 10:49:52.195359 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-574c6969fc-sjxzg" Nov 25 10:49:52 crc kubenswrapper[4821]: I1125 10:49:52.201608 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7558555969-g2z2s" Nov 25 10:49:52 crc kubenswrapper[4821]: I1125 10:49:52.324508 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/77e0b5ce-8867-4d7e-b89e-170e67f3a548-scripts\") pod \"77e0b5ce-8867-4d7e-b89e-170e67f3a548\" (UID: \"77e0b5ce-8867-4d7e-b89e-170e67f3a548\") " Nov 25 10:49:52 crc kubenswrapper[4821]: I1125 10:49:52.324625 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/79a12e89-074b-48f2-bf44-b177fe2e84e2-scripts\") pod \"79a12e89-074b-48f2-bf44-b177fe2e84e2\" (UID: \"79a12e89-074b-48f2-bf44-b177fe2e84e2\") " Nov 25 10:49:52 crc kubenswrapper[4821]: I1125 10:49:52.324791 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/77e0b5ce-8867-4d7e-b89e-170e67f3a548-logs\") pod \"77e0b5ce-8867-4d7e-b89e-170e67f3a548\" (UID: \"77e0b5ce-8867-4d7e-b89e-170e67f3a548\") " Nov 25 10:49:52 crc kubenswrapper[4821]: I1125 10:49:52.324907 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/79a12e89-074b-48f2-bf44-b177fe2e84e2-config-data\") pod \"79a12e89-074b-48f2-bf44-b177fe2e84e2\" (UID: \"79a12e89-074b-48f2-bf44-b177fe2e84e2\") " Nov 25 10:49:52 crc kubenswrapper[4821]: I1125 10:49:52.324931 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-frk9p\" (UniqueName: \"kubernetes.io/projected/79a12e89-074b-48f2-bf44-b177fe2e84e2-kube-api-access-frk9p\") pod \"79a12e89-074b-48f2-bf44-b177fe2e84e2\" (UID: \"79a12e89-074b-48f2-bf44-b177fe2e84e2\") " Nov 25 10:49:52 crc kubenswrapper[4821]: I1125 10:49:52.324948 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/77e0b5ce-8867-4d7e-b89e-170e67f3a548-config-data\") pod \"77e0b5ce-8867-4d7e-b89e-170e67f3a548\" (UID: \"77e0b5ce-8867-4d7e-b89e-170e67f3a548\") " Nov 25 10:49:52 crc kubenswrapper[4821]: I1125 10:49:52.324985 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/79a12e89-074b-48f2-bf44-b177fe2e84e2-logs\") pod \"79a12e89-074b-48f2-bf44-b177fe2e84e2\" (UID: \"79a12e89-074b-48f2-bf44-b177fe2e84e2\") " Nov 25 10:49:52 crc kubenswrapper[4821]: I1125 10:49:52.325019 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/79a12e89-074b-48f2-bf44-b177fe2e84e2-horizon-secret-key\") pod \"79a12e89-074b-48f2-bf44-b177fe2e84e2\" (UID: \"79a12e89-074b-48f2-bf44-b177fe2e84e2\") " Nov 25 10:49:52 crc kubenswrapper[4821]: I1125 10:49:52.325094 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v2wrj\" (UniqueName: \"kubernetes.io/projected/77e0b5ce-8867-4d7e-b89e-170e67f3a548-kube-api-access-v2wrj\") pod \"77e0b5ce-8867-4d7e-b89e-170e67f3a548\" (UID: \"77e0b5ce-8867-4d7e-b89e-170e67f3a548\") " Nov 25 10:49:52 crc kubenswrapper[4821]: I1125 10:49:52.325137 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/77e0b5ce-8867-4d7e-b89e-170e67f3a548-horizon-secret-key\") pod \"77e0b5ce-8867-4d7e-b89e-170e67f3a548\" (UID: \"77e0b5ce-8867-4d7e-b89e-170e67f3a548\") " Nov 25 10:49:52 crc kubenswrapper[4821]: I1125 10:49:52.328298 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/79a12e89-074b-48f2-bf44-b177fe2e84e2-logs" (OuterVolumeSpecName: "logs") pod "79a12e89-074b-48f2-bf44-b177fe2e84e2" (UID: "79a12e89-074b-48f2-bf44-b177fe2e84e2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:49:52 crc kubenswrapper[4821]: I1125 10:49:52.330815 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/77e0b5ce-8867-4d7e-b89e-170e67f3a548-logs" (OuterVolumeSpecName: "logs") pod "77e0b5ce-8867-4d7e-b89e-170e67f3a548" (UID: "77e0b5ce-8867-4d7e-b89e-170e67f3a548"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:49:52 crc kubenswrapper[4821]: I1125 10:49:52.350328 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79a12e89-074b-48f2-bf44-b177fe2e84e2-kube-api-access-frk9p" (OuterVolumeSpecName: "kube-api-access-frk9p") pod "79a12e89-074b-48f2-bf44-b177fe2e84e2" (UID: "79a12e89-074b-48f2-bf44-b177fe2e84e2"). InnerVolumeSpecName "kube-api-access-frk9p". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:49:52 crc kubenswrapper[4821]: I1125 10:49:52.359913 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/79a12e89-074b-48f2-bf44-b177fe2e84e2-scripts" (OuterVolumeSpecName: "scripts") pod "79a12e89-074b-48f2-bf44-b177fe2e84e2" (UID: "79a12e89-074b-48f2-bf44-b177fe2e84e2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:49:52 crc kubenswrapper[4821]: I1125 10:49:52.372488 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77e0b5ce-8867-4d7e-b89e-170e67f3a548-kube-api-access-v2wrj" (OuterVolumeSpecName: "kube-api-access-v2wrj") pod "77e0b5ce-8867-4d7e-b89e-170e67f3a548" (UID: "77e0b5ce-8867-4d7e-b89e-170e67f3a548"). InnerVolumeSpecName "kube-api-access-v2wrj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:49:52 crc kubenswrapper[4821]: I1125 10:49:52.385995 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79a12e89-074b-48f2-bf44-b177fe2e84e2-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "79a12e89-074b-48f2-bf44-b177fe2e84e2" (UID: "79a12e89-074b-48f2-bf44-b177fe2e84e2"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:49:52 crc kubenswrapper[4821]: I1125 10:49:52.386766 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77e0b5ce-8867-4d7e-b89e-170e67f3a548-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "77e0b5ce-8867-4d7e-b89e-170e67f3a548" (UID: "77e0b5ce-8867-4d7e-b89e-170e67f3a548"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:49:52 crc kubenswrapper[4821]: I1125 10:49:52.394942 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/77e0b5ce-8867-4d7e-b89e-170e67f3a548-config-data" (OuterVolumeSpecName: "config-data") pod "77e0b5ce-8867-4d7e-b89e-170e67f3a548" (UID: "77e0b5ce-8867-4d7e-b89e-170e67f3a548"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:49:52 crc kubenswrapper[4821]: I1125 10:49:52.415754 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/77e0b5ce-8867-4d7e-b89e-170e67f3a548-scripts" (OuterVolumeSpecName: "scripts") pod "77e0b5ce-8867-4d7e-b89e-170e67f3a548" (UID: "77e0b5ce-8867-4d7e-b89e-170e67f3a548"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:49:52 crc kubenswrapper[4821]: I1125 10:49:52.419500 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/79a12e89-074b-48f2-bf44-b177fe2e84e2-config-data" (OuterVolumeSpecName: "config-data") pod "79a12e89-074b-48f2-bf44-b177fe2e84e2" (UID: "79a12e89-074b-48f2-bf44-b177fe2e84e2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:49:52 crc kubenswrapper[4821]: I1125 10:49:52.427299 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v2wrj\" (UniqueName: \"kubernetes.io/projected/77e0b5ce-8867-4d7e-b89e-170e67f3a548-kube-api-access-v2wrj\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:52 crc kubenswrapper[4821]: I1125 10:49:52.427321 4821 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/77e0b5ce-8867-4d7e-b89e-170e67f3a548-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:52 crc kubenswrapper[4821]: I1125 10:49:52.427331 4821 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/77e0b5ce-8867-4d7e-b89e-170e67f3a548-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:52 crc kubenswrapper[4821]: I1125 10:49:52.427339 4821 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/79a12e89-074b-48f2-bf44-b177fe2e84e2-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:52 crc kubenswrapper[4821]: I1125 10:49:52.427347 4821 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/77e0b5ce-8867-4d7e-b89e-170e67f3a548-logs\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:52 crc kubenswrapper[4821]: I1125 10:49:52.427354 4821 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/79a12e89-074b-48f2-bf44-b177fe2e84e2-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:52 crc kubenswrapper[4821]: I1125 10:49:52.427363 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-frk9p\" (UniqueName: \"kubernetes.io/projected/79a12e89-074b-48f2-bf44-b177fe2e84e2-kube-api-access-frk9p\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:52 crc kubenswrapper[4821]: I1125 10:49:52.427373 4821 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/77e0b5ce-8867-4d7e-b89e-170e67f3a548-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:52 crc kubenswrapper[4821]: I1125 10:49:52.427381 4821 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/79a12e89-074b-48f2-bf44-b177fe2e84e2-logs\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:52 crc kubenswrapper[4821]: I1125 10:49:52.427389 4821 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/79a12e89-074b-48f2-bf44-b177fe2e84e2-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:52 crc kubenswrapper[4821]: I1125 10:49:52.774577 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"4f5bff28-70e9-4b56-ab38-49511f85852d","Type":"ContainerStarted","Data":"42e6a15ed6523ba17eead013e533d86e41662911fe061607b0113ad40f221233"} Nov 25 10:49:52 crc kubenswrapper[4821]: I1125 10:49:52.782242 4821 generic.go:334] "Generic (PLEG): container finished" podID="f9398a81-673d-46b4-b560-91c6e2885532" containerID="71773c54cfbe03e75458bf24e2a51cb38377a4ced5a4d0293161a4ebc1ea7207" exitCode=137 Nov 25 10:49:52 crc kubenswrapper[4821]: I1125 10:49:52.782276 4821 generic.go:334] "Generic (PLEG): container finished" podID="f9398a81-673d-46b4-b560-91c6e2885532" containerID="3561021b89555fbb5e74f9d8972eb9d892f00a8f4362cad3f466a085cb92eab4" exitCode=137 Nov 25 10:49:52 crc kubenswrapper[4821]: I1125 10:49:52.782366 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-64dc894fb9-5xf8b" event={"ID":"f9398a81-673d-46b4-b560-91c6e2885532","Type":"ContainerDied","Data":"71773c54cfbe03e75458bf24e2a51cb38377a4ced5a4d0293161a4ebc1ea7207"} Nov 25 10:49:52 crc kubenswrapper[4821]: I1125 10:49:52.782394 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-64dc894fb9-5xf8b" event={"ID":"f9398a81-673d-46b4-b560-91c6e2885532","Type":"ContainerDied","Data":"3561021b89555fbb5e74f9d8972eb9d892f00a8f4362cad3f466a085cb92eab4"} Nov 25 10:49:52 crc kubenswrapper[4821]: I1125 10:49:52.785936 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-574c6969fc-sjxzg" event={"ID":"79a12e89-074b-48f2-bf44-b177fe2e84e2","Type":"ContainerDied","Data":"a796dab8becc201a1a5893ba83b9780af86245e2776ce6e5ed298b535f85e975"} Nov 25 10:49:52 crc kubenswrapper[4821]: I1125 10:49:52.785972 4821 scope.go:117] "RemoveContainer" containerID="704790aed35077ae5429fdc8a0d61310a824faa317eb4be8fe7009dd3293d6f7" Nov 25 10:49:52 crc kubenswrapper[4821]: I1125 10:49:52.785977 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-574c6969fc-sjxzg" Nov 25 10:49:52 crc kubenswrapper[4821]: I1125 10:49:52.806884 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7558555969-g2z2s" Nov 25 10:49:52 crc kubenswrapper[4821]: I1125 10:49:52.806890 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7558555969-g2z2s" event={"ID":"77e0b5ce-8867-4d7e-b89e-170e67f3a548","Type":"ContainerDied","Data":"f6e1c6c23c32f47b087d392cbf5933988edde50c045e7baa12e5de4c8bf6982e"} Nov 25 10:49:52 crc kubenswrapper[4821]: I1125 10:49:52.822315 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="51b2c969-f09a-4c90-a897-8b6291716ee7" containerName="cinder-api-log" containerID="cri-o://bfcd58ab3bfcac6d66961f3ccea2866bf6fd3ae05317283d51b75ec0f336edee" gracePeriod=30 Nov 25 10:49:52 crc kubenswrapper[4821]: I1125 10:49:52.822493 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"51b2c969-f09a-4c90-a897-8b6291716ee7","Type":"ContainerStarted","Data":"dd3b43875d28c2b6e542e602fa80e40c8a008021f40ec73bda730d66d4f16f20"} Nov 25 10:49:52 crc kubenswrapper[4821]: I1125 10:49:52.822589 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="51b2c969-f09a-4c90-a897-8b6291716ee7" containerName="cinder-api" containerID="cri-o://dd3b43875d28c2b6e542e602fa80e40c8a008021f40ec73bda730d66d4f16f20" gracePeriod=30 Nov 25 10:49:52 crc kubenswrapper[4821]: I1125 10:49:52.823104 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 25 10:49:52 crc kubenswrapper[4821]: I1125 10:49:52.831267 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 25 10:49:52 crc kubenswrapper[4821]: I1125 10:49:52.835780 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.801304457 podStartE2EDuration="5.835756305s" podCreationTimestamp="2025-11-25 10:49:47 +0000 UTC" firstStartedPulling="2025-11-25 10:49:48.771267669 +0000 UTC m=+1059.307587516" lastFinishedPulling="2025-11-25 10:49:49.805719517 +0000 UTC m=+1060.342039364" observedRunningTime="2025-11-25 10:49:52.798301536 +0000 UTC m=+1063.334621383" watchObservedRunningTime="2025-11-25 10:49:52.835756305 +0000 UTC m=+1063.372076152" Nov 25 10:49:52 crc kubenswrapper[4821]: I1125 10:49:52.858222 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-574c6969fc-sjxzg"] Nov 25 10:49:52 crc kubenswrapper[4821]: I1125 10:49:52.896595 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-574c6969fc-sjxzg"] Nov 25 10:49:52 crc kubenswrapper[4821]: I1125 10:49:52.906507 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=5.906481848 podStartE2EDuration="5.906481848s" podCreationTimestamp="2025-11-25 10:49:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:49:52.864600985 +0000 UTC m=+1063.400920832" watchObservedRunningTime="2025-11-25 10:49:52.906481848 +0000 UTC m=+1063.442801695" Nov 25 10:49:52 crc kubenswrapper[4821]: I1125 10:49:52.920906 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7558555969-g2z2s"] Nov 25 10:49:52 crc kubenswrapper[4821]: I1125 10:49:52.927750 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7558555969-g2z2s"] Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.051912 4821 scope.go:117] "RemoveContainer" containerID="4006d528ae6d37112c7f67c5d9d59de50d95a411c30d1ffd8b665efe54524f95" Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.082261 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-64dc894fb9-5xf8b" Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.119907 4821 scope.go:117] "RemoveContainer" containerID="72ace1a8cf21a307bdc5ebcad22a3b55aefccfd5287aca26c9642a08991a80cc" Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.151121 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4xr2t\" (UniqueName: \"kubernetes.io/projected/f9398a81-673d-46b4-b560-91c6e2885532-kube-api-access-4xr2t\") pod \"f9398a81-673d-46b4-b560-91c6e2885532\" (UID: \"f9398a81-673d-46b4-b560-91c6e2885532\") " Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.151255 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f9398a81-673d-46b4-b560-91c6e2885532-horizon-secret-key\") pod \"f9398a81-673d-46b4-b560-91c6e2885532\" (UID: \"f9398a81-673d-46b4-b560-91c6e2885532\") " Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.151460 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f9398a81-673d-46b4-b560-91c6e2885532-logs\") pod \"f9398a81-673d-46b4-b560-91c6e2885532\" (UID: \"f9398a81-673d-46b4-b560-91c6e2885532\") " Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.151542 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f9398a81-673d-46b4-b560-91c6e2885532-scripts\") pod \"f9398a81-673d-46b4-b560-91c6e2885532\" (UID: \"f9398a81-673d-46b4-b560-91c6e2885532\") " Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.151603 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f9398a81-673d-46b4-b560-91c6e2885532-config-data\") pod \"f9398a81-673d-46b4-b560-91c6e2885532\" (UID: \"f9398a81-673d-46b4-b560-91c6e2885532\") " Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.162944 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f9398a81-673d-46b4-b560-91c6e2885532-logs" (OuterVolumeSpecName: "logs") pod "f9398a81-673d-46b4-b560-91c6e2885532" (UID: "f9398a81-673d-46b4-b560-91c6e2885532"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.182077 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9398a81-673d-46b4-b560-91c6e2885532-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "f9398a81-673d-46b4-b560-91c6e2885532" (UID: "f9398a81-673d-46b4-b560-91c6e2885532"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.232676 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9398a81-673d-46b4-b560-91c6e2885532-kube-api-access-4xr2t" (OuterVolumeSpecName: "kube-api-access-4xr2t") pod "f9398a81-673d-46b4-b560-91c6e2885532" (UID: "f9398a81-673d-46b4-b560-91c6e2885532"). InnerVolumeSpecName "kube-api-access-4xr2t". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.233922 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f9398a81-673d-46b4-b560-91c6e2885532-scripts" (OuterVolumeSpecName: "scripts") pod "f9398a81-673d-46b4-b560-91c6e2885532" (UID: "f9398a81-673d-46b4-b560-91c6e2885532"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.239845 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f9398a81-673d-46b4-b560-91c6e2885532-config-data" (OuterVolumeSpecName: "config-data") pod "f9398a81-673d-46b4-b560-91c6e2885532" (UID: "f9398a81-673d-46b4-b560-91c6e2885532"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.253839 4821 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f9398a81-673d-46b4-b560-91c6e2885532-logs\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.254092 4821 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f9398a81-673d-46b4-b560-91c6e2885532-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.254251 4821 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f9398a81-673d-46b4-b560-91c6e2885532-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.254340 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4xr2t\" (UniqueName: \"kubernetes.io/projected/f9398a81-673d-46b4-b560-91c6e2885532-kube-api-access-4xr2t\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.254421 4821 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f9398a81-673d-46b4-b560-91c6e2885532-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.423195 4821 scope.go:117] "RemoveContainer" containerID="f3607e3b49353ca4be3512145e78ea3d93f1aaf59d3670587ad46301a2656aa3" Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.481694 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.562911 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9cb63eff-3e7e-436e-aea6-ce8254c3db77-scripts\") pod \"9cb63eff-3e7e-436e-aea6-ce8254c3db77\" (UID: \"9cb63eff-3e7e-436e-aea6-ce8254c3db77\") " Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.564242 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9cb63eff-3e7e-436e-aea6-ce8254c3db77-run-httpd\") pod \"9cb63eff-3e7e-436e-aea6-ce8254c3db77\" (UID: \"9cb63eff-3e7e-436e-aea6-ce8254c3db77\") " Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.564294 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9cb63eff-3e7e-436e-aea6-ce8254c3db77-log-httpd\") pod \"9cb63eff-3e7e-436e-aea6-ce8254c3db77\" (UID: \"9cb63eff-3e7e-436e-aea6-ce8254c3db77\") " Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.564353 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f5cz7\" (UniqueName: \"kubernetes.io/projected/9cb63eff-3e7e-436e-aea6-ce8254c3db77-kube-api-access-f5cz7\") pod \"9cb63eff-3e7e-436e-aea6-ce8254c3db77\" (UID: \"9cb63eff-3e7e-436e-aea6-ce8254c3db77\") " Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.564425 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9cb63eff-3e7e-436e-aea6-ce8254c3db77-sg-core-conf-yaml\") pod \"9cb63eff-3e7e-436e-aea6-ce8254c3db77\" (UID: \"9cb63eff-3e7e-436e-aea6-ce8254c3db77\") " Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.564451 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cb63eff-3e7e-436e-aea6-ce8254c3db77-config-data\") pod \"9cb63eff-3e7e-436e-aea6-ce8254c3db77\" (UID: \"9cb63eff-3e7e-436e-aea6-ce8254c3db77\") " Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.564502 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cb63eff-3e7e-436e-aea6-ce8254c3db77-combined-ca-bundle\") pod \"9cb63eff-3e7e-436e-aea6-ce8254c3db77\" (UID: \"9cb63eff-3e7e-436e-aea6-ce8254c3db77\") " Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.564572 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9cb63eff-3e7e-436e-aea6-ce8254c3db77-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "9cb63eff-3e7e-436e-aea6-ce8254c3db77" (UID: "9cb63eff-3e7e-436e-aea6-ce8254c3db77"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.565065 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9cb63eff-3e7e-436e-aea6-ce8254c3db77-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "9cb63eff-3e7e-436e-aea6-ce8254c3db77" (UID: "9cb63eff-3e7e-436e-aea6-ce8254c3db77"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.565293 4821 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9cb63eff-3e7e-436e-aea6-ce8254c3db77-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.565318 4821 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9cb63eff-3e7e-436e-aea6-ce8254c3db77-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.569388 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cb63eff-3e7e-436e-aea6-ce8254c3db77-scripts" (OuterVolumeSpecName: "scripts") pod "9cb63eff-3e7e-436e-aea6-ce8254c3db77" (UID: "9cb63eff-3e7e-436e-aea6-ce8254c3db77"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.587226 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9cb63eff-3e7e-436e-aea6-ce8254c3db77-kube-api-access-f5cz7" (OuterVolumeSpecName: "kube-api-access-f5cz7") pod "9cb63eff-3e7e-436e-aea6-ce8254c3db77" (UID: "9cb63eff-3e7e-436e-aea6-ce8254c3db77"). InnerVolumeSpecName "kube-api-access-f5cz7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.604799 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cb63eff-3e7e-436e-aea6-ce8254c3db77-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "9cb63eff-3e7e-436e-aea6-ce8254c3db77" (UID: "9cb63eff-3e7e-436e-aea6-ce8254c3db77"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.645294 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cb63eff-3e7e-436e-aea6-ce8254c3db77-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9cb63eff-3e7e-436e-aea6-ce8254c3db77" (UID: "9cb63eff-3e7e-436e-aea6-ce8254c3db77"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.645585 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-869c6f5bfb-l785w" Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.664376 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cb63eff-3e7e-436e-aea6-ce8254c3db77-config-data" (OuterVolumeSpecName: "config-data") pod "9cb63eff-3e7e-436e-aea6-ce8254c3db77" (UID: "9cb63eff-3e7e-436e-aea6-ce8254c3db77"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.669772 4821 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cb63eff-3e7e-436e-aea6-ce8254c3db77-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.669816 4821 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9cb63eff-3e7e-436e-aea6-ce8254c3db77-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.669831 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f5cz7\" (UniqueName: \"kubernetes.io/projected/9cb63eff-3e7e-436e-aea6-ce8254c3db77-kube-api-access-f5cz7\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.669845 4821 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9cb63eff-3e7e-436e-aea6-ce8254c3db77-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.669856 4821 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cb63eff-3e7e-436e-aea6-ce8254c3db77-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.729234 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-67f765d856-ncc6t" Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.836065 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5c9c8c87dc-mqk8l" event={"ID":"62fc74a6-526c-4953-8439-b4af8ce5f39b","Type":"ContainerStarted","Data":"1463c342b33ac69fcbe5674a80853721c793dcbfc7193bb82b4f2e28e8f24941"} Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.836105 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5c9c8c87dc-mqk8l" event={"ID":"62fc74a6-526c-4953-8439-b4af8ce5f39b","Type":"ContainerStarted","Data":"32401690c464ab9658f1de24e6aba3697a0bc5883655cc41d5bdabe2bda54294"} Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.838608 4821 generic.go:334] "Generic (PLEG): container finished" podID="51b2c969-f09a-4c90-a897-8b6291716ee7" containerID="dd3b43875d28c2b6e542e602fa80e40c8a008021f40ec73bda730d66d4f16f20" exitCode=0 Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.838636 4821 generic.go:334] "Generic (PLEG): container finished" podID="51b2c969-f09a-4c90-a897-8b6291716ee7" containerID="bfcd58ab3bfcac6d66961f3ccea2866bf6fd3ae05317283d51b75ec0f336edee" exitCode=143 Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.838683 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"51b2c969-f09a-4c90-a897-8b6291716ee7","Type":"ContainerDied","Data":"dd3b43875d28c2b6e542e602fa80e40c8a008021f40ec73bda730d66d4f16f20"} Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.838698 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"51b2c969-f09a-4c90-a897-8b6291716ee7","Type":"ContainerDied","Data":"bfcd58ab3bfcac6d66961f3ccea2866bf6fd3ae05317283d51b75ec0f336edee"} Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.838707 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"51b2c969-f09a-4c90-a897-8b6291716ee7","Type":"ContainerDied","Data":"3489affce0bd0bc4dc75c1b7ce2eb1a6d4226cf0b2524407e17eaf765dd7adb4"} Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.838716 4821 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3489affce0bd0bc4dc75c1b7ce2eb1a6d4226cf0b2524407e17eaf765dd7adb4" Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.840271 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-f75f66468-2k6lr" event={"ID":"1e6fee50-9d84-4965-a222-a55f6dcfc4cc","Type":"ContainerStarted","Data":"0b9731ef6ac9262060653d498247b674a17f07008596d5db60d445a37429e324"} Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.844715 4821 generic.go:334] "Generic (PLEG): container finished" podID="9cb63eff-3e7e-436e-aea6-ce8254c3db77" containerID="cf5c6db7d1dca33b423304d3462893904bd7f34d04492e4bdf0db6099734590c" exitCode=0 Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.844785 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.844756 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9cb63eff-3e7e-436e-aea6-ce8254c3db77","Type":"ContainerDied","Data":"cf5c6db7d1dca33b423304d3462893904bd7f34d04492e4bdf0db6099734590c"} Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.844840 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9cb63eff-3e7e-436e-aea6-ce8254c3db77","Type":"ContainerDied","Data":"3b31424b0b3432b8ccdba74c9228c79dbbfde9d71e6fd17d2545670fd87680df"} Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.844870 4821 scope.go:117] "RemoveContainer" containerID="0764f0b45f4c9a4cc4d1c4342d5f4f2d448117f396e64fdd31e641a7ca96dd03" Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.848078 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-64dc894fb9-5xf8b" Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.849409 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-64dc894fb9-5xf8b" event={"ID":"f9398a81-673d-46b4-b560-91c6e2885532","Type":"ContainerDied","Data":"72b85ac1dd8b375e0cbd78d7a4025b3b5766312305a02ad8d0e6bd5cfeec8ee1"} Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.881372 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-5c9c8c87dc-mqk8l" podStartSLOduration=3.518639821 podStartE2EDuration="6.88134981s" podCreationTimestamp="2025-11-25 10:49:47 +0000 UTC" firstStartedPulling="2025-11-25 10:49:49.259349152 +0000 UTC m=+1059.795668999" lastFinishedPulling="2025-11-25 10:49:52.622059141 +0000 UTC m=+1063.158378988" observedRunningTime="2025-11-25 10:49:53.866205652 +0000 UTC m=+1064.402525499" watchObservedRunningTime="2025-11-25 10:49:53.88134981 +0000 UTC m=+1064.417669657" Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.900773 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.901036 4821 scope.go:117] "RemoveContainer" containerID="666f50101704e276ddca6f07195913aeb8365df95b6639c5717f6c54b431976e" Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.924474 4821 scope.go:117] "RemoveContainer" containerID="cf5c6db7d1dca33b423304d3462893904bd7f34d04492e4bdf0db6099734590c" Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.944421 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.973528 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/51b2c969-f09a-4c90-a897-8b6291716ee7-config-data-custom\") pod \"51b2c969-f09a-4c90-a897-8b6291716ee7\" (UID: \"51b2c969-f09a-4c90-a897-8b6291716ee7\") " Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.973679 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51b2c969-f09a-4c90-a897-8b6291716ee7-config-data\") pod \"51b2c969-f09a-4c90-a897-8b6291716ee7\" (UID: \"51b2c969-f09a-4c90-a897-8b6291716ee7\") " Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.973792 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mwxk5\" (UniqueName: \"kubernetes.io/projected/51b2c969-f09a-4c90-a897-8b6291716ee7-kube-api-access-mwxk5\") pod \"51b2c969-f09a-4c90-a897-8b6291716ee7\" (UID: \"51b2c969-f09a-4c90-a897-8b6291716ee7\") " Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.973824 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/51b2c969-f09a-4c90-a897-8b6291716ee7-logs\") pod \"51b2c969-f09a-4c90-a897-8b6291716ee7\" (UID: \"51b2c969-f09a-4c90-a897-8b6291716ee7\") " Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.973868 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/51b2c969-f09a-4c90-a897-8b6291716ee7-etc-machine-id\") pod \"51b2c969-f09a-4c90-a897-8b6291716ee7\" (UID: \"51b2c969-f09a-4c90-a897-8b6291716ee7\") " Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.973901 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51b2c969-f09a-4c90-a897-8b6291716ee7-combined-ca-bundle\") pod \"51b2c969-f09a-4c90-a897-8b6291716ee7\" (UID: \"51b2c969-f09a-4c90-a897-8b6291716ee7\") " Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.973924 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/51b2c969-f09a-4c90-a897-8b6291716ee7-scripts\") pod \"51b2c969-f09a-4c90-a897-8b6291716ee7\" (UID: \"51b2c969-f09a-4c90-a897-8b6291716ee7\") " Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.975144 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/51b2c969-f09a-4c90-a897-8b6291716ee7-logs" (OuterVolumeSpecName: "logs") pod "51b2c969-f09a-4c90-a897-8b6291716ee7" (UID: "51b2c969-f09a-4c90-a897-8b6291716ee7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.975175 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/51b2c969-f09a-4c90-a897-8b6291716ee7-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "51b2c969-f09a-4c90-a897-8b6291716ee7" (UID: "51b2c969-f09a-4c90-a897-8b6291716ee7"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.977109 4821 scope.go:117] "RemoveContainer" containerID="0764f0b45f4c9a4cc4d1c4342d5f4f2d448117f396e64fdd31e641a7ca96dd03" Nov 25 10:49:53 crc kubenswrapper[4821]: E1125 10:49:53.977852 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0764f0b45f4c9a4cc4d1c4342d5f4f2d448117f396e64fdd31e641a7ca96dd03\": container with ID starting with 0764f0b45f4c9a4cc4d1c4342d5f4f2d448117f396e64fdd31e641a7ca96dd03 not found: ID does not exist" containerID="0764f0b45f4c9a4cc4d1c4342d5f4f2d448117f396e64fdd31e641a7ca96dd03" Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.977919 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0764f0b45f4c9a4cc4d1c4342d5f4f2d448117f396e64fdd31e641a7ca96dd03"} err="failed to get container status \"0764f0b45f4c9a4cc4d1c4342d5f4f2d448117f396e64fdd31e641a7ca96dd03\": rpc error: code = NotFound desc = could not find container \"0764f0b45f4c9a4cc4d1c4342d5f4f2d448117f396e64fdd31e641a7ca96dd03\": container with ID starting with 0764f0b45f4c9a4cc4d1c4342d5f4f2d448117f396e64fdd31e641a7ca96dd03 not found: ID does not exist" Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.977963 4821 scope.go:117] "RemoveContainer" containerID="666f50101704e276ddca6f07195913aeb8365df95b6639c5717f6c54b431976e" Nov 25 10:49:53 crc kubenswrapper[4821]: E1125 10:49:53.978427 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"666f50101704e276ddca6f07195913aeb8365df95b6639c5717f6c54b431976e\": container with ID starting with 666f50101704e276ddca6f07195913aeb8365df95b6639c5717f6c54b431976e not found: ID does not exist" containerID="666f50101704e276ddca6f07195913aeb8365df95b6639c5717f6c54b431976e" Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.978443 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"666f50101704e276ddca6f07195913aeb8365df95b6639c5717f6c54b431976e"} err="failed to get container status \"666f50101704e276ddca6f07195913aeb8365df95b6639c5717f6c54b431976e\": rpc error: code = NotFound desc = could not find container \"666f50101704e276ddca6f07195913aeb8365df95b6639c5717f6c54b431976e\": container with ID starting with 666f50101704e276ddca6f07195913aeb8365df95b6639c5717f6c54b431976e not found: ID does not exist" Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.978456 4821 scope.go:117] "RemoveContainer" containerID="cf5c6db7d1dca33b423304d3462893904bd7f34d04492e4bdf0db6099734590c" Nov 25 10:49:53 crc kubenswrapper[4821]: E1125 10:49:53.978667 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf5c6db7d1dca33b423304d3462893904bd7f34d04492e4bdf0db6099734590c\": container with ID starting with cf5c6db7d1dca33b423304d3462893904bd7f34d04492e4bdf0db6099734590c not found: ID does not exist" containerID="cf5c6db7d1dca33b423304d3462893904bd7f34d04492e4bdf0db6099734590c" Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.978683 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf5c6db7d1dca33b423304d3462893904bd7f34d04492e4bdf0db6099734590c"} err="failed to get container status \"cf5c6db7d1dca33b423304d3462893904bd7f34d04492e4bdf0db6099734590c\": rpc error: code = NotFound desc = could not find container \"cf5c6db7d1dca33b423304d3462893904bd7f34d04492e4bdf0db6099734590c\": container with ID starting with cf5c6db7d1dca33b423304d3462893904bd7f34d04492e4bdf0db6099734590c not found: ID does not exist" Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.978695 4821 scope.go:117] "RemoveContainer" containerID="71773c54cfbe03e75458bf24e2a51cb38377a4ced5a4d0293161a4ebc1ea7207" Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.986519 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51b2c969-f09a-4c90-a897-8b6291716ee7-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "51b2c969-f09a-4c90-a897-8b6291716ee7" (UID: "51b2c969-f09a-4c90-a897-8b6291716ee7"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.987145 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51b2c969-f09a-4c90-a897-8b6291716ee7-scripts" (OuterVolumeSpecName: "scripts") pod "51b2c969-f09a-4c90-a897-8b6291716ee7" (UID: "51b2c969-f09a-4c90-a897-8b6291716ee7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.987716 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51b2c969-f09a-4c90-a897-8b6291716ee7-kube-api-access-mwxk5" (OuterVolumeSpecName: "kube-api-access-mwxk5") pod "51b2c969-f09a-4c90-a897-8b6291716ee7" (UID: "51b2c969-f09a-4c90-a897-8b6291716ee7"). InnerVolumeSpecName "kube-api-access-mwxk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.995350 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:49:53 crc kubenswrapper[4821]: I1125 10:49:53.995406 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-64dc894fb9-5xf8b"] Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.003983 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-64dc894fb9-5xf8b"] Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.010041 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:49:54 crc kubenswrapper[4821]: E1125 10:49:54.010856 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77e0b5ce-8867-4d7e-b89e-170e67f3a548" containerName="horizon-log" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.010893 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="77e0b5ce-8867-4d7e-b89e-170e67f3a548" containerName="horizon-log" Nov 25 10:49:54 crc kubenswrapper[4821]: E1125 10:49:54.010910 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cb63eff-3e7e-436e-aea6-ce8254c3db77" containerName="sg-core" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.010917 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cb63eff-3e7e-436e-aea6-ce8254c3db77" containerName="sg-core" Nov 25 10:49:54 crc kubenswrapper[4821]: E1125 10:49:54.010937 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79a12e89-074b-48f2-bf44-b177fe2e84e2" containerName="horizon" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.010943 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="79a12e89-074b-48f2-bf44-b177fe2e84e2" containerName="horizon" Nov 25 10:49:54 crc kubenswrapper[4821]: E1125 10:49:54.010975 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9398a81-673d-46b4-b560-91c6e2885532" containerName="horizon-log" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.010983 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9398a81-673d-46b4-b560-91c6e2885532" containerName="horizon-log" Nov 25 10:49:54 crc kubenswrapper[4821]: E1125 10:49:54.010994 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79a12e89-074b-48f2-bf44-b177fe2e84e2" containerName="horizon-log" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.011000 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="79a12e89-074b-48f2-bf44-b177fe2e84e2" containerName="horizon-log" Nov 25 10:49:54 crc kubenswrapper[4821]: E1125 10:49:54.011011 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9398a81-673d-46b4-b560-91c6e2885532" containerName="horizon" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.011017 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9398a81-673d-46b4-b560-91c6e2885532" containerName="horizon" Nov 25 10:49:54 crc kubenswrapper[4821]: E1125 10:49:54.011026 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cb63eff-3e7e-436e-aea6-ce8254c3db77" containerName="proxy-httpd" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.011052 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cb63eff-3e7e-436e-aea6-ce8254c3db77" containerName="proxy-httpd" Nov 25 10:49:54 crc kubenswrapper[4821]: E1125 10:49:54.011075 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cb63eff-3e7e-436e-aea6-ce8254c3db77" containerName="ceilometer-notification-agent" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.011082 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cb63eff-3e7e-436e-aea6-ce8254c3db77" containerName="ceilometer-notification-agent" Nov 25 10:49:54 crc kubenswrapper[4821]: E1125 10:49:54.011095 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51b2c969-f09a-4c90-a897-8b6291716ee7" containerName="cinder-api-log" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.011100 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="51b2c969-f09a-4c90-a897-8b6291716ee7" containerName="cinder-api-log" Nov 25 10:49:54 crc kubenswrapper[4821]: E1125 10:49:54.011108 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77e0b5ce-8867-4d7e-b89e-170e67f3a548" containerName="horizon" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.011115 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="77e0b5ce-8867-4d7e-b89e-170e67f3a548" containerName="horizon" Nov 25 10:49:54 crc kubenswrapper[4821]: E1125 10:49:54.011125 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51b2c969-f09a-4c90-a897-8b6291716ee7" containerName="cinder-api" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.011130 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="51b2c969-f09a-4c90-a897-8b6291716ee7" containerName="cinder-api" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.011338 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9398a81-673d-46b4-b560-91c6e2885532" containerName="horizon-log" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.011353 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="9cb63eff-3e7e-436e-aea6-ce8254c3db77" containerName="sg-core" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.011369 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="79a12e89-074b-48f2-bf44-b177fe2e84e2" containerName="horizon-log" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.011395 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="77e0b5ce-8867-4d7e-b89e-170e67f3a548" containerName="horizon-log" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.011406 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="51b2c969-f09a-4c90-a897-8b6291716ee7" containerName="cinder-api" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.011532 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="77e0b5ce-8867-4d7e-b89e-170e67f3a548" containerName="horizon" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.011546 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="9cb63eff-3e7e-436e-aea6-ce8254c3db77" containerName="ceilometer-notification-agent" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.011592 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="79a12e89-074b-48f2-bf44-b177fe2e84e2" containerName="horizon" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.011632 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="51b2c969-f09a-4c90-a897-8b6291716ee7" containerName="cinder-api-log" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.011641 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="9cb63eff-3e7e-436e-aea6-ce8254c3db77" containerName="proxy-httpd" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.011650 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9398a81-673d-46b4-b560-91c6e2885532" containerName="horizon" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.013409 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.017945 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51b2c969-f09a-4c90-a897-8b6291716ee7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "51b2c969-f09a-4c90-a897-8b6291716ee7" (UID: "51b2c969-f09a-4c90-a897-8b6291716ee7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.021610 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.023430 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.024509 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.068418 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51b2c969-f09a-4c90-a897-8b6291716ee7-config-data" (OuterVolumeSpecName: "config-data") pod "51b2c969-f09a-4c90-a897-8b6291716ee7" (UID: "51b2c969-f09a-4c90-a897-8b6291716ee7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.075447 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1647cf99-f6d6-421d-a00b-86538c7a7575-scripts\") pod \"ceilometer-0\" (UID: \"1647cf99-f6d6-421d-a00b-86538c7a7575\") " pod="openstack/ceilometer-0" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.075505 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1647cf99-f6d6-421d-a00b-86538c7a7575-config-data\") pod \"ceilometer-0\" (UID: \"1647cf99-f6d6-421d-a00b-86538c7a7575\") " pod="openstack/ceilometer-0" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.075532 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1647cf99-f6d6-421d-a00b-86538c7a7575-run-httpd\") pod \"ceilometer-0\" (UID: \"1647cf99-f6d6-421d-a00b-86538c7a7575\") " pod="openstack/ceilometer-0" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.075554 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1647cf99-f6d6-421d-a00b-86538c7a7575-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1647cf99-f6d6-421d-a00b-86538c7a7575\") " pod="openstack/ceilometer-0" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.075657 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pm9qx\" (UniqueName: \"kubernetes.io/projected/1647cf99-f6d6-421d-a00b-86538c7a7575-kube-api-access-pm9qx\") pod \"ceilometer-0\" (UID: \"1647cf99-f6d6-421d-a00b-86538c7a7575\") " pod="openstack/ceilometer-0" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.075694 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1647cf99-f6d6-421d-a00b-86538c7a7575-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1647cf99-f6d6-421d-a00b-86538c7a7575\") " pod="openstack/ceilometer-0" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.075769 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1647cf99-f6d6-421d-a00b-86538c7a7575-log-httpd\") pod \"ceilometer-0\" (UID: \"1647cf99-f6d6-421d-a00b-86538c7a7575\") " pod="openstack/ceilometer-0" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.075875 4821 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/51b2c969-f09a-4c90-a897-8b6291716ee7-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.075887 4821 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51b2c969-f09a-4c90-a897-8b6291716ee7-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.075898 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mwxk5\" (UniqueName: \"kubernetes.io/projected/51b2c969-f09a-4c90-a897-8b6291716ee7-kube-api-access-mwxk5\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.075906 4821 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/51b2c969-f09a-4c90-a897-8b6291716ee7-logs\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.075916 4821 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/51b2c969-f09a-4c90-a897-8b6291716ee7-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.075924 4821 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51b2c969-f09a-4c90-a897-8b6291716ee7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.075933 4821 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/51b2c969-f09a-4c90-a897-8b6291716ee7-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.124213 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="77e0b5ce-8867-4d7e-b89e-170e67f3a548" path="/var/lib/kubelet/pods/77e0b5ce-8867-4d7e-b89e-170e67f3a548/volumes" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.125118 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79a12e89-074b-48f2-bf44-b177fe2e84e2" path="/var/lib/kubelet/pods/79a12e89-074b-48f2-bf44-b177fe2e84e2/volumes" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.125750 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9cb63eff-3e7e-436e-aea6-ce8254c3db77" path="/var/lib/kubelet/pods/9cb63eff-3e7e-436e-aea6-ce8254c3db77/volumes" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.127331 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f9398a81-673d-46b4-b560-91c6e2885532" path="/var/lib/kubelet/pods/f9398a81-673d-46b4-b560-91c6e2885532/volumes" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.177059 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1647cf99-f6d6-421d-a00b-86538c7a7575-config-data\") pod \"ceilometer-0\" (UID: \"1647cf99-f6d6-421d-a00b-86538c7a7575\") " pod="openstack/ceilometer-0" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.177773 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1647cf99-f6d6-421d-a00b-86538c7a7575-run-httpd\") pod \"ceilometer-0\" (UID: \"1647cf99-f6d6-421d-a00b-86538c7a7575\") " pod="openstack/ceilometer-0" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.177121 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1647cf99-f6d6-421d-a00b-86538c7a7575-run-httpd\") pod \"ceilometer-0\" (UID: \"1647cf99-f6d6-421d-a00b-86538c7a7575\") " pod="openstack/ceilometer-0" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.178494 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1647cf99-f6d6-421d-a00b-86538c7a7575-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1647cf99-f6d6-421d-a00b-86538c7a7575\") " pod="openstack/ceilometer-0" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.179037 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pm9qx\" (UniqueName: \"kubernetes.io/projected/1647cf99-f6d6-421d-a00b-86538c7a7575-kube-api-access-pm9qx\") pod \"ceilometer-0\" (UID: \"1647cf99-f6d6-421d-a00b-86538c7a7575\") " pod="openstack/ceilometer-0" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.179322 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1647cf99-f6d6-421d-a00b-86538c7a7575-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1647cf99-f6d6-421d-a00b-86538c7a7575\") " pod="openstack/ceilometer-0" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.179637 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1647cf99-f6d6-421d-a00b-86538c7a7575-log-httpd\") pod \"ceilometer-0\" (UID: \"1647cf99-f6d6-421d-a00b-86538c7a7575\") " pod="openstack/ceilometer-0" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.179760 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1647cf99-f6d6-421d-a00b-86538c7a7575-scripts\") pod \"ceilometer-0\" (UID: \"1647cf99-f6d6-421d-a00b-86538c7a7575\") " pod="openstack/ceilometer-0" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.180797 4821 scope.go:117] "RemoveContainer" containerID="3561021b89555fbb5e74f9d8972eb9d892f00a8f4362cad3f466a085cb92eab4" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.180152 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1647cf99-f6d6-421d-a00b-86538c7a7575-log-httpd\") pod \"ceilometer-0\" (UID: \"1647cf99-f6d6-421d-a00b-86538c7a7575\") " pod="openstack/ceilometer-0" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.183335 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1647cf99-f6d6-421d-a00b-86538c7a7575-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1647cf99-f6d6-421d-a00b-86538c7a7575\") " pod="openstack/ceilometer-0" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.183536 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1647cf99-f6d6-421d-a00b-86538c7a7575-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1647cf99-f6d6-421d-a00b-86538c7a7575\") " pod="openstack/ceilometer-0" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.183653 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1647cf99-f6d6-421d-a00b-86538c7a7575-config-data\") pod \"ceilometer-0\" (UID: \"1647cf99-f6d6-421d-a00b-86538c7a7575\") " pod="openstack/ceilometer-0" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.184775 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1647cf99-f6d6-421d-a00b-86538c7a7575-scripts\") pod \"ceilometer-0\" (UID: \"1647cf99-f6d6-421d-a00b-86538c7a7575\") " pod="openstack/ceilometer-0" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.198449 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pm9qx\" (UniqueName: \"kubernetes.io/projected/1647cf99-f6d6-421d-a00b-86538c7a7575-kube-api-access-pm9qx\") pod \"ceilometer-0\" (UID: \"1647cf99-f6d6-421d-a00b-86538c7a7575\") " pod="openstack/ceilometer-0" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.348136 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.403899 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-67b899d6f8-pscf9"] Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.405372 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-67b899d6f8-pscf9" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.411710 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.411847 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.424613 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-67b899d6f8-pscf9"] Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.485101 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b5bf27f5-8585-42d9-9d58-3c6e7f3a34d5-logs\") pod \"barbican-api-67b899d6f8-pscf9\" (UID: \"b5bf27f5-8585-42d9-9d58-3c6e7f3a34d5\") " pod="openstack/barbican-api-67b899d6f8-pscf9" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.485501 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5bf27f5-8585-42d9-9d58-3c6e7f3a34d5-config-data\") pod \"barbican-api-67b899d6f8-pscf9\" (UID: \"b5bf27f5-8585-42d9-9d58-3c6e7f3a34d5\") " pod="openstack/barbican-api-67b899d6f8-pscf9" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.485534 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b5bf27f5-8585-42d9-9d58-3c6e7f3a34d5-public-tls-certs\") pod \"barbican-api-67b899d6f8-pscf9\" (UID: \"b5bf27f5-8585-42d9-9d58-3c6e7f3a34d5\") " pod="openstack/barbican-api-67b899d6f8-pscf9" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.485574 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b5bf27f5-8585-42d9-9d58-3c6e7f3a34d5-internal-tls-certs\") pod \"barbican-api-67b899d6f8-pscf9\" (UID: \"b5bf27f5-8585-42d9-9d58-3c6e7f3a34d5\") " pod="openstack/barbican-api-67b899d6f8-pscf9" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.485614 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6w6n\" (UniqueName: \"kubernetes.io/projected/b5bf27f5-8585-42d9-9d58-3c6e7f3a34d5-kube-api-access-c6w6n\") pod \"barbican-api-67b899d6f8-pscf9\" (UID: \"b5bf27f5-8585-42d9-9d58-3c6e7f3a34d5\") " pod="openstack/barbican-api-67b899d6f8-pscf9" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.485638 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5bf27f5-8585-42d9-9d58-3c6e7f3a34d5-combined-ca-bundle\") pod \"barbican-api-67b899d6f8-pscf9\" (UID: \"b5bf27f5-8585-42d9-9d58-3c6e7f3a34d5\") " pod="openstack/barbican-api-67b899d6f8-pscf9" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.485664 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b5bf27f5-8585-42d9-9d58-3c6e7f3a34d5-config-data-custom\") pod \"barbican-api-67b899d6f8-pscf9\" (UID: \"b5bf27f5-8585-42d9-9d58-3c6e7f3a34d5\") " pod="openstack/barbican-api-67b899d6f8-pscf9" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.587689 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b5bf27f5-8585-42d9-9d58-3c6e7f3a34d5-public-tls-certs\") pod \"barbican-api-67b899d6f8-pscf9\" (UID: \"b5bf27f5-8585-42d9-9d58-3c6e7f3a34d5\") " pod="openstack/barbican-api-67b899d6f8-pscf9" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.587751 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b5bf27f5-8585-42d9-9d58-3c6e7f3a34d5-internal-tls-certs\") pod \"barbican-api-67b899d6f8-pscf9\" (UID: \"b5bf27f5-8585-42d9-9d58-3c6e7f3a34d5\") " pod="openstack/barbican-api-67b899d6f8-pscf9" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.587793 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c6w6n\" (UniqueName: \"kubernetes.io/projected/b5bf27f5-8585-42d9-9d58-3c6e7f3a34d5-kube-api-access-c6w6n\") pod \"barbican-api-67b899d6f8-pscf9\" (UID: \"b5bf27f5-8585-42d9-9d58-3c6e7f3a34d5\") " pod="openstack/barbican-api-67b899d6f8-pscf9" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.587815 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5bf27f5-8585-42d9-9d58-3c6e7f3a34d5-combined-ca-bundle\") pod \"barbican-api-67b899d6f8-pscf9\" (UID: \"b5bf27f5-8585-42d9-9d58-3c6e7f3a34d5\") " pod="openstack/barbican-api-67b899d6f8-pscf9" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.587843 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b5bf27f5-8585-42d9-9d58-3c6e7f3a34d5-config-data-custom\") pod \"barbican-api-67b899d6f8-pscf9\" (UID: \"b5bf27f5-8585-42d9-9d58-3c6e7f3a34d5\") " pod="openstack/barbican-api-67b899d6f8-pscf9" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.587867 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b5bf27f5-8585-42d9-9d58-3c6e7f3a34d5-logs\") pod \"barbican-api-67b899d6f8-pscf9\" (UID: \"b5bf27f5-8585-42d9-9d58-3c6e7f3a34d5\") " pod="openstack/barbican-api-67b899d6f8-pscf9" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.587932 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5bf27f5-8585-42d9-9d58-3c6e7f3a34d5-config-data\") pod \"barbican-api-67b899d6f8-pscf9\" (UID: \"b5bf27f5-8585-42d9-9d58-3c6e7f3a34d5\") " pod="openstack/barbican-api-67b899d6f8-pscf9" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.591348 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b5bf27f5-8585-42d9-9d58-3c6e7f3a34d5-logs\") pod \"barbican-api-67b899d6f8-pscf9\" (UID: \"b5bf27f5-8585-42d9-9d58-3c6e7f3a34d5\") " pod="openstack/barbican-api-67b899d6f8-pscf9" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.593896 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5bf27f5-8585-42d9-9d58-3c6e7f3a34d5-combined-ca-bundle\") pod \"barbican-api-67b899d6f8-pscf9\" (UID: \"b5bf27f5-8585-42d9-9d58-3c6e7f3a34d5\") " pod="openstack/barbican-api-67b899d6f8-pscf9" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.594207 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5bf27f5-8585-42d9-9d58-3c6e7f3a34d5-config-data\") pod \"barbican-api-67b899d6f8-pscf9\" (UID: \"b5bf27f5-8585-42d9-9d58-3c6e7f3a34d5\") " pod="openstack/barbican-api-67b899d6f8-pscf9" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.597524 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b5bf27f5-8585-42d9-9d58-3c6e7f3a34d5-internal-tls-certs\") pod \"barbican-api-67b899d6f8-pscf9\" (UID: \"b5bf27f5-8585-42d9-9d58-3c6e7f3a34d5\") " pod="openstack/barbican-api-67b899d6f8-pscf9" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.597926 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b5bf27f5-8585-42d9-9d58-3c6e7f3a34d5-public-tls-certs\") pod \"barbican-api-67b899d6f8-pscf9\" (UID: \"b5bf27f5-8585-42d9-9d58-3c6e7f3a34d5\") " pod="openstack/barbican-api-67b899d6f8-pscf9" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.600883 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b5bf27f5-8585-42d9-9d58-3c6e7f3a34d5-config-data-custom\") pod \"barbican-api-67b899d6f8-pscf9\" (UID: \"b5bf27f5-8585-42d9-9d58-3c6e7f3a34d5\") " pod="openstack/barbican-api-67b899d6f8-pscf9" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.608132 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c6w6n\" (UniqueName: \"kubernetes.io/projected/b5bf27f5-8585-42d9-9d58-3c6e7f3a34d5-kube-api-access-c6w6n\") pod \"barbican-api-67b899d6f8-pscf9\" (UID: \"b5bf27f5-8585-42d9-9d58-3c6e7f3a34d5\") " pod="openstack/barbican-api-67b899d6f8-pscf9" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.775108 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-67b899d6f8-pscf9" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.896514 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-f75f66468-2k6lr" event={"ID":"1e6fee50-9d84-4965-a222-a55f6dcfc4cc","Type":"ContainerStarted","Data":"794e2d3e21741598c2d0226bbbf884a14c5f84310c292b92615ac77c4f3e88e3"} Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.901270 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.959292 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:49:54 crc kubenswrapper[4821]: I1125 10:49:54.960141 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-f75f66468-2k6lr" podStartSLOduration=4.476331892 podStartE2EDuration="7.960131583s" podCreationTimestamp="2025-11-25 10:49:47 +0000 UTC" firstStartedPulling="2025-11-25 10:49:49.132568684 +0000 UTC m=+1059.668888531" lastFinishedPulling="2025-11-25 10:49:52.616368375 +0000 UTC m=+1063.152688222" observedRunningTime="2025-11-25 10:49:54.9315065 +0000 UTC m=+1065.467826347" watchObservedRunningTime="2025-11-25 10:49:54.960131583 +0000 UTC m=+1065.496451430" Nov 25 10:49:55 crc kubenswrapper[4821]: I1125 10:49:55.028304 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 25 10:49:55 crc kubenswrapper[4821]: I1125 10:49:55.049600 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Nov 25 10:49:55 crc kubenswrapper[4821]: I1125 10:49:55.058223 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 25 10:49:55 crc kubenswrapper[4821]: I1125 10:49:55.072585 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 25 10:49:55 crc kubenswrapper[4821]: I1125 10:49:55.078724 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 25 10:49:55 crc kubenswrapper[4821]: I1125 10:49:55.079496 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Nov 25 10:49:55 crc kubenswrapper[4821]: I1125 10:49:55.079867 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Nov 25 10:49:55 crc kubenswrapper[4821]: I1125 10:49:55.079976 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 25 10:49:55 crc kubenswrapper[4821]: I1125 10:49:55.201195 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4fd0ccca-05dd-43e3-80b1-1c312d130056-config-data\") pod \"cinder-api-0\" (UID: \"4fd0ccca-05dd-43e3-80b1-1c312d130056\") " pod="openstack/cinder-api-0" Nov 25 10:49:55 crc kubenswrapper[4821]: I1125 10:49:55.201447 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4fd0ccca-05dd-43e3-80b1-1c312d130056-config-data-custom\") pod \"cinder-api-0\" (UID: \"4fd0ccca-05dd-43e3-80b1-1c312d130056\") " pod="openstack/cinder-api-0" Nov 25 10:49:55 crc kubenswrapper[4821]: I1125 10:49:55.201523 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4fd0ccca-05dd-43e3-80b1-1c312d130056-etc-machine-id\") pod \"cinder-api-0\" (UID: \"4fd0ccca-05dd-43e3-80b1-1c312d130056\") " pod="openstack/cinder-api-0" Nov 25 10:49:55 crc kubenswrapper[4821]: I1125 10:49:55.201574 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4fd0ccca-05dd-43e3-80b1-1c312d130056-scripts\") pod \"cinder-api-0\" (UID: \"4fd0ccca-05dd-43e3-80b1-1c312d130056\") " pod="openstack/cinder-api-0" Nov 25 10:49:55 crc kubenswrapper[4821]: I1125 10:49:55.201653 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4fd0ccca-05dd-43e3-80b1-1c312d130056-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"4fd0ccca-05dd-43e3-80b1-1c312d130056\") " pod="openstack/cinder-api-0" Nov 25 10:49:55 crc kubenswrapper[4821]: I1125 10:49:55.201751 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46kwc\" (UniqueName: \"kubernetes.io/projected/4fd0ccca-05dd-43e3-80b1-1c312d130056-kube-api-access-46kwc\") pod \"cinder-api-0\" (UID: \"4fd0ccca-05dd-43e3-80b1-1c312d130056\") " pod="openstack/cinder-api-0" Nov 25 10:49:55 crc kubenswrapper[4821]: I1125 10:49:55.201797 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4fd0ccca-05dd-43e3-80b1-1c312d130056-public-tls-certs\") pod \"cinder-api-0\" (UID: \"4fd0ccca-05dd-43e3-80b1-1c312d130056\") " pod="openstack/cinder-api-0" Nov 25 10:49:55 crc kubenswrapper[4821]: I1125 10:49:55.201830 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4fd0ccca-05dd-43e3-80b1-1c312d130056-logs\") pod \"cinder-api-0\" (UID: \"4fd0ccca-05dd-43e3-80b1-1c312d130056\") " pod="openstack/cinder-api-0" Nov 25 10:49:55 crc kubenswrapper[4821]: I1125 10:49:55.201877 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4fd0ccca-05dd-43e3-80b1-1c312d130056-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"4fd0ccca-05dd-43e3-80b1-1c312d130056\") " pod="openstack/cinder-api-0" Nov 25 10:49:55 crc kubenswrapper[4821]: I1125 10:49:55.303098 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4fd0ccca-05dd-43e3-80b1-1c312d130056-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"4fd0ccca-05dd-43e3-80b1-1c312d130056\") " pod="openstack/cinder-api-0" Nov 25 10:49:55 crc kubenswrapper[4821]: I1125 10:49:55.303389 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4fd0ccca-05dd-43e3-80b1-1c312d130056-config-data\") pod \"cinder-api-0\" (UID: \"4fd0ccca-05dd-43e3-80b1-1c312d130056\") " pod="openstack/cinder-api-0" Nov 25 10:49:55 crc kubenswrapper[4821]: I1125 10:49:55.303539 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4fd0ccca-05dd-43e3-80b1-1c312d130056-config-data-custom\") pod \"cinder-api-0\" (UID: \"4fd0ccca-05dd-43e3-80b1-1c312d130056\") " pod="openstack/cinder-api-0" Nov 25 10:49:55 crc kubenswrapper[4821]: I1125 10:49:55.303681 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4fd0ccca-05dd-43e3-80b1-1c312d130056-etc-machine-id\") pod \"cinder-api-0\" (UID: \"4fd0ccca-05dd-43e3-80b1-1c312d130056\") " pod="openstack/cinder-api-0" Nov 25 10:49:55 crc kubenswrapper[4821]: I1125 10:49:55.303801 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4fd0ccca-05dd-43e3-80b1-1c312d130056-scripts\") pod \"cinder-api-0\" (UID: \"4fd0ccca-05dd-43e3-80b1-1c312d130056\") " pod="openstack/cinder-api-0" Nov 25 10:49:55 crc kubenswrapper[4821]: I1125 10:49:55.303910 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4fd0ccca-05dd-43e3-80b1-1c312d130056-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"4fd0ccca-05dd-43e3-80b1-1c312d130056\") " pod="openstack/cinder-api-0" Nov 25 10:49:55 crc kubenswrapper[4821]: I1125 10:49:55.304030 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-46kwc\" (UniqueName: \"kubernetes.io/projected/4fd0ccca-05dd-43e3-80b1-1c312d130056-kube-api-access-46kwc\") pod \"cinder-api-0\" (UID: \"4fd0ccca-05dd-43e3-80b1-1c312d130056\") " pod="openstack/cinder-api-0" Nov 25 10:49:55 crc kubenswrapper[4821]: I1125 10:49:55.304153 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4fd0ccca-05dd-43e3-80b1-1c312d130056-public-tls-certs\") pod \"cinder-api-0\" (UID: \"4fd0ccca-05dd-43e3-80b1-1c312d130056\") " pod="openstack/cinder-api-0" Nov 25 10:49:55 crc kubenswrapper[4821]: I1125 10:49:55.304280 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4fd0ccca-05dd-43e3-80b1-1c312d130056-logs\") pod \"cinder-api-0\" (UID: \"4fd0ccca-05dd-43e3-80b1-1c312d130056\") " pod="openstack/cinder-api-0" Nov 25 10:49:55 crc kubenswrapper[4821]: I1125 10:49:55.305008 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4fd0ccca-05dd-43e3-80b1-1c312d130056-logs\") pod \"cinder-api-0\" (UID: \"4fd0ccca-05dd-43e3-80b1-1c312d130056\") " pod="openstack/cinder-api-0" Nov 25 10:49:55 crc kubenswrapper[4821]: I1125 10:49:55.306968 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4fd0ccca-05dd-43e3-80b1-1c312d130056-etc-machine-id\") pod \"cinder-api-0\" (UID: \"4fd0ccca-05dd-43e3-80b1-1c312d130056\") " pod="openstack/cinder-api-0" Nov 25 10:49:55 crc kubenswrapper[4821]: I1125 10:49:55.312938 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4fd0ccca-05dd-43e3-80b1-1c312d130056-scripts\") pod \"cinder-api-0\" (UID: \"4fd0ccca-05dd-43e3-80b1-1c312d130056\") " pod="openstack/cinder-api-0" Nov 25 10:49:55 crc kubenswrapper[4821]: I1125 10:49:55.312970 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4fd0ccca-05dd-43e3-80b1-1c312d130056-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"4fd0ccca-05dd-43e3-80b1-1c312d130056\") " pod="openstack/cinder-api-0" Nov 25 10:49:55 crc kubenswrapper[4821]: I1125 10:49:55.313046 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4fd0ccca-05dd-43e3-80b1-1c312d130056-config-data-custom\") pod \"cinder-api-0\" (UID: \"4fd0ccca-05dd-43e3-80b1-1c312d130056\") " pod="openstack/cinder-api-0" Nov 25 10:49:55 crc kubenswrapper[4821]: I1125 10:49:55.314998 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4fd0ccca-05dd-43e3-80b1-1c312d130056-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"4fd0ccca-05dd-43e3-80b1-1c312d130056\") " pod="openstack/cinder-api-0" Nov 25 10:49:55 crc kubenswrapper[4821]: I1125 10:49:55.318327 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-67b899d6f8-pscf9"] Nov 25 10:49:55 crc kubenswrapper[4821]: I1125 10:49:55.320839 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4fd0ccca-05dd-43e3-80b1-1c312d130056-config-data\") pod \"cinder-api-0\" (UID: \"4fd0ccca-05dd-43e3-80b1-1c312d130056\") " pod="openstack/cinder-api-0" Nov 25 10:49:55 crc kubenswrapper[4821]: I1125 10:49:55.322027 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4fd0ccca-05dd-43e3-80b1-1c312d130056-public-tls-certs\") pod \"cinder-api-0\" (UID: \"4fd0ccca-05dd-43e3-80b1-1c312d130056\") " pod="openstack/cinder-api-0" Nov 25 10:49:55 crc kubenswrapper[4821]: I1125 10:49:55.329273 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-46kwc\" (UniqueName: \"kubernetes.io/projected/4fd0ccca-05dd-43e3-80b1-1c312d130056-kube-api-access-46kwc\") pod \"cinder-api-0\" (UID: \"4fd0ccca-05dd-43e3-80b1-1c312d130056\") " pod="openstack/cinder-api-0" Nov 25 10:49:55 crc kubenswrapper[4821]: I1125 10:49:55.410269 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 25 10:49:55 crc kubenswrapper[4821]: I1125 10:49:55.880576 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-67f765d856-ncc6t" Nov 25 10:49:55 crc kubenswrapper[4821]: I1125 10:49:55.917610 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 25 10:49:55 crc kubenswrapper[4821]: I1125 10:49:55.923672 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-67b899d6f8-pscf9" event={"ID":"b5bf27f5-8585-42d9-9d58-3c6e7f3a34d5","Type":"ContainerStarted","Data":"dfa2214d4e87ce63f7c4dc1b8bd9bb9d76ae91656001d28e01c75b62255a9de5"} Nov 25 10:49:55 crc kubenswrapper[4821]: I1125 10:49:55.923736 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-67b899d6f8-pscf9" event={"ID":"b5bf27f5-8585-42d9-9d58-3c6e7f3a34d5","Type":"ContainerStarted","Data":"a17eceb8d22876a112eda4e0bc0153c862c291ed367fbd89002e45115336e691"} Nov 25 10:49:55 crc kubenswrapper[4821]: I1125 10:49:55.926821 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1647cf99-f6d6-421d-a00b-86538c7a7575","Type":"ContainerStarted","Data":"3c19f2f14248485cd53184e9f83662d006581b8e8df8c8b58a517328fe292758"} Nov 25 10:49:55 crc kubenswrapper[4821]: I1125 10:49:55.985487 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-869c6f5bfb-l785w"] Nov 25 10:49:55 crc kubenswrapper[4821]: I1125 10:49:55.985763 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-869c6f5bfb-l785w" podUID="e2327ad3-5d88-4f4b-91ea-91deaf7fc249" containerName="horizon-log" containerID="cri-o://6e8007ccc7348305c4f9b913d2dfa8dbf30b5f4417db7e25937dc070c2fa5264" gracePeriod=30 Nov 25 10:49:55 crc kubenswrapper[4821]: I1125 10:49:55.986244 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-869c6f5bfb-l785w" podUID="e2327ad3-5d88-4f4b-91ea-91deaf7fc249" containerName="horizon" containerID="cri-o://0d6baf8077dd2e1ec034e0a15bb7d7957d591590cdfe3a7b244fa1f9081cdfcc" gracePeriod=30 Nov 25 10:49:56 crc kubenswrapper[4821]: I1125 10:49:56.014980 4821 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-869c6f5bfb-l785w" podUID="e2327ad3-5d88-4f4b-91ea-91deaf7fc249" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": EOF" Nov 25 10:49:56 crc kubenswrapper[4821]: I1125 10:49:56.134740 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51b2c969-f09a-4c90-a897-8b6291716ee7" path="/var/lib/kubelet/pods/51b2c969-f09a-4c90-a897-8b6291716ee7/volumes" Nov 25 10:49:56 crc kubenswrapper[4821]: I1125 10:49:56.944905 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1647cf99-f6d6-421d-a00b-86538c7a7575","Type":"ContainerStarted","Data":"e7403b14fd72da5014034f95716d4e786968445593065d9e47e2dba3a1b41e4b"} Nov 25 10:49:56 crc kubenswrapper[4821]: I1125 10:49:56.949844 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"4fd0ccca-05dd-43e3-80b1-1c312d130056","Type":"ContainerStarted","Data":"3a3467c680d8fdbaa3fb7194c8e23cd4ff906e4bd7889e62e537bd5f098adaa7"} Nov 25 10:49:56 crc kubenswrapper[4821]: I1125 10:49:56.949916 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"4fd0ccca-05dd-43e3-80b1-1c312d130056","Type":"ContainerStarted","Data":"c6623609cea977857369deea475e2bdebbbeba35b664bd5c7bf703c4217de92f"} Nov 25 10:49:56 crc kubenswrapper[4821]: I1125 10:49:56.952219 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-67b899d6f8-pscf9" event={"ID":"b5bf27f5-8585-42d9-9d58-3c6e7f3a34d5","Type":"ContainerStarted","Data":"434e9ed288f2028e91bbd3bf0fe2a5bf5f992b60b9a227b736d671c65709b839"} Nov 25 10:49:56 crc kubenswrapper[4821]: I1125 10:49:56.954079 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-67b899d6f8-pscf9" Nov 25 10:49:56 crc kubenswrapper[4821]: I1125 10:49:56.954130 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-67b899d6f8-pscf9" Nov 25 10:49:56 crc kubenswrapper[4821]: I1125 10:49:56.980631 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-67b899d6f8-pscf9" podStartSLOduration=2.980585499 podStartE2EDuration="2.980585499s" podCreationTimestamp="2025-11-25 10:49:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:49:56.976706511 +0000 UTC m=+1067.513026368" watchObservedRunningTime="2025-11-25 10:49:56.980585499 +0000 UTC m=+1067.516905346" Nov 25 10:49:57 crc kubenswrapper[4821]: I1125 10:49:57.967016 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1647cf99-f6d6-421d-a00b-86538c7a7575","Type":"ContainerStarted","Data":"0073627f3d04f03ca74fbb9e3c020f654c964b11997f56a69f24e0f4144dd25f"} Nov 25 10:49:57 crc kubenswrapper[4821]: I1125 10:49:57.970700 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"4fd0ccca-05dd-43e3-80b1-1c312d130056","Type":"ContainerStarted","Data":"a05d0e4b22adbff0f1cbff45c85d3f2a58cfcca077254b98d8913def3fe586a5"} Nov 25 10:49:57 crc kubenswrapper[4821]: I1125 10:49:57.970796 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 25 10:49:57 crc kubenswrapper[4821]: I1125 10:49:57.995040 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.995016466 podStartE2EDuration="3.995016466s" podCreationTimestamp="2025-11-25 10:49:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:49:57.986541038 +0000 UTC m=+1068.522860885" watchObservedRunningTime="2025-11-25 10:49:57.995016466 +0000 UTC m=+1068.531336313" Nov 25 10:49:58 crc kubenswrapper[4821]: I1125 10:49:58.057595 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 25 10:49:58 crc kubenswrapper[4821]: I1125 10:49:58.110648 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 25 10:49:58 crc kubenswrapper[4821]: I1125 10:49:58.742375 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6578955fd5-svh8j" Nov 25 10:49:58 crc kubenswrapper[4821]: I1125 10:49:58.807835 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-dpjr6"] Nov 25 10:49:58 crc kubenswrapper[4821]: I1125 10:49:58.808128 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6b7b667979-dpjr6" podUID="53f9f168-1063-4bb0-8d1c-03482761ae7f" containerName="dnsmasq-dns" containerID="cri-o://065de96b781ff5e2a331bea1a4e20284824bef6044fdf9fc5450f3ff83be103c" gracePeriod=10 Nov 25 10:49:58 crc kubenswrapper[4821]: I1125 10:49:58.981713 4821 generic.go:334] "Generic (PLEG): container finished" podID="53f9f168-1063-4bb0-8d1c-03482761ae7f" containerID="065de96b781ff5e2a331bea1a4e20284824bef6044fdf9fc5450f3ff83be103c" exitCode=0 Nov 25 10:49:58 crc kubenswrapper[4821]: I1125 10:49:58.982035 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-dpjr6" event={"ID":"53f9f168-1063-4bb0-8d1c-03482761ae7f","Type":"ContainerDied","Data":"065de96b781ff5e2a331bea1a4e20284824bef6044fdf9fc5450f3ff83be103c"} Nov 25 10:49:58 crc kubenswrapper[4821]: I1125 10:49:58.989371 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1647cf99-f6d6-421d-a00b-86538c7a7575","Type":"ContainerStarted","Data":"7b894abf033fdc87c1547963b5822943b6b998050779618be7999ac3afa4aa3b"} Nov 25 10:49:58 crc kubenswrapper[4821]: I1125 10:49:58.989621 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="4f5bff28-70e9-4b56-ab38-49511f85852d" containerName="cinder-scheduler" containerID="cri-o://a4e878765149277342071df06b9a3d4de0e2483bb18a27e40f85f6f2947fde9b" gracePeriod=30 Nov 25 10:49:58 crc kubenswrapper[4821]: I1125 10:49:58.991042 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="4f5bff28-70e9-4b56-ab38-49511f85852d" containerName="probe" containerID="cri-o://42e6a15ed6523ba17eead013e533d86e41662911fe061607b0113ad40f221233" gracePeriod=30 Nov 25 10:49:59 crc kubenswrapper[4821]: I1125 10:49:59.481395 4821 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-869c6f5bfb-l785w" podUID="e2327ad3-5d88-4f4b-91ea-91deaf7fc249" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": read tcp 10.217.0.2:37458->10.217.0.148:8443: read: connection reset by peer" Nov 25 10:49:59 crc kubenswrapper[4821]: I1125 10:49:59.518937 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-dpjr6" Nov 25 10:49:59 crc kubenswrapper[4821]: I1125 10:49:59.630241 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cbgd9\" (UniqueName: \"kubernetes.io/projected/53f9f168-1063-4bb0-8d1c-03482761ae7f-kube-api-access-cbgd9\") pod \"53f9f168-1063-4bb0-8d1c-03482761ae7f\" (UID: \"53f9f168-1063-4bb0-8d1c-03482761ae7f\") " Nov 25 10:49:59 crc kubenswrapper[4821]: I1125 10:49:59.630405 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/53f9f168-1063-4bb0-8d1c-03482761ae7f-dns-svc\") pod \"53f9f168-1063-4bb0-8d1c-03482761ae7f\" (UID: \"53f9f168-1063-4bb0-8d1c-03482761ae7f\") " Nov 25 10:49:59 crc kubenswrapper[4821]: I1125 10:49:59.631189 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53f9f168-1063-4bb0-8d1c-03482761ae7f-config\") pod \"53f9f168-1063-4bb0-8d1c-03482761ae7f\" (UID: \"53f9f168-1063-4bb0-8d1c-03482761ae7f\") " Nov 25 10:49:59 crc kubenswrapper[4821]: I1125 10:49:59.631227 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/53f9f168-1063-4bb0-8d1c-03482761ae7f-ovsdbserver-sb\") pod \"53f9f168-1063-4bb0-8d1c-03482761ae7f\" (UID: \"53f9f168-1063-4bb0-8d1c-03482761ae7f\") " Nov 25 10:49:59 crc kubenswrapper[4821]: I1125 10:49:59.631282 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/53f9f168-1063-4bb0-8d1c-03482761ae7f-ovsdbserver-nb\") pod \"53f9f168-1063-4bb0-8d1c-03482761ae7f\" (UID: \"53f9f168-1063-4bb0-8d1c-03482761ae7f\") " Nov 25 10:49:59 crc kubenswrapper[4821]: I1125 10:49:59.631312 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/53f9f168-1063-4bb0-8d1c-03482761ae7f-dns-swift-storage-0\") pod \"53f9f168-1063-4bb0-8d1c-03482761ae7f\" (UID: \"53f9f168-1063-4bb0-8d1c-03482761ae7f\") " Nov 25 10:49:59 crc kubenswrapper[4821]: I1125 10:49:59.642415 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53f9f168-1063-4bb0-8d1c-03482761ae7f-kube-api-access-cbgd9" (OuterVolumeSpecName: "kube-api-access-cbgd9") pod "53f9f168-1063-4bb0-8d1c-03482761ae7f" (UID: "53f9f168-1063-4bb0-8d1c-03482761ae7f"). InnerVolumeSpecName "kube-api-access-cbgd9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:49:59 crc kubenswrapper[4821]: I1125 10:49:59.734896 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cbgd9\" (UniqueName: \"kubernetes.io/projected/53f9f168-1063-4bb0-8d1c-03482761ae7f-kube-api-access-cbgd9\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:59 crc kubenswrapper[4821]: I1125 10:49:59.744918 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53f9f168-1063-4bb0-8d1c-03482761ae7f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "53f9f168-1063-4bb0-8d1c-03482761ae7f" (UID: "53f9f168-1063-4bb0-8d1c-03482761ae7f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:49:59 crc kubenswrapper[4821]: I1125 10:49:59.759731 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53f9f168-1063-4bb0-8d1c-03482761ae7f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "53f9f168-1063-4bb0-8d1c-03482761ae7f" (UID: "53f9f168-1063-4bb0-8d1c-03482761ae7f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:49:59 crc kubenswrapper[4821]: I1125 10:49:59.764565 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53f9f168-1063-4bb0-8d1c-03482761ae7f-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "53f9f168-1063-4bb0-8d1c-03482761ae7f" (UID: "53f9f168-1063-4bb0-8d1c-03482761ae7f"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:49:59 crc kubenswrapper[4821]: E1125 10:49:59.782101 4821 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/53f9f168-1063-4bb0-8d1c-03482761ae7f-config podName:53f9f168-1063-4bb0-8d1c-03482761ae7f nodeName:}" failed. No retries permitted until 2025-11-25 10:50:00.282073233 +0000 UTC m=+1070.818393080 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "config" (UniqueName: "kubernetes.io/configmap/53f9f168-1063-4bb0-8d1c-03482761ae7f-config") pod "53f9f168-1063-4bb0-8d1c-03482761ae7f" (UID: "53f9f168-1063-4bb0-8d1c-03482761ae7f") : error deleting /var/lib/kubelet/pods/53f9f168-1063-4bb0-8d1c-03482761ae7f/volume-subpaths: remove /var/lib/kubelet/pods/53f9f168-1063-4bb0-8d1c-03482761ae7f/volume-subpaths: no such file or directory Nov 25 10:49:59 crc kubenswrapper[4821]: I1125 10:49:59.782387 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53f9f168-1063-4bb0-8d1c-03482761ae7f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "53f9f168-1063-4bb0-8d1c-03482761ae7f" (UID: "53f9f168-1063-4bb0-8d1c-03482761ae7f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:49:59 crc kubenswrapper[4821]: I1125 10:49:59.836470 4821 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/53f9f168-1063-4bb0-8d1c-03482761ae7f-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:59 crc kubenswrapper[4821]: I1125 10:49:59.836510 4821 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/53f9f168-1063-4bb0-8d1c-03482761ae7f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:59 crc kubenswrapper[4821]: I1125 10:49:59.836524 4821 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/53f9f168-1063-4bb0-8d1c-03482761ae7f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:59 crc kubenswrapper[4821]: I1125 10:49:59.836536 4821 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/53f9f168-1063-4bb0-8d1c-03482761ae7f-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:00 crc kubenswrapper[4821]: I1125 10:50:00.001668 4821 generic.go:334] "Generic (PLEG): container finished" podID="e2327ad3-5d88-4f4b-91ea-91deaf7fc249" containerID="0d6baf8077dd2e1ec034e0a15bb7d7957d591590cdfe3a7b244fa1f9081cdfcc" exitCode=0 Nov 25 10:50:00 crc kubenswrapper[4821]: I1125 10:50:00.001747 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-869c6f5bfb-l785w" event={"ID":"e2327ad3-5d88-4f4b-91ea-91deaf7fc249","Type":"ContainerDied","Data":"0d6baf8077dd2e1ec034e0a15bb7d7957d591590cdfe3a7b244fa1f9081cdfcc"} Nov 25 10:50:00 crc kubenswrapper[4821]: I1125 10:50:00.003484 4821 generic.go:334] "Generic (PLEG): container finished" podID="4f5bff28-70e9-4b56-ab38-49511f85852d" containerID="42e6a15ed6523ba17eead013e533d86e41662911fe061607b0113ad40f221233" exitCode=0 Nov 25 10:50:00 crc kubenswrapper[4821]: I1125 10:50:00.003522 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"4f5bff28-70e9-4b56-ab38-49511f85852d","Type":"ContainerDied","Data":"42e6a15ed6523ba17eead013e533d86e41662911fe061607b0113ad40f221233"} Nov 25 10:50:00 crc kubenswrapper[4821]: I1125 10:50:00.006020 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1647cf99-f6d6-421d-a00b-86538c7a7575","Type":"ContainerStarted","Data":"b57c43f68f8bdd2464535935e220533ae71fe723e2b8bf249a896b43a1b9bc9c"} Nov 25 10:50:00 crc kubenswrapper[4821]: I1125 10:50:00.007051 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 25 10:50:00 crc kubenswrapper[4821]: I1125 10:50:00.009683 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-dpjr6" event={"ID":"53f9f168-1063-4bb0-8d1c-03482761ae7f","Type":"ContainerDied","Data":"958ca843b2ab4ec81ead689869cb03397b72daa4519d47bf3f5386e2894b2290"} Nov 25 10:50:00 crc kubenswrapper[4821]: I1125 10:50:00.009718 4821 scope.go:117] "RemoveContainer" containerID="065de96b781ff5e2a331bea1a4e20284824bef6044fdf9fc5450f3ff83be103c" Nov 25 10:50:00 crc kubenswrapper[4821]: I1125 10:50:00.009835 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-dpjr6" Nov 25 10:50:00 crc kubenswrapper[4821]: I1125 10:50:00.036599 4821 scope.go:117] "RemoveContainer" containerID="8e12b5f21ce0e86d4aebb00d928fcd06fd64a985b3d050b09d26a3e16cea1647" Nov 25 10:50:00 crc kubenswrapper[4821]: I1125 10:50:00.040122 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.809946613 podStartE2EDuration="7.040099243s" podCreationTimestamp="2025-11-25 10:49:53 +0000 UTC" firstStartedPulling="2025-11-25 10:49:55.021333051 +0000 UTC m=+1065.557652898" lastFinishedPulling="2025-11-25 10:49:59.251485681 +0000 UTC m=+1069.787805528" observedRunningTime="2025-11-25 10:50:00.028663409 +0000 UTC m=+1070.564983256" watchObservedRunningTime="2025-11-25 10:50:00.040099243 +0000 UTC m=+1070.576419090" Nov 25 10:50:00 crc kubenswrapper[4821]: I1125 10:50:00.343305 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53f9f168-1063-4bb0-8d1c-03482761ae7f-config\") pod \"53f9f168-1063-4bb0-8d1c-03482761ae7f\" (UID: \"53f9f168-1063-4bb0-8d1c-03482761ae7f\") " Nov 25 10:50:00 crc kubenswrapper[4821]: I1125 10:50:00.343760 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53f9f168-1063-4bb0-8d1c-03482761ae7f-config" (OuterVolumeSpecName: "config") pod "53f9f168-1063-4bb0-8d1c-03482761ae7f" (UID: "53f9f168-1063-4bb0-8d1c-03482761ae7f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:50:00 crc kubenswrapper[4821]: I1125 10:50:00.445736 4821 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53f9f168-1063-4bb0-8d1c-03482761ae7f-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:00 crc kubenswrapper[4821]: I1125 10:50:00.648555 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-dpjr6"] Nov 25 10:50:00 crc kubenswrapper[4821]: I1125 10:50:00.649600 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-55c965854d-xpf4q" Nov 25 10:50:00 crc kubenswrapper[4821]: I1125 10:50:00.663583 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-dpjr6"] Nov 25 10:50:00 crc kubenswrapper[4821]: I1125 10:50:00.875214 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-55c965854d-xpf4q" Nov 25 10:50:00 crc kubenswrapper[4821]: I1125 10:50:00.993402 4821 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-869c6f5bfb-l785w" podUID="e2327ad3-5d88-4f4b-91ea-91deaf7fc249" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Nov 25 10:50:02 crc kubenswrapper[4821]: I1125 10:50:02.136056 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53f9f168-1063-4bb0-8d1c-03482761ae7f" path="/var/lib/kubelet/pods/53f9f168-1063-4bb0-8d1c-03482761ae7f/volumes" Nov 25 10:50:02 crc kubenswrapper[4821]: I1125 10:50:02.672765 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 25 10:50:02 crc kubenswrapper[4821]: I1125 10:50:02.786297 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4f5bff28-70e9-4b56-ab38-49511f85852d-config-data-custom\") pod \"4f5bff28-70e9-4b56-ab38-49511f85852d\" (UID: \"4f5bff28-70e9-4b56-ab38-49511f85852d\") " Nov 25 10:50:02 crc kubenswrapper[4821]: I1125 10:50:02.786425 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4f5bff28-70e9-4b56-ab38-49511f85852d-etc-machine-id\") pod \"4f5bff28-70e9-4b56-ab38-49511f85852d\" (UID: \"4f5bff28-70e9-4b56-ab38-49511f85852d\") " Nov 25 10:50:02 crc kubenswrapper[4821]: I1125 10:50:02.786471 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vvqhz\" (UniqueName: \"kubernetes.io/projected/4f5bff28-70e9-4b56-ab38-49511f85852d-kube-api-access-vvqhz\") pod \"4f5bff28-70e9-4b56-ab38-49511f85852d\" (UID: \"4f5bff28-70e9-4b56-ab38-49511f85852d\") " Nov 25 10:50:02 crc kubenswrapper[4821]: I1125 10:50:02.786503 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4f5bff28-70e9-4b56-ab38-49511f85852d-scripts\") pod \"4f5bff28-70e9-4b56-ab38-49511f85852d\" (UID: \"4f5bff28-70e9-4b56-ab38-49511f85852d\") " Nov 25 10:50:02 crc kubenswrapper[4821]: I1125 10:50:02.786586 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f5bff28-70e9-4b56-ab38-49511f85852d-combined-ca-bundle\") pod \"4f5bff28-70e9-4b56-ab38-49511f85852d\" (UID: \"4f5bff28-70e9-4b56-ab38-49511f85852d\") " Nov 25 10:50:02 crc kubenswrapper[4821]: I1125 10:50:02.786706 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f5bff28-70e9-4b56-ab38-49511f85852d-config-data\") pod \"4f5bff28-70e9-4b56-ab38-49511f85852d\" (UID: \"4f5bff28-70e9-4b56-ab38-49511f85852d\") " Nov 25 10:50:02 crc kubenswrapper[4821]: I1125 10:50:02.787028 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4f5bff28-70e9-4b56-ab38-49511f85852d-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "4f5bff28-70e9-4b56-ab38-49511f85852d" (UID: "4f5bff28-70e9-4b56-ab38-49511f85852d"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:50:02 crc kubenswrapper[4821]: I1125 10:50:02.787319 4821 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4f5bff28-70e9-4b56-ab38-49511f85852d-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:02 crc kubenswrapper[4821]: I1125 10:50:02.793259 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f5bff28-70e9-4b56-ab38-49511f85852d-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "4f5bff28-70e9-4b56-ab38-49511f85852d" (UID: "4f5bff28-70e9-4b56-ab38-49511f85852d"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:50:02 crc kubenswrapper[4821]: I1125 10:50:02.793456 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f5bff28-70e9-4b56-ab38-49511f85852d-kube-api-access-vvqhz" (OuterVolumeSpecName: "kube-api-access-vvqhz") pod "4f5bff28-70e9-4b56-ab38-49511f85852d" (UID: "4f5bff28-70e9-4b56-ab38-49511f85852d"). InnerVolumeSpecName "kube-api-access-vvqhz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:50:02 crc kubenswrapper[4821]: I1125 10:50:02.796507 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f5bff28-70e9-4b56-ab38-49511f85852d-scripts" (OuterVolumeSpecName: "scripts") pod "4f5bff28-70e9-4b56-ab38-49511f85852d" (UID: "4f5bff28-70e9-4b56-ab38-49511f85852d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:50:02 crc kubenswrapper[4821]: I1125 10:50:02.861639 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f5bff28-70e9-4b56-ab38-49511f85852d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4f5bff28-70e9-4b56-ab38-49511f85852d" (UID: "4f5bff28-70e9-4b56-ab38-49511f85852d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:50:02 crc kubenswrapper[4821]: I1125 10:50:02.889429 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vvqhz\" (UniqueName: \"kubernetes.io/projected/4f5bff28-70e9-4b56-ab38-49511f85852d-kube-api-access-vvqhz\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:02 crc kubenswrapper[4821]: I1125 10:50:02.889474 4821 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4f5bff28-70e9-4b56-ab38-49511f85852d-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:02 crc kubenswrapper[4821]: I1125 10:50:02.889486 4821 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f5bff28-70e9-4b56-ab38-49511f85852d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:02 crc kubenswrapper[4821]: I1125 10:50:02.889496 4821 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4f5bff28-70e9-4b56-ab38-49511f85852d-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:02 crc kubenswrapper[4821]: I1125 10:50:02.960349 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f5bff28-70e9-4b56-ab38-49511f85852d-config-data" (OuterVolumeSpecName: "config-data") pod "4f5bff28-70e9-4b56-ab38-49511f85852d" (UID: "4f5bff28-70e9-4b56-ab38-49511f85852d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:50:02 crc kubenswrapper[4821]: I1125 10:50:02.990863 4821 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f5bff28-70e9-4b56-ab38-49511f85852d-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:03 crc kubenswrapper[4821]: I1125 10:50:03.044278 4821 generic.go:334] "Generic (PLEG): container finished" podID="4f5bff28-70e9-4b56-ab38-49511f85852d" containerID="a4e878765149277342071df06b9a3d4de0e2483bb18a27e40f85f6f2947fde9b" exitCode=0 Nov 25 10:50:03 crc kubenswrapper[4821]: I1125 10:50:03.044327 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"4f5bff28-70e9-4b56-ab38-49511f85852d","Type":"ContainerDied","Data":"a4e878765149277342071df06b9a3d4de0e2483bb18a27e40f85f6f2947fde9b"} Nov 25 10:50:03 crc kubenswrapper[4821]: I1125 10:50:03.044356 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"4f5bff28-70e9-4b56-ab38-49511f85852d","Type":"ContainerDied","Data":"3da9a016ff98c68c6d2cd152abf9e95deba6ba01d9f4f20698258cfea411b238"} Nov 25 10:50:03 crc kubenswrapper[4821]: I1125 10:50:03.044375 4821 scope.go:117] "RemoveContainer" containerID="42e6a15ed6523ba17eead013e533d86e41662911fe061607b0113ad40f221233" Nov 25 10:50:03 crc kubenswrapper[4821]: I1125 10:50:03.044374 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 25 10:50:03 crc kubenswrapper[4821]: I1125 10:50:03.076776 4821 scope.go:117] "RemoveContainer" containerID="a4e878765149277342071df06b9a3d4de0e2483bb18a27e40f85f6f2947fde9b" Nov 25 10:50:03 crc kubenswrapper[4821]: I1125 10:50:03.077040 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-5fb467464c-hlrl9" Nov 25 10:50:03 crc kubenswrapper[4821]: I1125 10:50:03.103935 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 25 10:50:03 crc kubenswrapper[4821]: I1125 10:50:03.140565 4821 scope.go:117] "RemoveContainer" containerID="42e6a15ed6523ba17eead013e533d86e41662911fe061607b0113ad40f221233" Nov 25 10:50:03 crc kubenswrapper[4821]: E1125 10:50:03.142379 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42e6a15ed6523ba17eead013e533d86e41662911fe061607b0113ad40f221233\": container with ID starting with 42e6a15ed6523ba17eead013e533d86e41662911fe061607b0113ad40f221233 not found: ID does not exist" containerID="42e6a15ed6523ba17eead013e533d86e41662911fe061607b0113ad40f221233" Nov 25 10:50:03 crc kubenswrapper[4821]: I1125 10:50:03.142459 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42e6a15ed6523ba17eead013e533d86e41662911fe061607b0113ad40f221233"} err="failed to get container status \"42e6a15ed6523ba17eead013e533d86e41662911fe061607b0113ad40f221233\": rpc error: code = NotFound desc = could not find container \"42e6a15ed6523ba17eead013e533d86e41662911fe061607b0113ad40f221233\": container with ID starting with 42e6a15ed6523ba17eead013e533d86e41662911fe061607b0113ad40f221233 not found: ID does not exist" Nov 25 10:50:03 crc kubenswrapper[4821]: I1125 10:50:03.142497 4821 scope.go:117] "RemoveContainer" containerID="a4e878765149277342071df06b9a3d4de0e2483bb18a27e40f85f6f2947fde9b" Nov 25 10:50:03 crc kubenswrapper[4821]: I1125 10:50:03.146068 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 25 10:50:03 crc kubenswrapper[4821]: E1125 10:50:03.146324 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4e878765149277342071df06b9a3d4de0e2483bb18a27e40f85f6f2947fde9b\": container with ID starting with a4e878765149277342071df06b9a3d4de0e2483bb18a27e40f85f6f2947fde9b not found: ID does not exist" containerID="a4e878765149277342071df06b9a3d4de0e2483bb18a27e40f85f6f2947fde9b" Nov 25 10:50:03 crc kubenswrapper[4821]: I1125 10:50:03.146382 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4e878765149277342071df06b9a3d4de0e2483bb18a27e40f85f6f2947fde9b"} err="failed to get container status \"a4e878765149277342071df06b9a3d4de0e2483bb18a27e40f85f6f2947fde9b\": rpc error: code = NotFound desc = could not find container \"a4e878765149277342071df06b9a3d4de0e2483bb18a27e40f85f6f2947fde9b\": container with ID starting with a4e878765149277342071df06b9a3d4de0e2483bb18a27e40f85f6f2947fde9b not found: ID does not exist" Nov 25 10:50:03 crc kubenswrapper[4821]: I1125 10:50:03.179434 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 25 10:50:03 crc kubenswrapper[4821]: E1125 10:50:03.180277 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53f9f168-1063-4bb0-8d1c-03482761ae7f" containerName="dnsmasq-dns" Nov 25 10:50:03 crc kubenswrapper[4821]: I1125 10:50:03.180298 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="53f9f168-1063-4bb0-8d1c-03482761ae7f" containerName="dnsmasq-dns" Nov 25 10:50:03 crc kubenswrapper[4821]: E1125 10:50:03.180434 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f5bff28-70e9-4b56-ab38-49511f85852d" containerName="probe" Nov 25 10:50:03 crc kubenswrapper[4821]: I1125 10:50:03.180445 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f5bff28-70e9-4b56-ab38-49511f85852d" containerName="probe" Nov 25 10:50:03 crc kubenswrapper[4821]: E1125 10:50:03.180488 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f5bff28-70e9-4b56-ab38-49511f85852d" containerName="cinder-scheduler" Nov 25 10:50:03 crc kubenswrapper[4821]: I1125 10:50:03.180497 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f5bff28-70e9-4b56-ab38-49511f85852d" containerName="cinder-scheduler" Nov 25 10:50:03 crc kubenswrapper[4821]: E1125 10:50:03.180521 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53f9f168-1063-4bb0-8d1c-03482761ae7f" containerName="init" Nov 25 10:50:03 crc kubenswrapper[4821]: I1125 10:50:03.180529 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="53f9f168-1063-4bb0-8d1c-03482761ae7f" containerName="init" Nov 25 10:50:03 crc kubenswrapper[4821]: I1125 10:50:03.182030 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f5bff28-70e9-4b56-ab38-49511f85852d" containerName="probe" Nov 25 10:50:03 crc kubenswrapper[4821]: I1125 10:50:03.182105 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f5bff28-70e9-4b56-ab38-49511f85852d" containerName="cinder-scheduler" Nov 25 10:50:03 crc kubenswrapper[4821]: I1125 10:50:03.182135 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="53f9f168-1063-4bb0-8d1c-03482761ae7f" containerName="dnsmasq-dns" Nov 25 10:50:03 crc kubenswrapper[4821]: I1125 10:50:03.190514 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 25 10:50:03 crc kubenswrapper[4821]: I1125 10:50:03.192754 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 25 10:50:03 crc kubenswrapper[4821]: I1125 10:50:03.208770 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 25 10:50:03 crc kubenswrapper[4821]: I1125 10:50:03.301064 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7b0c571f-6107-4ecd-8ffb-cd7b992d05a5-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"7b0c571f-6107-4ecd-8ffb-cd7b992d05a5\") " pod="openstack/cinder-scheduler-0" Nov 25 10:50:03 crc kubenswrapper[4821]: I1125 10:50:03.301105 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7b0c571f-6107-4ecd-8ffb-cd7b992d05a5-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"7b0c571f-6107-4ecd-8ffb-cd7b992d05a5\") " pod="openstack/cinder-scheduler-0" Nov 25 10:50:03 crc kubenswrapper[4821]: I1125 10:50:03.301202 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b0c571f-6107-4ecd-8ffb-cd7b992d05a5-scripts\") pod \"cinder-scheduler-0\" (UID: \"7b0c571f-6107-4ecd-8ffb-cd7b992d05a5\") " pod="openstack/cinder-scheduler-0" Nov 25 10:50:03 crc kubenswrapper[4821]: I1125 10:50:03.301232 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hh7fb\" (UniqueName: \"kubernetes.io/projected/7b0c571f-6107-4ecd-8ffb-cd7b992d05a5-kube-api-access-hh7fb\") pod \"cinder-scheduler-0\" (UID: \"7b0c571f-6107-4ecd-8ffb-cd7b992d05a5\") " pod="openstack/cinder-scheduler-0" Nov 25 10:50:03 crc kubenswrapper[4821]: I1125 10:50:03.301271 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b0c571f-6107-4ecd-8ffb-cd7b992d05a5-config-data\") pod \"cinder-scheduler-0\" (UID: \"7b0c571f-6107-4ecd-8ffb-cd7b992d05a5\") " pod="openstack/cinder-scheduler-0" Nov 25 10:50:03 crc kubenswrapper[4821]: I1125 10:50:03.301311 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b0c571f-6107-4ecd-8ffb-cd7b992d05a5-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"7b0c571f-6107-4ecd-8ffb-cd7b992d05a5\") " pod="openstack/cinder-scheduler-0" Nov 25 10:50:03 crc kubenswrapper[4821]: I1125 10:50:03.403437 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b0c571f-6107-4ecd-8ffb-cd7b992d05a5-scripts\") pod \"cinder-scheduler-0\" (UID: \"7b0c571f-6107-4ecd-8ffb-cd7b992d05a5\") " pod="openstack/cinder-scheduler-0" Nov 25 10:50:03 crc kubenswrapper[4821]: I1125 10:50:03.403504 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hh7fb\" (UniqueName: \"kubernetes.io/projected/7b0c571f-6107-4ecd-8ffb-cd7b992d05a5-kube-api-access-hh7fb\") pod \"cinder-scheduler-0\" (UID: \"7b0c571f-6107-4ecd-8ffb-cd7b992d05a5\") " pod="openstack/cinder-scheduler-0" Nov 25 10:50:03 crc kubenswrapper[4821]: I1125 10:50:03.403551 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b0c571f-6107-4ecd-8ffb-cd7b992d05a5-config-data\") pod \"cinder-scheduler-0\" (UID: \"7b0c571f-6107-4ecd-8ffb-cd7b992d05a5\") " pod="openstack/cinder-scheduler-0" Nov 25 10:50:03 crc kubenswrapper[4821]: I1125 10:50:03.403645 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b0c571f-6107-4ecd-8ffb-cd7b992d05a5-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"7b0c571f-6107-4ecd-8ffb-cd7b992d05a5\") " pod="openstack/cinder-scheduler-0" Nov 25 10:50:03 crc kubenswrapper[4821]: I1125 10:50:03.403698 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7b0c571f-6107-4ecd-8ffb-cd7b992d05a5-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"7b0c571f-6107-4ecd-8ffb-cd7b992d05a5\") " pod="openstack/cinder-scheduler-0" Nov 25 10:50:03 crc kubenswrapper[4821]: I1125 10:50:03.403728 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7b0c571f-6107-4ecd-8ffb-cd7b992d05a5-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"7b0c571f-6107-4ecd-8ffb-cd7b992d05a5\") " pod="openstack/cinder-scheduler-0" Nov 25 10:50:03 crc kubenswrapper[4821]: I1125 10:50:03.404698 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7b0c571f-6107-4ecd-8ffb-cd7b992d05a5-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"7b0c571f-6107-4ecd-8ffb-cd7b992d05a5\") " pod="openstack/cinder-scheduler-0" Nov 25 10:50:03 crc kubenswrapper[4821]: I1125 10:50:03.410563 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b0c571f-6107-4ecd-8ffb-cd7b992d05a5-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"7b0c571f-6107-4ecd-8ffb-cd7b992d05a5\") " pod="openstack/cinder-scheduler-0" Nov 25 10:50:03 crc kubenswrapper[4821]: I1125 10:50:03.410998 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b0c571f-6107-4ecd-8ffb-cd7b992d05a5-scripts\") pod \"cinder-scheduler-0\" (UID: \"7b0c571f-6107-4ecd-8ffb-cd7b992d05a5\") " pod="openstack/cinder-scheduler-0" Nov 25 10:50:03 crc kubenswrapper[4821]: I1125 10:50:03.411738 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7b0c571f-6107-4ecd-8ffb-cd7b992d05a5-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"7b0c571f-6107-4ecd-8ffb-cd7b992d05a5\") " pod="openstack/cinder-scheduler-0" Nov 25 10:50:03 crc kubenswrapper[4821]: I1125 10:50:03.412472 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b0c571f-6107-4ecd-8ffb-cd7b992d05a5-config-data\") pod \"cinder-scheduler-0\" (UID: \"7b0c571f-6107-4ecd-8ffb-cd7b992d05a5\") " pod="openstack/cinder-scheduler-0" Nov 25 10:50:03 crc kubenswrapper[4821]: I1125 10:50:03.439285 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hh7fb\" (UniqueName: \"kubernetes.io/projected/7b0c571f-6107-4ecd-8ffb-cd7b992d05a5-kube-api-access-hh7fb\") pod \"cinder-scheduler-0\" (UID: \"7b0c571f-6107-4ecd-8ffb-cd7b992d05a5\") " pod="openstack/cinder-scheduler-0" Nov 25 10:50:03 crc kubenswrapper[4821]: I1125 10:50:03.578034 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 25 10:50:04 crc kubenswrapper[4821]: I1125 10:50:04.126253 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4f5bff28-70e9-4b56-ab38-49511f85852d" path="/var/lib/kubelet/pods/4f5bff28-70e9-4b56-ab38-49511f85852d/volumes" Nov 25 10:50:04 crc kubenswrapper[4821]: I1125 10:50:04.155262 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 25 10:50:04 crc kubenswrapper[4821]: W1125 10:50:04.158682 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7b0c571f_6107_4ecd_8ffb_cd7b992d05a5.slice/crio-4f36788aee9ac31cd252c14dbbb25b25366bdc76516fd23c27f2621d091288b9 WatchSource:0}: Error finding container 4f36788aee9ac31cd252c14dbbb25b25366bdc76516fd23c27f2621d091288b9: Status 404 returned error can't find the container with id 4f36788aee9ac31cd252c14dbbb25b25366bdc76516fd23c27f2621d091288b9 Nov 25 10:50:05 crc kubenswrapper[4821]: I1125 10:50:05.070794 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"7b0c571f-6107-4ecd-8ffb-cd7b992d05a5","Type":"ContainerStarted","Data":"accbdbc47b5f9ddec6e794d85cf915208806783a7f7621c4b3b93eb4be308d97"} Nov 25 10:50:05 crc kubenswrapper[4821]: I1125 10:50:05.071139 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"7b0c571f-6107-4ecd-8ffb-cd7b992d05a5","Type":"ContainerStarted","Data":"4f36788aee9ac31cd252c14dbbb25b25366bdc76516fd23c27f2621d091288b9"} Nov 25 10:50:05 crc kubenswrapper[4821]: I1125 10:50:05.981738 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Nov 25 10:50:05 crc kubenswrapper[4821]: I1125 10:50:05.983704 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 25 10:50:05 crc kubenswrapper[4821]: I1125 10:50:05.987891 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-5kqm5" Nov 25 10:50:05 crc kubenswrapper[4821]: I1125 10:50:05.988432 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Nov 25 10:50:05 crc kubenswrapper[4821]: I1125 10:50:05.988969 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Nov 25 10:50:06 crc kubenswrapper[4821]: I1125 10:50:06.008039 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 25 10:50:06 crc kubenswrapper[4821]: I1125 10:50:06.054382 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d969bbb6-564d-4563-98c6-5e2ec965c2dc-openstack-config-secret\") pod \"openstackclient\" (UID: \"d969bbb6-564d-4563-98c6-5e2ec965c2dc\") " pod="openstack/openstackclient" Nov 25 10:50:06 crc kubenswrapper[4821]: I1125 10:50:06.054734 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d969bbb6-564d-4563-98c6-5e2ec965c2dc-combined-ca-bundle\") pod \"openstackclient\" (UID: \"d969bbb6-564d-4563-98c6-5e2ec965c2dc\") " pod="openstack/openstackclient" Nov 25 10:50:06 crc kubenswrapper[4821]: I1125 10:50:06.054931 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wddgw\" (UniqueName: \"kubernetes.io/projected/d969bbb6-564d-4563-98c6-5e2ec965c2dc-kube-api-access-wddgw\") pod \"openstackclient\" (UID: \"d969bbb6-564d-4563-98c6-5e2ec965c2dc\") " pod="openstack/openstackclient" Nov 25 10:50:06 crc kubenswrapper[4821]: I1125 10:50:06.055057 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d969bbb6-564d-4563-98c6-5e2ec965c2dc-openstack-config\") pod \"openstackclient\" (UID: \"d969bbb6-564d-4563-98c6-5e2ec965c2dc\") " pod="openstack/openstackclient" Nov 25 10:50:06 crc kubenswrapper[4821]: I1125 10:50:06.082722 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"7b0c571f-6107-4ecd-8ffb-cd7b992d05a5","Type":"ContainerStarted","Data":"7ae265877bd28f71ad4be389717e86b8e1890d024100bd142133bffc14edb113"} Nov 25 10:50:06 crc kubenswrapper[4821]: I1125 10:50:06.102407 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.102389255 podStartE2EDuration="3.102389255s" podCreationTimestamp="2025-11-25 10:50:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:50:06.099365828 +0000 UTC m=+1076.635685675" watchObservedRunningTime="2025-11-25 10:50:06.102389255 +0000 UTC m=+1076.638709102" Nov 25 10:50:06 crc kubenswrapper[4821]: I1125 10:50:06.156328 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d969bbb6-564d-4563-98c6-5e2ec965c2dc-openstack-config-secret\") pod \"openstackclient\" (UID: \"d969bbb6-564d-4563-98c6-5e2ec965c2dc\") " pod="openstack/openstackclient" Nov 25 10:50:06 crc kubenswrapper[4821]: I1125 10:50:06.156609 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d969bbb6-564d-4563-98c6-5e2ec965c2dc-combined-ca-bundle\") pod \"openstackclient\" (UID: \"d969bbb6-564d-4563-98c6-5e2ec965c2dc\") " pod="openstack/openstackclient" Nov 25 10:50:06 crc kubenswrapper[4821]: I1125 10:50:06.156756 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wddgw\" (UniqueName: \"kubernetes.io/projected/d969bbb6-564d-4563-98c6-5e2ec965c2dc-kube-api-access-wddgw\") pod \"openstackclient\" (UID: \"d969bbb6-564d-4563-98c6-5e2ec965c2dc\") " pod="openstack/openstackclient" Nov 25 10:50:06 crc kubenswrapper[4821]: I1125 10:50:06.156842 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d969bbb6-564d-4563-98c6-5e2ec965c2dc-openstack-config\") pod \"openstackclient\" (UID: \"d969bbb6-564d-4563-98c6-5e2ec965c2dc\") " pod="openstack/openstackclient" Nov 25 10:50:06 crc kubenswrapper[4821]: I1125 10:50:06.158741 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d969bbb6-564d-4563-98c6-5e2ec965c2dc-openstack-config\") pod \"openstackclient\" (UID: \"d969bbb6-564d-4563-98c6-5e2ec965c2dc\") " pod="openstack/openstackclient" Nov 25 10:50:06 crc kubenswrapper[4821]: I1125 10:50:06.175778 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d969bbb6-564d-4563-98c6-5e2ec965c2dc-openstack-config-secret\") pod \"openstackclient\" (UID: \"d969bbb6-564d-4563-98c6-5e2ec965c2dc\") " pod="openstack/openstackclient" Nov 25 10:50:06 crc kubenswrapper[4821]: I1125 10:50:06.176484 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d969bbb6-564d-4563-98c6-5e2ec965c2dc-combined-ca-bundle\") pod \"openstackclient\" (UID: \"d969bbb6-564d-4563-98c6-5e2ec965c2dc\") " pod="openstack/openstackclient" Nov 25 10:50:06 crc kubenswrapper[4821]: I1125 10:50:06.179625 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wddgw\" (UniqueName: \"kubernetes.io/projected/d969bbb6-564d-4563-98c6-5e2ec965c2dc-kube-api-access-wddgw\") pod \"openstackclient\" (UID: \"d969bbb6-564d-4563-98c6-5e2ec965c2dc\") " pod="openstack/openstackclient" Nov 25 10:50:06 crc kubenswrapper[4821]: I1125 10:50:06.305778 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 25 10:50:07 crc kubenswrapper[4821]: I1125 10:50:06.897134 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-67b899d6f8-pscf9" Nov 25 10:50:07 crc kubenswrapper[4821]: I1125 10:50:06.922624 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-67b899d6f8-pscf9" Nov 25 10:50:07 crc kubenswrapper[4821]: I1125 10:50:06.997405 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-55c965854d-xpf4q"] Nov 25 10:50:07 crc kubenswrapper[4821]: I1125 10:50:06.997627 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-55c965854d-xpf4q" podUID="73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7" containerName="barbican-api-log" containerID="cri-o://8bee610994bbfbd0e4f395db69e2e44e857f94bfbafa8c30ddba075694222c63" gracePeriod=30 Nov 25 10:50:07 crc kubenswrapper[4821]: I1125 10:50:06.998002 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-55c965854d-xpf4q" podUID="73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7" containerName="barbican-api" containerID="cri-o://1b4b4e05d8669cea149cb416589c262aeee2db1f26dafb11f5ec8795a3a85047" gracePeriod=30 Nov 25 10:50:07 crc kubenswrapper[4821]: I1125 10:50:07.058568 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-84cbbf9b54-k7shj" Nov 25 10:50:07 crc kubenswrapper[4821]: I1125 10:50:07.666830 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 25 10:50:08 crc kubenswrapper[4821]: I1125 10:50:08.108750 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"d969bbb6-564d-4563-98c6-5e2ec965c2dc","Type":"ContainerStarted","Data":"38f84dd5b6c250c79d1be7f1edecda571e6a5b9f3d916c19169c861f778e8566"} Nov 25 10:50:08 crc kubenswrapper[4821]: I1125 10:50:08.110937 4821 generic.go:334] "Generic (PLEG): container finished" podID="73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7" containerID="8bee610994bbfbd0e4f395db69e2e44e857f94bfbafa8c30ddba075694222c63" exitCode=143 Nov 25 10:50:08 crc kubenswrapper[4821]: I1125 10:50:08.110971 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-55c965854d-xpf4q" event={"ID":"73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7","Type":"ContainerDied","Data":"8bee610994bbfbd0e4f395db69e2e44e857f94bfbafa8c30ddba075694222c63"} Nov 25 10:50:08 crc kubenswrapper[4821]: I1125 10:50:08.329330 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Nov 25 10:50:08 crc kubenswrapper[4821]: I1125 10:50:08.579044 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 25 10:50:09 crc kubenswrapper[4821]: I1125 10:50:09.810728 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-f8667547d-9v5xh" Nov 25 10:50:09 crc kubenswrapper[4821]: I1125 10:50:09.874063 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-f8667547d-9v5xh" Nov 25 10:50:10 crc kubenswrapper[4821]: I1125 10:50:10.259826 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-55575b4d57-6t4kz" Nov 25 10:50:10 crc kubenswrapper[4821]: I1125 10:50:10.284616 4821 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-55c965854d-xpf4q" podUID="73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.162:9311/healthcheck\": read tcp 10.217.0.2:47972->10.217.0.162:9311: read: connection reset by peer" Nov 25 10:50:10 crc kubenswrapper[4821]: I1125 10:50:10.285879 4821 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-55c965854d-xpf4q" podUID="73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.162:9311/healthcheck\": read tcp 10.217.0.2:47970->10.217.0.162:9311: read: connection reset by peer" Nov 25 10:50:10 crc kubenswrapper[4821]: I1125 10:50:10.334630 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-84cbbf9b54-k7shj"] Nov 25 10:50:10 crc kubenswrapper[4821]: I1125 10:50:10.334888 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-84cbbf9b54-k7shj" podUID="690339d6-9d90-46b3-b471-afabc2da9e7c" containerName="neutron-api" containerID="cri-o://520722ded67872ccc942a42008a87b2908384fdc9b316a93acadb80181fde2e9" gracePeriod=30 Nov 25 10:50:10 crc kubenswrapper[4821]: I1125 10:50:10.335027 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-84cbbf9b54-k7shj" podUID="690339d6-9d90-46b3-b471-afabc2da9e7c" containerName="neutron-httpd" containerID="cri-o://9fc413e8b8971197f05d3f9d1675fd56f8c25a1a5c97327431248c71d7982d6f" gracePeriod=30 Nov 25 10:50:10 crc kubenswrapper[4821]: I1125 10:50:10.830742 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-55c965854d-xpf4q" Nov 25 10:50:10 crc kubenswrapper[4821]: I1125 10:50:10.850805 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7-config-data\") pod \"73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7\" (UID: \"73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7\") " Nov 25 10:50:10 crc kubenswrapper[4821]: I1125 10:50:10.850886 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7-combined-ca-bundle\") pod \"73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7\" (UID: \"73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7\") " Nov 25 10:50:10 crc kubenswrapper[4821]: I1125 10:50:10.850917 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m64k7\" (UniqueName: \"kubernetes.io/projected/73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7-kube-api-access-m64k7\") pod \"73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7\" (UID: \"73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7\") " Nov 25 10:50:10 crc kubenswrapper[4821]: I1125 10:50:10.851057 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7-logs\") pod \"73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7\" (UID: \"73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7\") " Nov 25 10:50:10 crc kubenswrapper[4821]: I1125 10:50:10.851112 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7-config-data-custom\") pod \"73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7\" (UID: \"73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7\") " Nov 25 10:50:10 crc kubenswrapper[4821]: I1125 10:50:10.863566 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7-logs" (OuterVolumeSpecName: "logs") pod "73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7" (UID: "73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:50:10 crc kubenswrapper[4821]: I1125 10:50:10.868423 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7" (UID: "73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:50:10 crc kubenswrapper[4821]: I1125 10:50:10.878370 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7-kube-api-access-m64k7" (OuterVolumeSpecName: "kube-api-access-m64k7") pod "73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7" (UID: "73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7"). InnerVolumeSpecName "kube-api-access-m64k7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:50:10 crc kubenswrapper[4821]: I1125 10:50:10.953318 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m64k7\" (UniqueName: \"kubernetes.io/projected/73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7-kube-api-access-m64k7\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:10 crc kubenswrapper[4821]: I1125 10:50:10.953490 4821 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7-logs\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:10 crc kubenswrapper[4821]: I1125 10:50:10.953542 4821 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:10 crc kubenswrapper[4821]: I1125 10:50:10.962377 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7" (UID: "73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:50:10 crc kubenswrapper[4821]: I1125 10:50:10.993054 4821 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-869c6f5bfb-l785w" podUID="e2327ad3-5d88-4f4b-91ea-91deaf7fc249" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Nov 25 10:50:11 crc kubenswrapper[4821]: I1125 10:50:11.013312 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7-config-data" (OuterVolumeSpecName: "config-data") pod "73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7" (UID: "73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:50:11 crc kubenswrapper[4821]: I1125 10:50:11.056287 4821 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:11 crc kubenswrapper[4821]: I1125 10:50:11.056325 4821 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:11 crc kubenswrapper[4821]: I1125 10:50:11.166840 4821 generic.go:334] "Generic (PLEG): container finished" podID="690339d6-9d90-46b3-b471-afabc2da9e7c" containerID="9fc413e8b8971197f05d3f9d1675fd56f8c25a1a5c97327431248c71d7982d6f" exitCode=0 Nov 25 10:50:11 crc kubenswrapper[4821]: I1125 10:50:11.166909 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-84cbbf9b54-k7shj" event={"ID":"690339d6-9d90-46b3-b471-afabc2da9e7c","Type":"ContainerDied","Data":"9fc413e8b8971197f05d3f9d1675fd56f8c25a1a5c97327431248c71d7982d6f"} Nov 25 10:50:11 crc kubenswrapper[4821]: I1125 10:50:11.169058 4821 generic.go:334] "Generic (PLEG): container finished" podID="73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7" containerID="1b4b4e05d8669cea149cb416589c262aeee2db1f26dafb11f5ec8795a3a85047" exitCode=0 Nov 25 10:50:11 crc kubenswrapper[4821]: I1125 10:50:11.169098 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-55c965854d-xpf4q" Nov 25 10:50:11 crc kubenswrapper[4821]: I1125 10:50:11.169103 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-55c965854d-xpf4q" event={"ID":"73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7","Type":"ContainerDied","Data":"1b4b4e05d8669cea149cb416589c262aeee2db1f26dafb11f5ec8795a3a85047"} Nov 25 10:50:11 crc kubenswrapper[4821]: I1125 10:50:11.169142 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-55c965854d-xpf4q" event={"ID":"73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7","Type":"ContainerDied","Data":"7563ef010033402324871f9d93411bb4645bbf2d61d1a16b84199972670d5566"} Nov 25 10:50:11 crc kubenswrapper[4821]: I1125 10:50:11.169179 4821 scope.go:117] "RemoveContainer" containerID="1b4b4e05d8669cea149cb416589c262aeee2db1f26dafb11f5ec8795a3a85047" Nov 25 10:50:11 crc kubenswrapper[4821]: I1125 10:50:11.200340 4821 scope.go:117] "RemoveContainer" containerID="8bee610994bbfbd0e4f395db69e2e44e857f94bfbafa8c30ddba075694222c63" Nov 25 10:50:11 crc kubenswrapper[4821]: I1125 10:50:11.222932 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-55c965854d-xpf4q"] Nov 25 10:50:11 crc kubenswrapper[4821]: I1125 10:50:11.232430 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-55c965854d-xpf4q"] Nov 25 10:50:11 crc kubenswrapper[4821]: I1125 10:50:11.237330 4821 scope.go:117] "RemoveContainer" containerID="1b4b4e05d8669cea149cb416589c262aeee2db1f26dafb11f5ec8795a3a85047" Nov 25 10:50:11 crc kubenswrapper[4821]: E1125 10:50:11.241329 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b4b4e05d8669cea149cb416589c262aeee2db1f26dafb11f5ec8795a3a85047\": container with ID starting with 1b4b4e05d8669cea149cb416589c262aeee2db1f26dafb11f5ec8795a3a85047 not found: ID does not exist" containerID="1b4b4e05d8669cea149cb416589c262aeee2db1f26dafb11f5ec8795a3a85047" Nov 25 10:50:11 crc kubenswrapper[4821]: I1125 10:50:11.241393 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b4b4e05d8669cea149cb416589c262aeee2db1f26dafb11f5ec8795a3a85047"} err="failed to get container status \"1b4b4e05d8669cea149cb416589c262aeee2db1f26dafb11f5ec8795a3a85047\": rpc error: code = NotFound desc = could not find container \"1b4b4e05d8669cea149cb416589c262aeee2db1f26dafb11f5ec8795a3a85047\": container with ID starting with 1b4b4e05d8669cea149cb416589c262aeee2db1f26dafb11f5ec8795a3a85047 not found: ID does not exist" Nov 25 10:50:11 crc kubenswrapper[4821]: I1125 10:50:11.241425 4821 scope.go:117] "RemoveContainer" containerID="8bee610994bbfbd0e4f395db69e2e44e857f94bfbafa8c30ddba075694222c63" Nov 25 10:50:11 crc kubenswrapper[4821]: E1125 10:50:11.243306 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8bee610994bbfbd0e4f395db69e2e44e857f94bfbafa8c30ddba075694222c63\": container with ID starting with 8bee610994bbfbd0e4f395db69e2e44e857f94bfbafa8c30ddba075694222c63 not found: ID does not exist" containerID="8bee610994bbfbd0e4f395db69e2e44e857f94bfbafa8c30ddba075694222c63" Nov 25 10:50:11 crc kubenswrapper[4821]: I1125 10:50:11.243339 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8bee610994bbfbd0e4f395db69e2e44e857f94bfbafa8c30ddba075694222c63"} err="failed to get container status \"8bee610994bbfbd0e4f395db69e2e44e857f94bfbafa8c30ddba075694222c63\": rpc error: code = NotFound desc = could not find container \"8bee610994bbfbd0e4f395db69e2e44e857f94bfbafa8c30ddba075694222c63\": container with ID starting with 8bee610994bbfbd0e4f395db69e2e44e857f94bfbafa8c30ddba075694222c63 not found: ID does not exist" Nov 25 10:50:11 crc kubenswrapper[4821]: I1125 10:50:11.505771 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-ff8c8c76c-ctf8b"] Nov 25 10:50:11 crc kubenswrapper[4821]: E1125 10:50:11.506133 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7" containerName="barbican-api-log" Nov 25 10:50:11 crc kubenswrapper[4821]: I1125 10:50:11.506148 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7" containerName="barbican-api-log" Nov 25 10:50:11 crc kubenswrapper[4821]: E1125 10:50:11.507676 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7" containerName="barbican-api" Nov 25 10:50:11 crc kubenswrapper[4821]: I1125 10:50:11.507692 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7" containerName="barbican-api" Nov 25 10:50:11 crc kubenswrapper[4821]: I1125 10:50:11.507858 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7" containerName="barbican-api-log" Nov 25 10:50:11 crc kubenswrapper[4821]: I1125 10:50:11.507883 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7" containerName="barbican-api" Nov 25 10:50:11 crc kubenswrapper[4821]: I1125 10:50:11.508768 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-ff8c8c76c-ctf8b" Nov 25 10:50:11 crc kubenswrapper[4821]: I1125 10:50:11.511335 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Nov 25 10:50:11 crc kubenswrapper[4821]: I1125 10:50:11.511615 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Nov 25 10:50:11 crc kubenswrapper[4821]: I1125 10:50:11.513011 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Nov 25 10:50:11 crc kubenswrapper[4821]: I1125 10:50:11.526337 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-ff8c8c76c-ctf8b"] Nov 25 10:50:11 crc kubenswrapper[4821]: I1125 10:50:11.564649 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/db085977-6941-483f-a800-411ff71612af-etc-swift\") pod \"swift-proxy-ff8c8c76c-ctf8b\" (UID: \"db085977-6941-483f-a800-411ff71612af\") " pod="openstack/swift-proxy-ff8c8c76c-ctf8b" Nov 25 10:50:11 crc kubenswrapper[4821]: I1125 10:50:11.565002 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/db085977-6941-483f-a800-411ff71612af-internal-tls-certs\") pod \"swift-proxy-ff8c8c76c-ctf8b\" (UID: \"db085977-6941-483f-a800-411ff71612af\") " pod="openstack/swift-proxy-ff8c8c76c-ctf8b" Nov 25 10:50:11 crc kubenswrapper[4821]: I1125 10:50:11.565050 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/db085977-6941-483f-a800-411ff71612af-public-tls-certs\") pod \"swift-proxy-ff8c8c76c-ctf8b\" (UID: \"db085977-6941-483f-a800-411ff71612af\") " pod="openstack/swift-proxy-ff8c8c76c-ctf8b" Nov 25 10:50:11 crc kubenswrapper[4821]: I1125 10:50:11.565093 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db085977-6941-483f-a800-411ff71612af-combined-ca-bundle\") pod \"swift-proxy-ff8c8c76c-ctf8b\" (UID: \"db085977-6941-483f-a800-411ff71612af\") " pod="openstack/swift-proxy-ff8c8c76c-ctf8b" Nov 25 10:50:11 crc kubenswrapper[4821]: I1125 10:50:11.565114 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db085977-6941-483f-a800-411ff71612af-log-httpd\") pod \"swift-proxy-ff8c8c76c-ctf8b\" (UID: \"db085977-6941-483f-a800-411ff71612af\") " pod="openstack/swift-proxy-ff8c8c76c-ctf8b" Nov 25 10:50:11 crc kubenswrapper[4821]: I1125 10:50:11.565137 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db085977-6941-483f-a800-411ff71612af-config-data\") pod \"swift-proxy-ff8c8c76c-ctf8b\" (UID: \"db085977-6941-483f-a800-411ff71612af\") " pod="openstack/swift-proxy-ff8c8c76c-ctf8b" Nov 25 10:50:11 crc kubenswrapper[4821]: I1125 10:50:11.565226 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db085977-6941-483f-a800-411ff71612af-run-httpd\") pod \"swift-proxy-ff8c8c76c-ctf8b\" (UID: \"db085977-6941-483f-a800-411ff71612af\") " pod="openstack/swift-proxy-ff8c8c76c-ctf8b" Nov 25 10:50:11 crc kubenswrapper[4821]: I1125 10:50:11.565289 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hq2kn\" (UniqueName: \"kubernetes.io/projected/db085977-6941-483f-a800-411ff71612af-kube-api-access-hq2kn\") pod \"swift-proxy-ff8c8c76c-ctf8b\" (UID: \"db085977-6941-483f-a800-411ff71612af\") " pod="openstack/swift-proxy-ff8c8c76c-ctf8b" Nov 25 10:50:11 crc kubenswrapper[4821]: I1125 10:50:11.667743 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/db085977-6941-483f-a800-411ff71612af-etc-swift\") pod \"swift-proxy-ff8c8c76c-ctf8b\" (UID: \"db085977-6941-483f-a800-411ff71612af\") " pod="openstack/swift-proxy-ff8c8c76c-ctf8b" Nov 25 10:50:11 crc kubenswrapper[4821]: I1125 10:50:11.667798 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/db085977-6941-483f-a800-411ff71612af-internal-tls-certs\") pod \"swift-proxy-ff8c8c76c-ctf8b\" (UID: \"db085977-6941-483f-a800-411ff71612af\") " pod="openstack/swift-proxy-ff8c8c76c-ctf8b" Nov 25 10:50:11 crc kubenswrapper[4821]: I1125 10:50:11.667839 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/db085977-6941-483f-a800-411ff71612af-public-tls-certs\") pod \"swift-proxy-ff8c8c76c-ctf8b\" (UID: \"db085977-6941-483f-a800-411ff71612af\") " pod="openstack/swift-proxy-ff8c8c76c-ctf8b" Nov 25 10:50:11 crc kubenswrapper[4821]: I1125 10:50:11.667870 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db085977-6941-483f-a800-411ff71612af-combined-ca-bundle\") pod \"swift-proxy-ff8c8c76c-ctf8b\" (UID: \"db085977-6941-483f-a800-411ff71612af\") " pod="openstack/swift-proxy-ff8c8c76c-ctf8b" Nov 25 10:50:11 crc kubenswrapper[4821]: I1125 10:50:11.667884 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db085977-6941-483f-a800-411ff71612af-log-httpd\") pod \"swift-proxy-ff8c8c76c-ctf8b\" (UID: \"db085977-6941-483f-a800-411ff71612af\") " pod="openstack/swift-proxy-ff8c8c76c-ctf8b" Nov 25 10:50:11 crc kubenswrapper[4821]: I1125 10:50:11.667905 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db085977-6941-483f-a800-411ff71612af-config-data\") pod \"swift-proxy-ff8c8c76c-ctf8b\" (UID: \"db085977-6941-483f-a800-411ff71612af\") " pod="openstack/swift-proxy-ff8c8c76c-ctf8b" Nov 25 10:50:11 crc kubenswrapper[4821]: I1125 10:50:11.668416 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db085977-6941-483f-a800-411ff71612af-log-httpd\") pod \"swift-proxy-ff8c8c76c-ctf8b\" (UID: \"db085977-6941-483f-a800-411ff71612af\") " pod="openstack/swift-proxy-ff8c8c76c-ctf8b" Nov 25 10:50:11 crc kubenswrapper[4821]: I1125 10:50:11.668614 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db085977-6941-483f-a800-411ff71612af-run-httpd\") pod \"swift-proxy-ff8c8c76c-ctf8b\" (UID: \"db085977-6941-483f-a800-411ff71612af\") " pod="openstack/swift-proxy-ff8c8c76c-ctf8b" Nov 25 10:50:11 crc kubenswrapper[4821]: I1125 10:50:11.668869 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hq2kn\" (UniqueName: \"kubernetes.io/projected/db085977-6941-483f-a800-411ff71612af-kube-api-access-hq2kn\") pod \"swift-proxy-ff8c8c76c-ctf8b\" (UID: \"db085977-6941-483f-a800-411ff71612af\") " pod="openstack/swift-proxy-ff8c8c76c-ctf8b" Nov 25 10:50:11 crc kubenswrapper[4821]: I1125 10:50:11.669036 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db085977-6941-483f-a800-411ff71612af-run-httpd\") pod \"swift-proxy-ff8c8c76c-ctf8b\" (UID: \"db085977-6941-483f-a800-411ff71612af\") " pod="openstack/swift-proxy-ff8c8c76c-ctf8b" Nov 25 10:50:11 crc kubenswrapper[4821]: I1125 10:50:11.673741 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/db085977-6941-483f-a800-411ff71612af-etc-swift\") pod \"swift-proxy-ff8c8c76c-ctf8b\" (UID: \"db085977-6941-483f-a800-411ff71612af\") " pod="openstack/swift-proxy-ff8c8c76c-ctf8b" Nov 25 10:50:11 crc kubenswrapper[4821]: I1125 10:50:11.674108 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/db085977-6941-483f-a800-411ff71612af-public-tls-certs\") pod \"swift-proxy-ff8c8c76c-ctf8b\" (UID: \"db085977-6941-483f-a800-411ff71612af\") " pod="openstack/swift-proxy-ff8c8c76c-ctf8b" Nov 25 10:50:11 crc kubenswrapper[4821]: I1125 10:50:11.674653 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/db085977-6941-483f-a800-411ff71612af-internal-tls-certs\") pod \"swift-proxy-ff8c8c76c-ctf8b\" (UID: \"db085977-6941-483f-a800-411ff71612af\") " pod="openstack/swift-proxy-ff8c8c76c-ctf8b" Nov 25 10:50:11 crc kubenswrapper[4821]: I1125 10:50:11.679561 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db085977-6941-483f-a800-411ff71612af-combined-ca-bundle\") pod \"swift-proxy-ff8c8c76c-ctf8b\" (UID: \"db085977-6941-483f-a800-411ff71612af\") " pod="openstack/swift-proxy-ff8c8c76c-ctf8b" Nov 25 10:50:11 crc kubenswrapper[4821]: I1125 10:50:11.690013 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hq2kn\" (UniqueName: \"kubernetes.io/projected/db085977-6941-483f-a800-411ff71612af-kube-api-access-hq2kn\") pod \"swift-proxy-ff8c8c76c-ctf8b\" (UID: \"db085977-6941-483f-a800-411ff71612af\") " pod="openstack/swift-proxy-ff8c8c76c-ctf8b" Nov 25 10:50:11 crc kubenswrapper[4821]: I1125 10:50:11.690154 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db085977-6941-483f-a800-411ff71612af-config-data\") pod \"swift-proxy-ff8c8c76c-ctf8b\" (UID: \"db085977-6941-483f-a800-411ff71612af\") " pod="openstack/swift-proxy-ff8c8c76c-ctf8b" Nov 25 10:50:11 crc kubenswrapper[4821]: I1125 10:50:11.829605 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-ff8c8c76c-ctf8b" Nov 25 10:50:12 crc kubenswrapper[4821]: I1125 10:50:12.128284 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7" path="/var/lib/kubelet/pods/73f82f36-5dd0-4abd-b2e9-11d41d4b4ef7/volumes" Nov 25 10:50:12 crc kubenswrapper[4821]: I1125 10:50:12.454778 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-ff8c8c76c-ctf8b"] Nov 25 10:50:12 crc kubenswrapper[4821]: W1125 10:50:12.470434 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddb085977_6941_483f_a800_411ff71612af.slice/crio-793b738c6bb4a35009c8b76b531888035506e4dc6f050e21ef543b1d32b82ae6 WatchSource:0}: Error finding container 793b738c6bb4a35009c8b76b531888035506e4dc6f050e21ef543b1d32b82ae6: Status 404 returned error can't find the container with id 793b738c6bb4a35009c8b76b531888035506e4dc6f050e21ef543b1d32b82ae6 Nov 25 10:50:13 crc kubenswrapper[4821]: I1125 10:50:13.200884 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-ff8c8c76c-ctf8b" event={"ID":"db085977-6941-483f-a800-411ff71612af","Type":"ContainerStarted","Data":"68fe202499d3aeab33c3540fa38cf21211909ee2eeb9f87f3a24248897fb6ef8"} Nov 25 10:50:13 crc kubenswrapper[4821]: I1125 10:50:13.201268 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-ff8c8c76c-ctf8b" event={"ID":"db085977-6941-483f-a800-411ff71612af","Type":"ContainerStarted","Data":"9289b664e1f425141140ec58c7b4a1b9a607968cd0d7555f6f16f34f39b88d63"} Nov 25 10:50:13 crc kubenswrapper[4821]: I1125 10:50:13.201289 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-ff8c8c76c-ctf8b" event={"ID":"db085977-6941-483f-a800-411ff71612af","Type":"ContainerStarted","Data":"793b738c6bb4a35009c8b76b531888035506e4dc6f050e21ef543b1d32b82ae6"} Nov 25 10:50:13 crc kubenswrapper[4821]: I1125 10:50:13.201340 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-ff8c8c76c-ctf8b" Nov 25 10:50:13 crc kubenswrapper[4821]: I1125 10:50:13.201374 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-ff8c8c76c-ctf8b" Nov 25 10:50:13 crc kubenswrapper[4821]: I1125 10:50:13.232885 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-ff8c8c76c-ctf8b" podStartSLOduration=2.232863491 podStartE2EDuration="2.232863491s" podCreationTimestamp="2025-11-25 10:50:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:50:13.232658126 +0000 UTC m=+1083.768977973" watchObservedRunningTime="2025-11-25 10:50:13.232863491 +0000 UTC m=+1083.769183338" Nov 25 10:50:13 crc kubenswrapper[4821]: I1125 10:50:13.810938 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 25 10:50:13 crc kubenswrapper[4821]: I1125 10:50:13.954359 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:50:13 crc kubenswrapper[4821]: I1125 10:50:13.954629 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1647cf99-f6d6-421d-a00b-86538c7a7575" containerName="ceilometer-central-agent" containerID="cri-o://e7403b14fd72da5014034f95716d4e786968445593065d9e47e2dba3a1b41e4b" gracePeriod=30 Nov 25 10:50:13 crc kubenswrapper[4821]: I1125 10:50:13.955388 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1647cf99-f6d6-421d-a00b-86538c7a7575" containerName="proxy-httpd" containerID="cri-o://b57c43f68f8bdd2464535935e220533ae71fe723e2b8bf249a896b43a1b9bc9c" gracePeriod=30 Nov 25 10:50:13 crc kubenswrapper[4821]: I1125 10:50:13.955457 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1647cf99-f6d6-421d-a00b-86538c7a7575" containerName="sg-core" containerID="cri-o://7b894abf033fdc87c1547963b5822943b6b998050779618be7999ac3afa4aa3b" gracePeriod=30 Nov 25 10:50:13 crc kubenswrapper[4821]: I1125 10:50:13.955494 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1647cf99-f6d6-421d-a00b-86538c7a7575" containerName="ceilometer-notification-agent" containerID="cri-o://0073627f3d04f03ca74fbb9e3c020f654c964b11997f56a69f24e0f4144dd25f" gracePeriod=30 Nov 25 10:50:13 crc kubenswrapper[4821]: I1125 10:50:13.966665 4821 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="1647cf99-f6d6-421d-a00b-86538c7a7575" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.165:3000/\": EOF" Nov 25 10:50:14 crc kubenswrapper[4821]: I1125 10:50:14.231686 4821 generic.go:334] "Generic (PLEG): container finished" podID="1647cf99-f6d6-421d-a00b-86538c7a7575" containerID="b57c43f68f8bdd2464535935e220533ae71fe723e2b8bf249a896b43a1b9bc9c" exitCode=0 Nov 25 10:50:14 crc kubenswrapper[4821]: I1125 10:50:14.231724 4821 generic.go:334] "Generic (PLEG): container finished" podID="1647cf99-f6d6-421d-a00b-86538c7a7575" containerID="7b894abf033fdc87c1547963b5822943b6b998050779618be7999ac3afa4aa3b" exitCode=2 Nov 25 10:50:14 crc kubenswrapper[4821]: I1125 10:50:14.231764 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1647cf99-f6d6-421d-a00b-86538c7a7575","Type":"ContainerDied","Data":"b57c43f68f8bdd2464535935e220533ae71fe723e2b8bf249a896b43a1b9bc9c"} Nov 25 10:50:14 crc kubenswrapper[4821]: I1125 10:50:14.231823 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1647cf99-f6d6-421d-a00b-86538c7a7575","Type":"ContainerDied","Data":"7b894abf033fdc87c1547963b5822943b6b998050779618be7999ac3afa4aa3b"} Nov 25 10:50:15 crc kubenswrapper[4821]: I1125 10:50:15.242813 4821 generic.go:334] "Generic (PLEG): container finished" podID="690339d6-9d90-46b3-b471-afabc2da9e7c" containerID="520722ded67872ccc942a42008a87b2908384fdc9b316a93acadb80181fde2e9" exitCode=0 Nov 25 10:50:15 crc kubenswrapper[4821]: I1125 10:50:15.242866 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-84cbbf9b54-k7shj" event={"ID":"690339d6-9d90-46b3-b471-afabc2da9e7c","Type":"ContainerDied","Data":"520722ded67872ccc942a42008a87b2908384fdc9b316a93acadb80181fde2e9"} Nov 25 10:50:15 crc kubenswrapper[4821]: I1125 10:50:15.248306 4821 generic.go:334] "Generic (PLEG): container finished" podID="1647cf99-f6d6-421d-a00b-86538c7a7575" containerID="e7403b14fd72da5014034f95716d4e786968445593065d9e47e2dba3a1b41e4b" exitCode=0 Nov 25 10:50:15 crc kubenswrapper[4821]: I1125 10:50:15.248332 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1647cf99-f6d6-421d-a00b-86538c7a7575","Type":"ContainerDied","Data":"e7403b14fd72da5014034f95716d4e786968445593065d9e47e2dba3a1b41e4b"} Nov 25 10:50:15 crc kubenswrapper[4821]: I1125 10:50:15.356209 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 10:50:15 crc kubenswrapper[4821]: I1125 10:50:15.356493 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="16fdcebd-383a-4945-a5bf-a299b332bf8f" containerName="glance-log" containerID="cri-o://969ba4671a57151d49576d78ea3b5bfd6a8af49264c276fc5f8b40805e909f57" gracePeriod=30 Nov 25 10:50:15 crc kubenswrapper[4821]: I1125 10:50:15.356561 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="16fdcebd-383a-4945-a5bf-a299b332bf8f" containerName="glance-httpd" containerID="cri-o://9c59a92b8bdcb2afaeb56232da0557ea8d992d2a693fb7bc91ebdcbb5d6e0940" gracePeriod=30 Nov 25 10:50:16 crc kubenswrapper[4821]: I1125 10:50:16.261691 4821 generic.go:334] "Generic (PLEG): container finished" podID="16fdcebd-383a-4945-a5bf-a299b332bf8f" containerID="969ba4671a57151d49576d78ea3b5bfd6a8af49264c276fc5f8b40805e909f57" exitCode=143 Nov 25 10:50:16 crc kubenswrapper[4821]: I1125 10:50:16.261984 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"16fdcebd-383a-4945-a5bf-a299b332bf8f","Type":"ContainerDied","Data":"969ba4671a57151d49576d78ea3b5bfd6a8af49264c276fc5f8b40805e909f57"} Nov 25 10:50:16 crc kubenswrapper[4821]: I1125 10:50:16.267948 4821 generic.go:334] "Generic (PLEG): container finished" podID="1647cf99-f6d6-421d-a00b-86538c7a7575" containerID="0073627f3d04f03ca74fbb9e3c020f654c964b11997f56a69f24e0f4144dd25f" exitCode=0 Nov 25 10:50:16 crc kubenswrapper[4821]: I1125 10:50:16.267987 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1647cf99-f6d6-421d-a00b-86538c7a7575","Type":"ContainerDied","Data":"0073627f3d04f03ca74fbb9e3c020f654c964b11997f56a69f24e0f4144dd25f"} Nov 25 10:50:16 crc kubenswrapper[4821]: I1125 10:50:16.980237 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 10:50:16 crc kubenswrapper[4821]: I1125 10:50:16.980811 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="2e08ffdf-0b33-475b-b040-c0ecda45ff57" containerName="glance-log" containerID="cri-o://ea6716a5af0656e1213bff87beb5471adc668c54ebd32c9e0f829ff8a91fdde6" gracePeriod=30 Nov 25 10:50:16 crc kubenswrapper[4821]: I1125 10:50:16.980906 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="2e08ffdf-0b33-475b-b040-c0ecda45ff57" containerName="glance-httpd" containerID="cri-o://583283375d3838767a1436fa879285183d56c9e61af3523ef043f60584199559" gracePeriod=30 Nov 25 10:50:17 crc kubenswrapper[4821]: I1125 10:50:17.281202 4821 generic.go:334] "Generic (PLEG): container finished" podID="2e08ffdf-0b33-475b-b040-c0ecda45ff57" containerID="ea6716a5af0656e1213bff87beb5471adc668c54ebd32c9e0f829ff8a91fdde6" exitCode=143 Nov 25 10:50:17 crc kubenswrapper[4821]: I1125 10:50:17.281278 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2e08ffdf-0b33-475b-b040-c0ecda45ff57","Type":"ContainerDied","Data":"ea6716a5af0656e1213bff87beb5471adc668c54ebd32c9e0f829ff8a91fdde6"} Nov 25 10:50:19 crc kubenswrapper[4821]: I1125 10:50:19.304450 4821 generic.go:334] "Generic (PLEG): container finished" podID="16fdcebd-383a-4945-a5bf-a299b332bf8f" containerID="9c59a92b8bdcb2afaeb56232da0557ea8d992d2a693fb7bc91ebdcbb5d6e0940" exitCode=0 Nov 25 10:50:19 crc kubenswrapper[4821]: I1125 10:50:19.305432 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"16fdcebd-383a-4945-a5bf-a299b332bf8f","Type":"ContainerDied","Data":"9c59a92b8bdcb2afaeb56232da0557ea8d992d2a693fb7bc91ebdcbb5d6e0940"} Nov 25 10:50:20 crc kubenswrapper[4821]: I1125 10:50:20.155252 4821 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="2e08ffdf-0b33-475b-b040-c0ecda45ff57" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.152:9292/healthcheck\": read tcp 10.217.0.2:49406->10.217.0.152:9292: read: connection reset by peer" Nov 25 10:50:20 crc kubenswrapper[4821]: I1125 10:50:20.155953 4821 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="2e08ffdf-0b33-475b-b040-c0ecda45ff57" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.152:9292/healthcheck\": read tcp 10.217.0.2:49408->10.217.0.152:9292: read: connection reset by peer" Nov 25 10:50:20 crc kubenswrapper[4821]: I1125 10:50:20.335248 4821 generic.go:334] "Generic (PLEG): container finished" podID="2e08ffdf-0b33-475b-b040-c0ecda45ff57" containerID="583283375d3838767a1436fa879285183d56c9e61af3523ef043f60584199559" exitCode=0 Nov 25 10:50:20 crc kubenswrapper[4821]: I1125 10:50:20.335292 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2e08ffdf-0b33-475b-b040-c0ecda45ff57","Type":"ContainerDied","Data":"583283375d3838767a1436fa879285183d56c9e61af3523ef043f60584199559"} Nov 25 10:50:20 crc kubenswrapper[4821]: I1125 10:50:20.996033 4821 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-869c6f5bfb-l785w" podUID="e2327ad3-5d88-4f4b-91ea-91deaf7fc249" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Nov 25 10:50:21 crc kubenswrapper[4821]: I1125 10:50:21.447677 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:50:21 crc kubenswrapper[4821]: I1125 10:50:21.489549 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1647cf99-f6d6-421d-a00b-86538c7a7575-sg-core-conf-yaml\") pod \"1647cf99-f6d6-421d-a00b-86538c7a7575\" (UID: \"1647cf99-f6d6-421d-a00b-86538c7a7575\") " Nov 25 10:50:21 crc kubenswrapper[4821]: I1125 10:50:21.489611 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pm9qx\" (UniqueName: \"kubernetes.io/projected/1647cf99-f6d6-421d-a00b-86538c7a7575-kube-api-access-pm9qx\") pod \"1647cf99-f6d6-421d-a00b-86538c7a7575\" (UID: \"1647cf99-f6d6-421d-a00b-86538c7a7575\") " Nov 25 10:50:21 crc kubenswrapper[4821]: I1125 10:50:21.489662 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1647cf99-f6d6-421d-a00b-86538c7a7575-combined-ca-bundle\") pod \"1647cf99-f6d6-421d-a00b-86538c7a7575\" (UID: \"1647cf99-f6d6-421d-a00b-86538c7a7575\") " Nov 25 10:50:21 crc kubenswrapper[4821]: I1125 10:50:21.489767 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1647cf99-f6d6-421d-a00b-86538c7a7575-scripts\") pod \"1647cf99-f6d6-421d-a00b-86538c7a7575\" (UID: \"1647cf99-f6d6-421d-a00b-86538c7a7575\") " Nov 25 10:50:21 crc kubenswrapper[4821]: I1125 10:50:21.489854 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1647cf99-f6d6-421d-a00b-86538c7a7575-config-data\") pod \"1647cf99-f6d6-421d-a00b-86538c7a7575\" (UID: \"1647cf99-f6d6-421d-a00b-86538c7a7575\") " Nov 25 10:50:21 crc kubenswrapper[4821]: I1125 10:50:21.489928 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1647cf99-f6d6-421d-a00b-86538c7a7575-log-httpd\") pod \"1647cf99-f6d6-421d-a00b-86538c7a7575\" (UID: \"1647cf99-f6d6-421d-a00b-86538c7a7575\") " Nov 25 10:50:21 crc kubenswrapper[4821]: I1125 10:50:21.489992 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1647cf99-f6d6-421d-a00b-86538c7a7575-run-httpd\") pod \"1647cf99-f6d6-421d-a00b-86538c7a7575\" (UID: \"1647cf99-f6d6-421d-a00b-86538c7a7575\") " Nov 25 10:50:21 crc kubenswrapper[4821]: I1125 10:50:21.491138 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1647cf99-f6d6-421d-a00b-86538c7a7575-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "1647cf99-f6d6-421d-a00b-86538c7a7575" (UID: "1647cf99-f6d6-421d-a00b-86538c7a7575"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:50:21 crc kubenswrapper[4821]: I1125 10:50:21.491495 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1647cf99-f6d6-421d-a00b-86538c7a7575-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "1647cf99-f6d6-421d-a00b-86538c7a7575" (UID: "1647cf99-f6d6-421d-a00b-86538c7a7575"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:50:21 crc kubenswrapper[4821]: I1125 10:50:21.502694 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1647cf99-f6d6-421d-a00b-86538c7a7575-kube-api-access-pm9qx" (OuterVolumeSpecName: "kube-api-access-pm9qx") pod "1647cf99-f6d6-421d-a00b-86538c7a7575" (UID: "1647cf99-f6d6-421d-a00b-86538c7a7575"). InnerVolumeSpecName "kube-api-access-pm9qx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:50:21 crc kubenswrapper[4821]: I1125 10:50:21.503754 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1647cf99-f6d6-421d-a00b-86538c7a7575-scripts" (OuterVolumeSpecName: "scripts") pod "1647cf99-f6d6-421d-a00b-86538c7a7575" (UID: "1647cf99-f6d6-421d-a00b-86538c7a7575"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:50:21 crc kubenswrapper[4821]: I1125 10:50:21.530068 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1647cf99-f6d6-421d-a00b-86538c7a7575-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "1647cf99-f6d6-421d-a00b-86538c7a7575" (UID: "1647cf99-f6d6-421d-a00b-86538c7a7575"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:50:21 crc kubenswrapper[4821]: I1125 10:50:21.559336 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 25 10:50:21 crc kubenswrapper[4821]: I1125 10:50:21.591636 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"16fdcebd-383a-4945-a5bf-a299b332bf8f\" (UID: \"16fdcebd-383a-4945-a5bf-a299b332bf8f\") " Nov 25 10:50:21 crc kubenswrapper[4821]: I1125 10:50:21.591940 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16fdcebd-383a-4945-a5bf-a299b332bf8f-scripts\") pod \"16fdcebd-383a-4945-a5bf-a299b332bf8f\" (UID: \"16fdcebd-383a-4945-a5bf-a299b332bf8f\") " Nov 25 10:50:21 crc kubenswrapper[4821]: I1125 10:50:21.591975 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8nntd\" (UniqueName: \"kubernetes.io/projected/16fdcebd-383a-4945-a5bf-a299b332bf8f-kube-api-access-8nntd\") pod \"16fdcebd-383a-4945-a5bf-a299b332bf8f\" (UID: \"16fdcebd-383a-4945-a5bf-a299b332bf8f\") " Nov 25 10:50:21 crc kubenswrapper[4821]: I1125 10:50:21.592015 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/16fdcebd-383a-4945-a5bf-a299b332bf8f-internal-tls-certs\") pod \"16fdcebd-383a-4945-a5bf-a299b332bf8f\" (UID: \"16fdcebd-383a-4945-a5bf-a299b332bf8f\") " Nov 25 10:50:21 crc kubenswrapper[4821]: I1125 10:50:21.592084 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/16fdcebd-383a-4945-a5bf-a299b332bf8f-httpd-run\") pod \"16fdcebd-383a-4945-a5bf-a299b332bf8f\" (UID: \"16fdcebd-383a-4945-a5bf-a299b332bf8f\") " Nov 25 10:50:21 crc kubenswrapper[4821]: I1125 10:50:21.592316 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/16fdcebd-383a-4945-a5bf-a299b332bf8f-logs\") pod \"16fdcebd-383a-4945-a5bf-a299b332bf8f\" (UID: \"16fdcebd-383a-4945-a5bf-a299b332bf8f\") " Nov 25 10:50:21 crc kubenswrapper[4821]: I1125 10:50:21.592345 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16fdcebd-383a-4945-a5bf-a299b332bf8f-combined-ca-bundle\") pod \"16fdcebd-383a-4945-a5bf-a299b332bf8f\" (UID: \"16fdcebd-383a-4945-a5bf-a299b332bf8f\") " Nov 25 10:50:21 crc kubenswrapper[4821]: I1125 10:50:21.592379 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16fdcebd-383a-4945-a5bf-a299b332bf8f-config-data\") pod \"16fdcebd-383a-4945-a5bf-a299b332bf8f\" (UID: \"16fdcebd-383a-4945-a5bf-a299b332bf8f\") " Nov 25 10:50:21 crc kubenswrapper[4821]: I1125 10:50:21.592807 4821 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1647cf99-f6d6-421d-a00b-86538c7a7575-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:21 crc kubenswrapper[4821]: I1125 10:50:21.592821 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pm9qx\" (UniqueName: \"kubernetes.io/projected/1647cf99-f6d6-421d-a00b-86538c7a7575-kube-api-access-pm9qx\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:21 crc kubenswrapper[4821]: I1125 10:50:21.592830 4821 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1647cf99-f6d6-421d-a00b-86538c7a7575-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:21 crc kubenswrapper[4821]: I1125 10:50:21.592838 4821 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1647cf99-f6d6-421d-a00b-86538c7a7575-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:21 crc kubenswrapper[4821]: I1125 10:50:21.592846 4821 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1647cf99-f6d6-421d-a00b-86538c7a7575-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:21 crc kubenswrapper[4821]: I1125 10:50:21.593319 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/16fdcebd-383a-4945-a5bf-a299b332bf8f-logs" (OuterVolumeSpecName: "logs") pod "16fdcebd-383a-4945-a5bf-a299b332bf8f" (UID: "16fdcebd-383a-4945-a5bf-a299b332bf8f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:50:21 crc kubenswrapper[4821]: I1125 10:50:21.593479 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/16fdcebd-383a-4945-a5bf-a299b332bf8f-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "16fdcebd-383a-4945-a5bf-a299b332bf8f" (UID: "16fdcebd-383a-4945-a5bf-a299b332bf8f"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:50:21 crc kubenswrapper[4821]: I1125 10:50:21.657771 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16fdcebd-383a-4945-a5bf-a299b332bf8f-kube-api-access-8nntd" (OuterVolumeSpecName: "kube-api-access-8nntd") pod "16fdcebd-383a-4945-a5bf-a299b332bf8f" (UID: "16fdcebd-383a-4945-a5bf-a299b332bf8f"). InnerVolumeSpecName "kube-api-access-8nntd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:50:21 crc kubenswrapper[4821]: I1125 10:50:21.657929 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16fdcebd-383a-4945-a5bf-a299b332bf8f-scripts" (OuterVolumeSpecName: "scripts") pod "16fdcebd-383a-4945-a5bf-a299b332bf8f" (UID: "16fdcebd-383a-4945-a5bf-a299b332bf8f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:50:21 crc kubenswrapper[4821]: I1125 10:50:21.658643 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "16fdcebd-383a-4945-a5bf-a299b332bf8f" (UID: "16fdcebd-383a-4945-a5bf-a299b332bf8f"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 10:50:21 crc kubenswrapper[4821]: I1125 10:50:21.658738 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1647cf99-f6d6-421d-a00b-86538c7a7575-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1647cf99-f6d6-421d-a00b-86538c7a7575" (UID: "1647cf99-f6d6-421d-a00b-86538c7a7575"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:50:21 crc kubenswrapper[4821]: I1125 10:50:21.677338 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16fdcebd-383a-4945-a5bf-a299b332bf8f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "16fdcebd-383a-4945-a5bf-a299b332bf8f" (UID: "16fdcebd-383a-4945-a5bf-a299b332bf8f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:50:21 crc kubenswrapper[4821]: I1125 10:50:21.677547 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1647cf99-f6d6-421d-a00b-86538c7a7575-config-data" (OuterVolumeSpecName: "config-data") pod "1647cf99-f6d6-421d-a00b-86538c7a7575" (UID: "1647cf99-f6d6-421d-a00b-86538c7a7575"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:50:21 crc kubenswrapper[4821]: I1125 10:50:21.695560 4821 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16fdcebd-383a-4945-a5bf-a299b332bf8f-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:21 crc kubenswrapper[4821]: I1125 10:50:21.695613 4821 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1647cf99-f6d6-421d-a00b-86538c7a7575-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:21 crc kubenswrapper[4821]: I1125 10:50:21.695625 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8nntd\" (UniqueName: \"kubernetes.io/projected/16fdcebd-383a-4945-a5bf-a299b332bf8f-kube-api-access-8nntd\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:21 crc kubenswrapper[4821]: I1125 10:50:21.695645 4821 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/16fdcebd-383a-4945-a5bf-a299b332bf8f-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:21 crc kubenswrapper[4821]: I1125 10:50:21.695656 4821 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1647cf99-f6d6-421d-a00b-86538c7a7575-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:21 crc kubenswrapper[4821]: I1125 10:50:21.695668 4821 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/16fdcebd-383a-4945-a5bf-a299b332bf8f-logs\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:21 crc kubenswrapper[4821]: I1125 10:50:21.695677 4821 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16fdcebd-383a-4945-a5bf-a299b332bf8f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:21 crc kubenswrapper[4821]: I1125 10:50:21.695707 4821 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Nov 25 10:50:21 crc kubenswrapper[4821]: I1125 10:50:21.716810 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16fdcebd-383a-4945-a5bf-a299b332bf8f-config-data" (OuterVolumeSpecName: "config-data") pod "16fdcebd-383a-4945-a5bf-a299b332bf8f" (UID: "16fdcebd-383a-4945-a5bf-a299b332bf8f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:50:21 crc kubenswrapper[4821]: I1125 10:50:21.733288 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16fdcebd-383a-4945-a5bf-a299b332bf8f-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "16fdcebd-383a-4945-a5bf-a299b332bf8f" (UID: "16fdcebd-383a-4945-a5bf-a299b332bf8f"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:50:21 crc kubenswrapper[4821]: I1125 10:50:21.749707 4821 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Nov 25 10:50:21 crc kubenswrapper[4821]: I1125 10:50:21.797798 4821 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16fdcebd-383a-4945-a5bf-a299b332bf8f-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:21 crc kubenswrapper[4821]: I1125 10:50:21.797834 4821 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:21 crc kubenswrapper[4821]: I1125 10:50:21.797844 4821 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/16fdcebd-383a-4945-a5bf-a299b332bf8f-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:21 crc kubenswrapper[4821]: I1125 10:50:21.842419 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-ff8c8c76c-ctf8b" Nov 25 10:50:21 crc kubenswrapper[4821]: I1125 10:50:21.843299 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-ff8c8c76c-ctf8b" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.193906 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-84cbbf9b54-k7shj" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.307959 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/690339d6-9d90-46b3-b471-afabc2da9e7c-config\") pod \"690339d6-9d90-46b3-b471-afabc2da9e7c\" (UID: \"690339d6-9d90-46b3-b471-afabc2da9e7c\") " Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.308113 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/690339d6-9d90-46b3-b471-afabc2da9e7c-httpd-config\") pod \"690339d6-9d90-46b3-b471-afabc2da9e7c\" (UID: \"690339d6-9d90-46b3-b471-afabc2da9e7c\") " Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.308213 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/690339d6-9d90-46b3-b471-afabc2da9e7c-combined-ca-bundle\") pod \"690339d6-9d90-46b3-b471-afabc2da9e7c\" (UID: \"690339d6-9d90-46b3-b471-afabc2da9e7c\") " Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.308262 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bmnv7\" (UniqueName: \"kubernetes.io/projected/690339d6-9d90-46b3-b471-afabc2da9e7c-kube-api-access-bmnv7\") pod \"690339d6-9d90-46b3-b471-afabc2da9e7c\" (UID: \"690339d6-9d90-46b3-b471-afabc2da9e7c\") " Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.308388 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/690339d6-9d90-46b3-b471-afabc2da9e7c-ovndb-tls-certs\") pod \"690339d6-9d90-46b3-b471-afabc2da9e7c\" (UID: \"690339d6-9d90-46b3-b471-afabc2da9e7c\") " Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.313445 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/690339d6-9d90-46b3-b471-afabc2da9e7c-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "690339d6-9d90-46b3-b471-afabc2da9e7c" (UID: "690339d6-9d90-46b3-b471-afabc2da9e7c"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.313612 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/690339d6-9d90-46b3-b471-afabc2da9e7c-kube-api-access-bmnv7" (OuterVolumeSpecName: "kube-api-access-bmnv7") pod "690339d6-9d90-46b3-b471-afabc2da9e7c" (UID: "690339d6-9d90-46b3-b471-afabc2da9e7c"). InnerVolumeSpecName "kube-api-access-bmnv7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.360494 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"16fdcebd-383a-4945-a5bf-a299b332bf8f","Type":"ContainerDied","Data":"0df33c820c10e12b24d38e8d4319533b7c9fd7ae73003b4ee9209b4d014aa9bd"} Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.360520 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.360547 4821 scope.go:117] "RemoveContainer" containerID="9c59a92b8bdcb2afaeb56232da0557ea8d992d2a693fb7bc91ebdcbb5d6e0940" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.365866 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1647cf99-f6d6-421d-a00b-86538c7a7575","Type":"ContainerDied","Data":"3c19f2f14248485cd53184e9f83662d006581b8e8df8c8b58a517328fe292758"} Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.365928 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.371139 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-84cbbf9b54-k7shj" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.371382 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-84cbbf9b54-k7shj" event={"ID":"690339d6-9d90-46b3-b471-afabc2da9e7c","Type":"ContainerDied","Data":"7c8f465e6e511bedd5549cf2ae1805de33731586756bc5ef47c073dce9b38b0c"} Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.390041 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/690339d6-9d90-46b3-b471-afabc2da9e7c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "690339d6-9d90-46b3-b471-afabc2da9e7c" (UID: "690339d6-9d90-46b3-b471-afabc2da9e7c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.404179 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/690339d6-9d90-46b3-b471-afabc2da9e7c-config" (OuterVolumeSpecName: "config") pod "690339d6-9d90-46b3-b471-afabc2da9e7c" (UID: "690339d6-9d90-46b3-b471-afabc2da9e7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.411391 4821 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/690339d6-9d90-46b3-b471-afabc2da9e7c-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.411424 4821 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/690339d6-9d90-46b3-b471-afabc2da9e7c-httpd-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.411434 4821 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/690339d6-9d90-46b3-b471-afabc2da9e7c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.411444 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bmnv7\" (UniqueName: \"kubernetes.io/projected/690339d6-9d90-46b3-b471-afabc2da9e7c-kube-api-access-bmnv7\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.426014 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.437183 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.437194 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/690339d6-9d90-46b3-b471-afabc2da9e7c-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "690339d6-9d90-46b3-b471-afabc2da9e7c" (UID: "690339d6-9d90-46b3-b471-afabc2da9e7c"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.448848 4821 scope.go:117] "RemoveContainer" containerID="969ba4671a57151d49576d78ea3b5bfd6a8af49264c276fc5f8b40805e909f57" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.456517 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.465176 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:50:22 crc kubenswrapper[4821]: E1125 10:50:22.465684 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16fdcebd-383a-4945-a5bf-a299b332bf8f" containerName="glance-log" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.465702 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="16fdcebd-383a-4945-a5bf-a299b332bf8f" containerName="glance-log" Nov 25 10:50:22 crc kubenswrapper[4821]: E1125 10:50:22.465718 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1647cf99-f6d6-421d-a00b-86538c7a7575" containerName="sg-core" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.465727 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="1647cf99-f6d6-421d-a00b-86538c7a7575" containerName="sg-core" Nov 25 10:50:22 crc kubenswrapper[4821]: E1125 10:50:22.465743 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1647cf99-f6d6-421d-a00b-86538c7a7575" containerName="ceilometer-central-agent" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.465751 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="1647cf99-f6d6-421d-a00b-86538c7a7575" containerName="ceilometer-central-agent" Nov 25 10:50:22 crc kubenswrapper[4821]: E1125 10:50:22.465765 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="690339d6-9d90-46b3-b471-afabc2da9e7c" containerName="neutron-httpd" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.465773 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="690339d6-9d90-46b3-b471-afabc2da9e7c" containerName="neutron-httpd" Nov 25 10:50:22 crc kubenswrapper[4821]: E1125 10:50:22.465791 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1647cf99-f6d6-421d-a00b-86538c7a7575" containerName="proxy-httpd" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.465799 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="1647cf99-f6d6-421d-a00b-86538c7a7575" containerName="proxy-httpd" Nov 25 10:50:22 crc kubenswrapper[4821]: E1125 10:50:22.465823 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="690339d6-9d90-46b3-b471-afabc2da9e7c" containerName="neutron-api" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.465830 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="690339d6-9d90-46b3-b471-afabc2da9e7c" containerName="neutron-api" Nov 25 10:50:22 crc kubenswrapper[4821]: E1125 10:50:22.465846 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1647cf99-f6d6-421d-a00b-86538c7a7575" containerName="ceilometer-notification-agent" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.465853 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="1647cf99-f6d6-421d-a00b-86538c7a7575" containerName="ceilometer-notification-agent" Nov 25 10:50:22 crc kubenswrapper[4821]: E1125 10:50:22.465868 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16fdcebd-383a-4945-a5bf-a299b332bf8f" containerName="glance-httpd" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.465875 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="16fdcebd-383a-4945-a5bf-a299b332bf8f" containerName="glance-httpd" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.466091 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="1647cf99-f6d6-421d-a00b-86538c7a7575" containerName="proxy-httpd" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.466104 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="16fdcebd-383a-4945-a5bf-a299b332bf8f" containerName="glance-log" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.466117 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="1647cf99-f6d6-421d-a00b-86538c7a7575" containerName="sg-core" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.466134 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="16fdcebd-383a-4945-a5bf-a299b332bf8f" containerName="glance-httpd" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.466154 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="1647cf99-f6d6-421d-a00b-86538c7a7575" containerName="ceilometer-central-agent" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.466182 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="690339d6-9d90-46b3-b471-afabc2da9e7c" containerName="neutron-httpd" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.466195 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="1647cf99-f6d6-421d-a00b-86538c7a7575" containerName="ceilometer-notification-agent" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.466210 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="690339d6-9d90-46b3-b471-afabc2da9e7c" containerName="neutron-api" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.469504 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.490235 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.505390 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.516262 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.516507 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bdacaaac-e3bf-4df9-8c19-e0e1315e5a44-log-httpd\") pod \"ceilometer-0\" (UID: \"bdacaaac-e3bf-4df9-8c19-e0e1315e5a44\") " pod="openstack/ceilometer-0" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.516653 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59lxj\" (UniqueName: \"kubernetes.io/projected/bdacaaac-e3bf-4df9-8c19-e0e1315e5a44-kube-api-access-59lxj\") pod \"ceilometer-0\" (UID: \"bdacaaac-e3bf-4df9-8c19-e0e1315e5a44\") " pod="openstack/ceilometer-0" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.516823 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bdacaaac-e3bf-4df9-8c19-e0e1315e5a44-scripts\") pod \"ceilometer-0\" (UID: \"bdacaaac-e3bf-4df9-8c19-e0e1315e5a44\") " pod="openstack/ceilometer-0" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.516863 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bdacaaac-e3bf-4df9-8c19-e0e1315e5a44-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bdacaaac-e3bf-4df9-8c19-e0e1315e5a44\") " pod="openstack/ceilometer-0" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.516942 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdacaaac-e3bf-4df9-8c19-e0e1315e5a44-config-data\") pod \"ceilometer-0\" (UID: \"bdacaaac-e3bf-4df9-8c19-e0e1315e5a44\") " pod="openstack/ceilometer-0" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.517002 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdacaaac-e3bf-4df9-8c19-e0e1315e5a44-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bdacaaac-e3bf-4df9-8c19-e0e1315e5a44\") " pod="openstack/ceilometer-0" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.517036 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bdacaaac-e3bf-4df9-8c19-e0e1315e5a44-run-httpd\") pod \"ceilometer-0\" (UID: \"bdacaaac-e3bf-4df9-8c19-e0e1315e5a44\") " pod="openstack/ceilometer-0" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.517194 4821 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/690339d6-9d90-46b3-b471-afabc2da9e7c-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.517495 4821 scope.go:117] "RemoveContainer" containerID="b57c43f68f8bdd2464535935e220533ae71fe723e2b8bf249a896b43a1b9bc9c" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.543077 4821 scope.go:117] "RemoveContainer" containerID="7b894abf033fdc87c1547963b5822943b6b998050779618be7999ac3afa4aa3b" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.560393 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.580048 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.581939 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.584483 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.584809 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.601757 4821 scope.go:117] "RemoveContainer" containerID="0073627f3d04f03ca74fbb9e3c020f654c964b11997f56a69f24e0f4144dd25f" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.606473 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.618664 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bdacaaac-e3bf-4df9-8c19-e0e1315e5a44-scripts\") pod \"ceilometer-0\" (UID: \"bdacaaac-e3bf-4df9-8c19-e0e1315e5a44\") " pod="openstack/ceilometer-0" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.618727 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bdacaaac-e3bf-4df9-8c19-e0e1315e5a44-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bdacaaac-e3bf-4df9-8c19-e0e1315e5a44\") " pod="openstack/ceilometer-0" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.618786 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdacaaac-e3bf-4df9-8c19-e0e1315e5a44-config-data\") pod \"ceilometer-0\" (UID: \"bdacaaac-e3bf-4df9-8c19-e0e1315e5a44\") " pod="openstack/ceilometer-0" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.618828 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdacaaac-e3bf-4df9-8c19-e0e1315e5a44-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bdacaaac-e3bf-4df9-8c19-e0e1315e5a44\") " pod="openstack/ceilometer-0" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.618850 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bdacaaac-e3bf-4df9-8c19-e0e1315e5a44-run-httpd\") pod \"ceilometer-0\" (UID: \"bdacaaac-e3bf-4df9-8c19-e0e1315e5a44\") " pod="openstack/ceilometer-0" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.618903 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bdacaaac-e3bf-4df9-8c19-e0e1315e5a44-log-httpd\") pod \"ceilometer-0\" (UID: \"bdacaaac-e3bf-4df9-8c19-e0e1315e5a44\") " pod="openstack/ceilometer-0" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.618970 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59lxj\" (UniqueName: \"kubernetes.io/projected/bdacaaac-e3bf-4df9-8c19-e0e1315e5a44-kube-api-access-59lxj\") pod \"ceilometer-0\" (UID: \"bdacaaac-e3bf-4df9-8c19-e0e1315e5a44\") " pod="openstack/ceilometer-0" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.621760 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bdacaaac-e3bf-4df9-8c19-e0e1315e5a44-log-httpd\") pod \"ceilometer-0\" (UID: \"bdacaaac-e3bf-4df9-8c19-e0e1315e5a44\") " pod="openstack/ceilometer-0" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.622957 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bdacaaac-e3bf-4df9-8c19-e0e1315e5a44-run-httpd\") pod \"ceilometer-0\" (UID: \"bdacaaac-e3bf-4df9-8c19-e0e1315e5a44\") " pod="openstack/ceilometer-0" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.626214 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdacaaac-e3bf-4df9-8c19-e0e1315e5a44-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bdacaaac-e3bf-4df9-8c19-e0e1315e5a44\") " pod="openstack/ceilometer-0" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.626374 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bdacaaac-e3bf-4df9-8c19-e0e1315e5a44-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bdacaaac-e3bf-4df9-8c19-e0e1315e5a44\") " pod="openstack/ceilometer-0" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.629299 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdacaaac-e3bf-4df9-8c19-e0e1315e5a44-config-data\") pod \"ceilometer-0\" (UID: \"bdacaaac-e3bf-4df9-8c19-e0e1315e5a44\") " pod="openstack/ceilometer-0" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.634636 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bdacaaac-e3bf-4df9-8c19-e0e1315e5a44-scripts\") pod \"ceilometer-0\" (UID: \"bdacaaac-e3bf-4df9-8c19-e0e1315e5a44\") " pod="openstack/ceilometer-0" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.635883 4821 scope.go:117] "RemoveContainer" containerID="e7403b14fd72da5014034f95716d4e786968445593065d9e47e2dba3a1b41e4b" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.638453 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59lxj\" (UniqueName: \"kubernetes.io/projected/bdacaaac-e3bf-4df9-8c19-e0e1315e5a44-kube-api-access-59lxj\") pod \"ceilometer-0\" (UID: \"bdacaaac-e3bf-4df9-8c19-e0e1315e5a44\") " pod="openstack/ceilometer-0" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.638609 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.662396 4821 scope.go:117] "RemoveContainer" containerID="9fc413e8b8971197f05d3f9d1675fd56f8c25a1a5c97327431248c71d7982d6f" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.702620 4821 scope.go:117] "RemoveContainer" containerID="520722ded67872ccc942a42008a87b2908384fdc9b316a93acadb80181fde2e9" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.710423 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-84cbbf9b54-k7shj"] Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.719051 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-84cbbf9b54-k7shj"] Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.723404 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1816fd90-5883-4de2-9124-64e326d1743f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"1816fd90-5883-4de2-9124-64e326d1743f\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.724121 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1816fd90-5883-4de2-9124-64e326d1743f-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"1816fd90-5883-4de2-9124-64e326d1743f\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.724170 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1816fd90-5883-4de2-9124-64e326d1743f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1816fd90-5883-4de2-9124-64e326d1743f\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.724195 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"1816fd90-5883-4de2-9124-64e326d1743f\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.724372 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1816fd90-5883-4de2-9124-64e326d1743f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1816fd90-5883-4de2-9124-64e326d1743f\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.724403 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1816fd90-5883-4de2-9124-64e326d1743f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1816fd90-5883-4de2-9124-64e326d1743f\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.724470 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1816fd90-5883-4de2-9124-64e326d1743f-logs\") pod \"glance-default-internal-api-0\" (UID: \"1816fd90-5883-4de2-9124-64e326d1743f\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.724587 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-klklz\" (UniqueName: \"kubernetes.io/projected/1816fd90-5883-4de2-9124-64e326d1743f-kube-api-access-klklz\") pod \"glance-default-internal-api-0\" (UID: \"1816fd90-5883-4de2-9124-64e326d1743f\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.824732 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.825965 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"2e08ffdf-0b33-475b-b040-c0ecda45ff57\" (UID: \"2e08ffdf-0b33-475b-b040-c0ecda45ff57\") " Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.826066 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e08ffdf-0b33-475b-b040-c0ecda45ff57-config-data\") pod \"2e08ffdf-0b33-475b-b040-c0ecda45ff57\" (UID: \"2e08ffdf-0b33-475b-b040-c0ecda45ff57\") " Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.826265 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e08ffdf-0b33-475b-b040-c0ecda45ff57-public-tls-certs\") pod \"2e08ffdf-0b33-475b-b040-c0ecda45ff57\" (UID: \"2e08ffdf-0b33-475b-b040-c0ecda45ff57\") " Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.826317 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2e08ffdf-0b33-475b-b040-c0ecda45ff57-httpd-run\") pod \"2e08ffdf-0b33-475b-b040-c0ecda45ff57\" (UID: \"2e08ffdf-0b33-475b-b040-c0ecda45ff57\") " Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.826342 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e08ffdf-0b33-475b-b040-c0ecda45ff57-scripts\") pod \"2e08ffdf-0b33-475b-b040-c0ecda45ff57\" (UID: \"2e08ffdf-0b33-475b-b040-c0ecda45ff57\") " Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.826372 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2p6qs\" (UniqueName: \"kubernetes.io/projected/2e08ffdf-0b33-475b-b040-c0ecda45ff57-kube-api-access-2p6qs\") pod \"2e08ffdf-0b33-475b-b040-c0ecda45ff57\" (UID: \"2e08ffdf-0b33-475b-b040-c0ecda45ff57\") " Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.826447 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e08ffdf-0b33-475b-b040-c0ecda45ff57-logs\") pod \"2e08ffdf-0b33-475b-b040-c0ecda45ff57\" (UID: \"2e08ffdf-0b33-475b-b040-c0ecda45ff57\") " Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.826483 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e08ffdf-0b33-475b-b040-c0ecda45ff57-combined-ca-bundle\") pod \"2e08ffdf-0b33-475b-b040-c0ecda45ff57\" (UID: \"2e08ffdf-0b33-475b-b040-c0ecda45ff57\") " Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.827872 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e08ffdf-0b33-475b-b040-c0ecda45ff57-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "2e08ffdf-0b33-475b-b040-c0ecda45ff57" (UID: "2e08ffdf-0b33-475b-b040-c0ecda45ff57"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.828416 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e08ffdf-0b33-475b-b040-c0ecda45ff57-logs" (OuterVolumeSpecName: "logs") pod "2e08ffdf-0b33-475b-b040-c0ecda45ff57" (UID: "2e08ffdf-0b33-475b-b040-c0ecda45ff57"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.829943 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1816fd90-5883-4de2-9124-64e326d1743f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"1816fd90-5883-4de2-9124-64e326d1743f\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.830028 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1816fd90-5883-4de2-9124-64e326d1743f-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"1816fd90-5883-4de2-9124-64e326d1743f\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.830053 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1816fd90-5883-4de2-9124-64e326d1743f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1816fd90-5883-4de2-9124-64e326d1743f\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.830082 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"1816fd90-5883-4de2-9124-64e326d1743f\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.830231 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1816fd90-5883-4de2-9124-64e326d1743f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1816fd90-5883-4de2-9124-64e326d1743f\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.830265 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1816fd90-5883-4de2-9124-64e326d1743f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1816fd90-5883-4de2-9124-64e326d1743f\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.830313 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1816fd90-5883-4de2-9124-64e326d1743f-logs\") pod \"glance-default-internal-api-0\" (UID: \"1816fd90-5883-4de2-9124-64e326d1743f\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.830400 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-klklz\" (UniqueName: \"kubernetes.io/projected/1816fd90-5883-4de2-9124-64e326d1743f-kube-api-access-klklz\") pod \"glance-default-internal-api-0\" (UID: \"1816fd90-5883-4de2-9124-64e326d1743f\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.830496 4821 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2e08ffdf-0b33-475b-b040-c0ecda45ff57-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.830519 4821 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e08ffdf-0b33-475b-b040-c0ecda45ff57-logs\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.831225 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e08ffdf-0b33-475b-b040-c0ecda45ff57-kube-api-access-2p6qs" (OuterVolumeSpecName: "kube-api-access-2p6qs") pod "2e08ffdf-0b33-475b-b040-c0ecda45ff57" (UID: "2e08ffdf-0b33-475b-b040-c0ecda45ff57"). InnerVolumeSpecName "kube-api-access-2p6qs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.831361 4821 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"1816fd90-5883-4de2-9124-64e326d1743f\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-internal-api-0" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.831661 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1816fd90-5883-4de2-9124-64e326d1743f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1816fd90-5883-4de2-9124-64e326d1743f\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.832102 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1816fd90-5883-4de2-9124-64e326d1743f-logs\") pod \"glance-default-internal-api-0\" (UID: \"1816fd90-5883-4de2-9124-64e326d1743f\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.834999 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance") pod "2e08ffdf-0b33-475b-b040-c0ecda45ff57" (UID: "2e08ffdf-0b33-475b-b040-c0ecda45ff57"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.836818 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1816fd90-5883-4de2-9124-64e326d1743f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1816fd90-5883-4de2-9124-64e326d1743f\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.838898 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e08ffdf-0b33-475b-b040-c0ecda45ff57-scripts" (OuterVolumeSpecName: "scripts") pod "2e08ffdf-0b33-475b-b040-c0ecda45ff57" (UID: "2e08ffdf-0b33-475b-b040-c0ecda45ff57"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.840279 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1816fd90-5883-4de2-9124-64e326d1743f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1816fd90-5883-4de2-9124-64e326d1743f\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.849254 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1816fd90-5883-4de2-9124-64e326d1743f-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"1816fd90-5883-4de2-9124-64e326d1743f\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.850487 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1816fd90-5883-4de2-9124-64e326d1743f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"1816fd90-5883-4de2-9124-64e326d1743f\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.854049 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-klklz\" (UniqueName: \"kubernetes.io/projected/1816fd90-5883-4de2-9124-64e326d1743f-kube-api-access-klklz\") pod \"glance-default-internal-api-0\" (UID: \"1816fd90-5883-4de2-9124-64e326d1743f\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.884740 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e08ffdf-0b33-475b-b040-c0ecda45ff57-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2e08ffdf-0b33-475b-b040-c0ecda45ff57" (UID: "2e08ffdf-0b33-475b-b040-c0ecda45ff57"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.885066 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"1816fd90-5883-4de2-9124-64e326d1743f\") " pod="openstack/glance-default-internal-api-0" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.910659 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.911389 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e08ffdf-0b33-475b-b040-c0ecda45ff57-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "2e08ffdf-0b33-475b-b040-c0ecda45ff57" (UID: "2e08ffdf-0b33-475b-b040-c0ecda45ff57"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.932483 4821 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e08ffdf-0b33-475b-b040-c0ecda45ff57-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.932519 4821 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e08ffdf-0b33-475b-b040-c0ecda45ff57-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.932529 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2p6qs\" (UniqueName: \"kubernetes.io/projected/2e08ffdf-0b33-475b-b040-c0ecda45ff57-kube-api-access-2p6qs\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.932539 4821 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e08ffdf-0b33-475b-b040-c0ecda45ff57-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.932568 4821 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.952278 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e08ffdf-0b33-475b-b040-c0ecda45ff57-config-data" (OuterVolumeSpecName: "config-data") pod "2e08ffdf-0b33-475b-b040-c0ecda45ff57" (UID: "2e08ffdf-0b33-475b-b040-c0ecda45ff57"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:50:22 crc kubenswrapper[4821]: I1125 10:50:22.967477 4821 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Nov 25 10:50:23 crc kubenswrapper[4821]: I1125 10:50:23.034898 4821 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:23 crc kubenswrapper[4821]: I1125 10:50:23.034924 4821 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e08ffdf-0b33-475b-b040-c0ecda45ff57-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:23 crc kubenswrapper[4821]: I1125 10:50:23.311389 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:50:23 crc kubenswrapper[4821]: W1125 10:50:23.315033 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbdacaaac_e3bf_4df9_8c19_e0e1315e5a44.slice/crio-f27722b39dc103c305a9fc6f15526c02c0796801f8c195a5e1249dba6a0017d6 WatchSource:0}: Error finding container f27722b39dc103c305a9fc6f15526c02c0796801f8c195a5e1249dba6a0017d6: Status 404 returned error can't find the container with id f27722b39dc103c305a9fc6f15526c02c0796801f8c195a5e1249dba6a0017d6 Nov 25 10:50:23 crc kubenswrapper[4821]: I1125 10:50:23.383735 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bdacaaac-e3bf-4df9-8c19-e0e1315e5a44","Type":"ContainerStarted","Data":"f27722b39dc103c305a9fc6f15526c02c0796801f8c195a5e1249dba6a0017d6"} Nov 25 10:50:23 crc kubenswrapper[4821]: I1125 10:50:23.386702 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2e08ffdf-0b33-475b-b040-c0ecda45ff57","Type":"ContainerDied","Data":"8e7596ffeecbe1d869e74b9e5bd43193776fd6896b8cae38882cda50b189d643"} Nov 25 10:50:23 crc kubenswrapper[4821]: I1125 10:50:23.386737 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 25 10:50:23 crc kubenswrapper[4821]: I1125 10:50:23.386745 4821 scope.go:117] "RemoveContainer" containerID="583283375d3838767a1436fa879285183d56c9e61af3523ef043f60584199559" Nov 25 10:50:23 crc kubenswrapper[4821]: I1125 10:50:23.420947 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 10:50:23 crc kubenswrapper[4821]: I1125 10:50:23.428947 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 10:50:23 crc kubenswrapper[4821]: I1125 10:50:23.429592 4821 scope.go:117] "RemoveContainer" containerID="ea6716a5af0656e1213bff87beb5471adc668c54ebd32c9e0f829ff8a91fdde6" Nov 25 10:50:23 crc kubenswrapper[4821]: I1125 10:50:23.450326 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 10:50:23 crc kubenswrapper[4821]: E1125 10:50:23.450791 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e08ffdf-0b33-475b-b040-c0ecda45ff57" containerName="glance-httpd" Nov 25 10:50:23 crc kubenswrapper[4821]: I1125 10:50:23.450810 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e08ffdf-0b33-475b-b040-c0ecda45ff57" containerName="glance-httpd" Nov 25 10:50:23 crc kubenswrapper[4821]: E1125 10:50:23.450833 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e08ffdf-0b33-475b-b040-c0ecda45ff57" containerName="glance-log" Nov 25 10:50:23 crc kubenswrapper[4821]: I1125 10:50:23.450840 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e08ffdf-0b33-475b-b040-c0ecda45ff57" containerName="glance-log" Nov 25 10:50:23 crc kubenswrapper[4821]: I1125 10:50:23.452491 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e08ffdf-0b33-475b-b040-c0ecda45ff57" containerName="glance-log" Nov 25 10:50:23 crc kubenswrapper[4821]: I1125 10:50:23.452543 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e08ffdf-0b33-475b-b040-c0ecda45ff57" containerName="glance-httpd" Nov 25 10:50:23 crc kubenswrapper[4821]: I1125 10:50:23.453971 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 25 10:50:23 crc kubenswrapper[4821]: I1125 10:50:23.459553 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 25 10:50:23 crc kubenswrapper[4821]: I1125 10:50:23.459888 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 25 10:50:23 crc kubenswrapper[4821]: I1125 10:50:23.463420 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 10:50:23 crc kubenswrapper[4821]: I1125 10:50:23.554237 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fada5a81-0157-452d-9ade-dd6ce6303918-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"fada5a81-0157-452d-9ade-dd6ce6303918\") " pod="openstack/glance-default-external-api-0" Nov 25 10:50:23 crc kubenswrapper[4821]: I1125 10:50:23.554286 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fada5a81-0157-452d-9ade-dd6ce6303918-scripts\") pod \"glance-default-external-api-0\" (UID: \"fada5a81-0157-452d-9ade-dd6ce6303918\") " pod="openstack/glance-default-external-api-0" Nov 25 10:50:23 crc kubenswrapper[4821]: I1125 10:50:23.554319 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fada5a81-0157-452d-9ade-dd6ce6303918-logs\") pod \"glance-default-external-api-0\" (UID: \"fada5a81-0157-452d-9ade-dd6ce6303918\") " pod="openstack/glance-default-external-api-0" Nov 25 10:50:23 crc kubenswrapper[4821]: I1125 10:50:23.554361 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mklv5\" (UniqueName: \"kubernetes.io/projected/fada5a81-0157-452d-9ade-dd6ce6303918-kube-api-access-mklv5\") pod \"glance-default-external-api-0\" (UID: \"fada5a81-0157-452d-9ade-dd6ce6303918\") " pod="openstack/glance-default-external-api-0" Nov 25 10:50:23 crc kubenswrapper[4821]: I1125 10:50:23.554382 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fada5a81-0157-452d-9ade-dd6ce6303918-config-data\") pod \"glance-default-external-api-0\" (UID: \"fada5a81-0157-452d-9ade-dd6ce6303918\") " pod="openstack/glance-default-external-api-0" Nov 25 10:50:23 crc kubenswrapper[4821]: I1125 10:50:23.554409 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fada5a81-0157-452d-9ade-dd6ce6303918-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"fada5a81-0157-452d-9ade-dd6ce6303918\") " pod="openstack/glance-default-external-api-0" Nov 25 10:50:23 crc kubenswrapper[4821]: I1125 10:50:23.554469 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"fada5a81-0157-452d-9ade-dd6ce6303918\") " pod="openstack/glance-default-external-api-0" Nov 25 10:50:23 crc kubenswrapper[4821]: I1125 10:50:23.554500 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fada5a81-0157-452d-9ade-dd6ce6303918-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"fada5a81-0157-452d-9ade-dd6ce6303918\") " pod="openstack/glance-default-external-api-0" Nov 25 10:50:23 crc kubenswrapper[4821]: I1125 10:50:23.593375 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 10:50:23 crc kubenswrapper[4821]: W1125 10:50:23.593407 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1816fd90_5883_4de2_9124_64e326d1743f.slice/crio-1072de5bda8efe7f8239255558afe31bede763e57c868b004d21a18dba6e9dc4 WatchSource:0}: Error finding container 1072de5bda8efe7f8239255558afe31bede763e57c868b004d21a18dba6e9dc4: Status 404 returned error can't find the container with id 1072de5bda8efe7f8239255558afe31bede763e57c868b004d21a18dba6e9dc4 Nov 25 10:50:23 crc kubenswrapper[4821]: I1125 10:50:23.655858 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fada5a81-0157-452d-9ade-dd6ce6303918-logs\") pod \"glance-default-external-api-0\" (UID: \"fada5a81-0157-452d-9ade-dd6ce6303918\") " pod="openstack/glance-default-external-api-0" Nov 25 10:50:23 crc kubenswrapper[4821]: I1125 10:50:23.656095 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mklv5\" (UniqueName: \"kubernetes.io/projected/fada5a81-0157-452d-9ade-dd6ce6303918-kube-api-access-mklv5\") pod \"glance-default-external-api-0\" (UID: \"fada5a81-0157-452d-9ade-dd6ce6303918\") " pod="openstack/glance-default-external-api-0" Nov 25 10:50:23 crc kubenswrapper[4821]: I1125 10:50:23.656153 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fada5a81-0157-452d-9ade-dd6ce6303918-config-data\") pod \"glance-default-external-api-0\" (UID: \"fada5a81-0157-452d-9ade-dd6ce6303918\") " pod="openstack/glance-default-external-api-0" Nov 25 10:50:23 crc kubenswrapper[4821]: I1125 10:50:23.656259 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fada5a81-0157-452d-9ade-dd6ce6303918-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"fada5a81-0157-452d-9ade-dd6ce6303918\") " pod="openstack/glance-default-external-api-0" Nov 25 10:50:23 crc kubenswrapper[4821]: I1125 10:50:23.656381 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fada5a81-0157-452d-9ade-dd6ce6303918-logs\") pod \"glance-default-external-api-0\" (UID: \"fada5a81-0157-452d-9ade-dd6ce6303918\") " pod="openstack/glance-default-external-api-0" Nov 25 10:50:23 crc kubenswrapper[4821]: I1125 10:50:23.656442 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"fada5a81-0157-452d-9ade-dd6ce6303918\") " pod="openstack/glance-default-external-api-0" Nov 25 10:50:23 crc kubenswrapper[4821]: I1125 10:50:23.656813 4821 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"fada5a81-0157-452d-9ade-dd6ce6303918\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-external-api-0" Nov 25 10:50:23 crc kubenswrapper[4821]: I1125 10:50:23.657327 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fada5a81-0157-452d-9ade-dd6ce6303918-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"fada5a81-0157-452d-9ade-dd6ce6303918\") " pod="openstack/glance-default-external-api-0" Nov 25 10:50:23 crc kubenswrapper[4821]: I1125 10:50:23.657465 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fada5a81-0157-452d-9ade-dd6ce6303918-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"fada5a81-0157-452d-9ade-dd6ce6303918\") " pod="openstack/glance-default-external-api-0" Nov 25 10:50:23 crc kubenswrapper[4821]: I1125 10:50:23.657518 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fada5a81-0157-452d-9ade-dd6ce6303918-scripts\") pod \"glance-default-external-api-0\" (UID: \"fada5a81-0157-452d-9ade-dd6ce6303918\") " pod="openstack/glance-default-external-api-0" Nov 25 10:50:23 crc kubenswrapper[4821]: I1125 10:50:23.657661 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fada5a81-0157-452d-9ade-dd6ce6303918-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"fada5a81-0157-452d-9ade-dd6ce6303918\") " pod="openstack/glance-default-external-api-0" Nov 25 10:50:23 crc kubenswrapper[4821]: I1125 10:50:23.663744 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fada5a81-0157-452d-9ade-dd6ce6303918-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"fada5a81-0157-452d-9ade-dd6ce6303918\") " pod="openstack/glance-default-external-api-0" Nov 25 10:50:23 crc kubenswrapper[4821]: I1125 10:50:23.664875 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fada5a81-0157-452d-9ade-dd6ce6303918-scripts\") pod \"glance-default-external-api-0\" (UID: \"fada5a81-0157-452d-9ade-dd6ce6303918\") " pod="openstack/glance-default-external-api-0" Nov 25 10:50:23 crc kubenswrapper[4821]: I1125 10:50:23.670720 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fada5a81-0157-452d-9ade-dd6ce6303918-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"fada5a81-0157-452d-9ade-dd6ce6303918\") " pod="openstack/glance-default-external-api-0" Nov 25 10:50:23 crc kubenswrapper[4821]: I1125 10:50:23.678617 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fada5a81-0157-452d-9ade-dd6ce6303918-config-data\") pod \"glance-default-external-api-0\" (UID: \"fada5a81-0157-452d-9ade-dd6ce6303918\") " pod="openstack/glance-default-external-api-0" Nov 25 10:50:23 crc kubenswrapper[4821]: I1125 10:50:23.679932 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mklv5\" (UniqueName: \"kubernetes.io/projected/fada5a81-0157-452d-9ade-dd6ce6303918-kube-api-access-mklv5\") pod \"glance-default-external-api-0\" (UID: \"fada5a81-0157-452d-9ade-dd6ce6303918\") " pod="openstack/glance-default-external-api-0" Nov 25 10:50:23 crc kubenswrapper[4821]: I1125 10:50:23.713655 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"fada5a81-0157-452d-9ade-dd6ce6303918\") " pod="openstack/glance-default-external-api-0" Nov 25 10:50:23 crc kubenswrapper[4821]: I1125 10:50:23.777785 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 25 10:50:24 crc kubenswrapper[4821]: I1125 10:50:24.141581 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1647cf99-f6d6-421d-a00b-86538c7a7575" path="/var/lib/kubelet/pods/1647cf99-f6d6-421d-a00b-86538c7a7575/volumes" Nov 25 10:50:24 crc kubenswrapper[4821]: I1125 10:50:24.143363 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16fdcebd-383a-4945-a5bf-a299b332bf8f" path="/var/lib/kubelet/pods/16fdcebd-383a-4945-a5bf-a299b332bf8f/volumes" Nov 25 10:50:24 crc kubenswrapper[4821]: I1125 10:50:24.144975 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e08ffdf-0b33-475b-b040-c0ecda45ff57" path="/var/lib/kubelet/pods/2e08ffdf-0b33-475b-b040-c0ecda45ff57/volumes" Nov 25 10:50:24 crc kubenswrapper[4821]: I1125 10:50:24.145743 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="690339d6-9d90-46b3-b471-afabc2da9e7c" path="/var/lib/kubelet/pods/690339d6-9d90-46b3-b471-afabc2da9e7c/volumes" Nov 25 10:50:24 crc kubenswrapper[4821]: I1125 10:50:24.421397 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1816fd90-5883-4de2-9124-64e326d1743f","Type":"ContainerStarted","Data":"5e1432915cd207f32883814034073c67ee56aba0bb6acbca7d5f4f609d0ded60"} Nov 25 10:50:24 crc kubenswrapper[4821]: I1125 10:50:24.421533 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1816fd90-5883-4de2-9124-64e326d1743f","Type":"ContainerStarted","Data":"1072de5bda8efe7f8239255558afe31bede763e57c868b004d21a18dba6e9dc4"} Nov 25 10:50:24 crc kubenswrapper[4821]: I1125 10:50:24.424793 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"d969bbb6-564d-4563-98c6-5e2ec965c2dc","Type":"ContainerStarted","Data":"403facfebec747fb869dda0793fe75d7b6a82c224c8af320e58c49778207ebdc"} Nov 25 10:50:24 crc kubenswrapper[4821]: I1125 10:50:24.475650 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=6.000712691 podStartE2EDuration="19.475628577s" podCreationTimestamp="2025-11-25 10:50:05 +0000 UTC" firstStartedPulling="2025-11-25 10:50:07.699673781 +0000 UTC m=+1078.235993628" lastFinishedPulling="2025-11-25 10:50:21.174589667 +0000 UTC m=+1091.710909514" observedRunningTime="2025-11-25 10:50:24.460033718 +0000 UTC m=+1094.996353565" watchObservedRunningTime="2025-11-25 10:50:24.475628577 +0000 UTC m=+1095.011948434" Nov 25 10:50:24 crc kubenswrapper[4821]: I1125 10:50:24.484833 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 10:50:25 crc kubenswrapper[4821]: I1125 10:50:25.246431 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:50:25 crc kubenswrapper[4821]: I1125 10:50:25.437747 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1816fd90-5883-4de2-9124-64e326d1743f","Type":"ContainerStarted","Data":"02f9b313992de694d73b1d4c632ecf3896ee0490980cf4b3b908f161b7619440"} Nov 25 10:50:25 crc kubenswrapper[4821]: I1125 10:50:25.439840 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"fada5a81-0157-452d-9ade-dd6ce6303918","Type":"ContainerStarted","Data":"dc327d89ca2391f67e3f73a35037d2e57ff93e27da1a19dcfd0765eae1eef7ee"} Nov 25 10:50:25 crc kubenswrapper[4821]: I1125 10:50:25.441526 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bdacaaac-e3bf-4df9-8c19-e0e1315e5a44","Type":"ContainerStarted","Data":"ebde8cc019d2e6bc8b0f56dcc8c74918cc2fd155ea3c89261fce81c770361fff"} Nov 25 10:50:25 crc kubenswrapper[4821]: I1125 10:50:25.464590 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.46456325 podStartE2EDuration="3.46456325s" podCreationTimestamp="2025-11-25 10:50:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:50:25.459599262 +0000 UTC m=+1095.995919119" watchObservedRunningTime="2025-11-25 10:50:25.46456325 +0000 UTC m=+1096.000883097" Nov 25 10:50:26 crc kubenswrapper[4821]: E1125 10:50:26.107947 4821 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode2327ad3_5d88_4f4b_91ea_91deaf7fc249.slice/crio-6e8007ccc7348305c4f9b913d2dfa8dbf30b5f4417db7e25937dc070c2fa5264.scope\": RecentStats: unable to find data in memory cache]" Nov 25 10:50:26 crc kubenswrapper[4821]: I1125 10:50:26.445222 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-869c6f5bfb-l785w" Nov 25 10:50:26 crc kubenswrapper[4821]: I1125 10:50:26.452948 4821 generic.go:334] "Generic (PLEG): container finished" podID="e2327ad3-5d88-4f4b-91ea-91deaf7fc249" containerID="6e8007ccc7348305c4f9b913d2dfa8dbf30b5f4417db7e25937dc070c2fa5264" exitCode=137 Nov 25 10:50:26 crc kubenswrapper[4821]: I1125 10:50:26.452985 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-869c6f5bfb-l785w" event={"ID":"e2327ad3-5d88-4f4b-91ea-91deaf7fc249","Type":"ContainerDied","Data":"6e8007ccc7348305c4f9b913d2dfa8dbf30b5f4417db7e25937dc070c2fa5264"} Nov 25 10:50:26 crc kubenswrapper[4821]: I1125 10:50:26.453019 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-869c6f5bfb-l785w" event={"ID":"e2327ad3-5d88-4f4b-91ea-91deaf7fc249","Type":"ContainerDied","Data":"2b47a548244014b3290bfe8c4fd69cccaef99e1db0f5a41a31a5e9fc2a34c105"} Nov 25 10:50:26 crc kubenswrapper[4821]: I1125 10:50:26.453043 4821 scope.go:117] "RemoveContainer" containerID="0d6baf8077dd2e1ec034e0a15bb7d7957d591590cdfe3a7b244fa1f9081cdfcc" Nov 25 10:50:26 crc kubenswrapper[4821]: I1125 10:50:26.453072 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-869c6f5bfb-l785w" Nov 25 10:50:26 crc kubenswrapper[4821]: I1125 10:50:26.455489 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"fada5a81-0157-452d-9ade-dd6ce6303918","Type":"ContainerStarted","Data":"6fb7d12cbb6e0eb06de3cf3421859bf35a9f02f63b10148aa7ee65677a221000"} Nov 25 10:50:26 crc kubenswrapper[4821]: I1125 10:50:26.455606 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"fada5a81-0157-452d-9ade-dd6ce6303918","Type":"ContainerStarted","Data":"e3406750013ee13cb7a944be8f428226b86e7d3c95a0308a9f974b3a055918e5"} Nov 25 10:50:26 crc kubenswrapper[4821]: I1125 10:50:26.462124 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bdacaaac-e3bf-4df9-8c19-e0e1315e5a44","Type":"ContainerStarted","Data":"bff9fe05abde28189abb9c82700233bc4dd30325edad0e27869797b6920c6b54"} Nov 25 10:50:26 crc kubenswrapper[4821]: I1125 10:50:26.504456 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.504438688 podStartE2EDuration="3.504438688s" podCreationTimestamp="2025-11-25 10:50:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:50:26.500370193 +0000 UTC m=+1097.036690040" watchObservedRunningTime="2025-11-25 10:50:26.504438688 +0000 UTC m=+1097.040758535" Nov 25 10:50:26 crc kubenswrapper[4821]: I1125 10:50:26.627764 4821 scope.go:117] "RemoveContainer" containerID="6e8007ccc7348305c4f9b913d2dfa8dbf30b5f4417db7e25937dc070c2fa5264" Nov 25 10:50:26 crc kubenswrapper[4821]: I1125 10:50:26.630201 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e2327ad3-5d88-4f4b-91ea-91deaf7fc249-scripts\") pod \"e2327ad3-5d88-4f4b-91ea-91deaf7fc249\" (UID: \"e2327ad3-5d88-4f4b-91ea-91deaf7fc249\") " Nov 25 10:50:26 crc kubenswrapper[4821]: I1125 10:50:26.630285 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6tvvk\" (UniqueName: \"kubernetes.io/projected/e2327ad3-5d88-4f4b-91ea-91deaf7fc249-kube-api-access-6tvvk\") pod \"e2327ad3-5d88-4f4b-91ea-91deaf7fc249\" (UID: \"e2327ad3-5d88-4f4b-91ea-91deaf7fc249\") " Nov 25 10:50:26 crc kubenswrapper[4821]: I1125 10:50:26.630398 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2327ad3-5d88-4f4b-91ea-91deaf7fc249-combined-ca-bundle\") pod \"e2327ad3-5d88-4f4b-91ea-91deaf7fc249\" (UID: \"e2327ad3-5d88-4f4b-91ea-91deaf7fc249\") " Nov 25 10:50:26 crc kubenswrapper[4821]: I1125 10:50:26.630452 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/e2327ad3-5d88-4f4b-91ea-91deaf7fc249-horizon-tls-certs\") pod \"e2327ad3-5d88-4f4b-91ea-91deaf7fc249\" (UID: \"e2327ad3-5d88-4f4b-91ea-91deaf7fc249\") " Nov 25 10:50:26 crc kubenswrapper[4821]: I1125 10:50:26.630519 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e2327ad3-5d88-4f4b-91ea-91deaf7fc249-logs\") pod \"e2327ad3-5d88-4f4b-91ea-91deaf7fc249\" (UID: \"e2327ad3-5d88-4f4b-91ea-91deaf7fc249\") " Nov 25 10:50:26 crc kubenswrapper[4821]: I1125 10:50:26.630559 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e2327ad3-5d88-4f4b-91ea-91deaf7fc249-config-data\") pod \"e2327ad3-5d88-4f4b-91ea-91deaf7fc249\" (UID: \"e2327ad3-5d88-4f4b-91ea-91deaf7fc249\") " Nov 25 10:50:26 crc kubenswrapper[4821]: I1125 10:50:26.630669 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e2327ad3-5d88-4f4b-91ea-91deaf7fc249-horizon-secret-key\") pod \"e2327ad3-5d88-4f4b-91ea-91deaf7fc249\" (UID: \"e2327ad3-5d88-4f4b-91ea-91deaf7fc249\") " Nov 25 10:50:26 crc kubenswrapper[4821]: I1125 10:50:26.634794 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e2327ad3-5d88-4f4b-91ea-91deaf7fc249-logs" (OuterVolumeSpecName: "logs") pod "e2327ad3-5d88-4f4b-91ea-91deaf7fc249" (UID: "e2327ad3-5d88-4f4b-91ea-91deaf7fc249"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:50:26 crc kubenswrapper[4821]: I1125 10:50:26.638054 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2327ad3-5d88-4f4b-91ea-91deaf7fc249-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "e2327ad3-5d88-4f4b-91ea-91deaf7fc249" (UID: "e2327ad3-5d88-4f4b-91ea-91deaf7fc249"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:50:26 crc kubenswrapper[4821]: I1125 10:50:26.665739 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2327ad3-5d88-4f4b-91ea-91deaf7fc249-kube-api-access-6tvvk" (OuterVolumeSpecName: "kube-api-access-6tvvk") pod "e2327ad3-5d88-4f4b-91ea-91deaf7fc249" (UID: "e2327ad3-5d88-4f4b-91ea-91deaf7fc249"). InnerVolumeSpecName "kube-api-access-6tvvk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:50:26 crc kubenswrapper[4821]: I1125 10:50:26.668719 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e2327ad3-5d88-4f4b-91ea-91deaf7fc249-config-data" (OuterVolumeSpecName: "config-data") pod "e2327ad3-5d88-4f4b-91ea-91deaf7fc249" (UID: "e2327ad3-5d88-4f4b-91ea-91deaf7fc249"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:50:26 crc kubenswrapper[4821]: I1125 10:50:26.672822 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2327ad3-5d88-4f4b-91ea-91deaf7fc249-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e2327ad3-5d88-4f4b-91ea-91deaf7fc249" (UID: "e2327ad3-5d88-4f4b-91ea-91deaf7fc249"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:50:26 crc kubenswrapper[4821]: I1125 10:50:26.673976 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e2327ad3-5d88-4f4b-91ea-91deaf7fc249-scripts" (OuterVolumeSpecName: "scripts") pod "e2327ad3-5d88-4f4b-91ea-91deaf7fc249" (UID: "e2327ad3-5d88-4f4b-91ea-91deaf7fc249"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:50:26 crc kubenswrapper[4821]: I1125 10:50:26.699117 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2327ad3-5d88-4f4b-91ea-91deaf7fc249-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "e2327ad3-5d88-4f4b-91ea-91deaf7fc249" (UID: "e2327ad3-5d88-4f4b-91ea-91deaf7fc249"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:50:26 crc kubenswrapper[4821]: I1125 10:50:26.710529 4821 scope.go:117] "RemoveContainer" containerID="0d6baf8077dd2e1ec034e0a15bb7d7957d591590cdfe3a7b244fa1f9081cdfcc" Nov 25 10:50:26 crc kubenswrapper[4821]: E1125 10:50:26.711430 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d6baf8077dd2e1ec034e0a15bb7d7957d591590cdfe3a7b244fa1f9081cdfcc\": container with ID starting with 0d6baf8077dd2e1ec034e0a15bb7d7957d591590cdfe3a7b244fa1f9081cdfcc not found: ID does not exist" containerID="0d6baf8077dd2e1ec034e0a15bb7d7957d591590cdfe3a7b244fa1f9081cdfcc" Nov 25 10:50:26 crc kubenswrapper[4821]: I1125 10:50:26.711473 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d6baf8077dd2e1ec034e0a15bb7d7957d591590cdfe3a7b244fa1f9081cdfcc"} err="failed to get container status \"0d6baf8077dd2e1ec034e0a15bb7d7957d591590cdfe3a7b244fa1f9081cdfcc\": rpc error: code = NotFound desc = could not find container \"0d6baf8077dd2e1ec034e0a15bb7d7957d591590cdfe3a7b244fa1f9081cdfcc\": container with ID starting with 0d6baf8077dd2e1ec034e0a15bb7d7957d591590cdfe3a7b244fa1f9081cdfcc not found: ID does not exist" Nov 25 10:50:26 crc kubenswrapper[4821]: I1125 10:50:26.711496 4821 scope.go:117] "RemoveContainer" containerID="6e8007ccc7348305c4f9b913d2dfa8dbf30b5f4417db7e25937dc070c2fa5264" Nov 25 10:50:26 crc kubenswrapper[4821]: E1125 10:50:26.711896 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e8007ccc7348305c4f9b913d2dfa8dbf30b5f4417db7e25937dc070c2fa5264\": container with ID starting with 6e8007ccc7348305c4f9b913d2dfa8dbf30b5f4417db7e25937dc070c2fa5264 not found: ID does not exist" containerID="6e8007ccc7348305c4f9b913d2dfa8dbf30b5f4417db7e25937dc070c2fa5264" Nov 25 10:50:26 crc kubenswrapper[4821]: I1125 10:50:26.711919 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e8007ccc7348305c4f9b913d2dfa8dbf30b5f4417db7e25937dc070c2fa5264"} err="failed to get container status \"6e8007ccc7348305c4f9b913d2dfa8dbf30b5f4417db7e25937dc070c2fa5264\": rpc error: code = NotFound desc = could not find container \"6e8007ccc7348305c4f9b913d2dfa8dbf30b5f4417db7e25937dc070c2fa5264\": container with ID starting with 6e8007ccc7348305c4f9b913d2dfa8dbf30b5f4417db7e25937dc070c2fa5264 not found: ID does not exist" Nov 25 10:50:26 crc kubenswrapper[4821]: I1125 10:50:26.733874 4821 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e2327ad3-5d88-4f4b-91ea-91deaf7fc249-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:26 crc kubenswrapper[4821]: I1125 10:50:26.733913 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6tvvk\" (UniqueName: \"kubernetes.io/projected/e2327ad3-5d88-4f4b-91ea-91deaf7fc249-kube-api-access-6tvvk\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:26 crc kubenswrapper[4821]: I1125 10:50:26.733928 4821 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2327ad3-5d88-4f4b-91ea-91deaf7fc249-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:26 crc kubenswrapper[4821]: I1125 10:50:26.733940 4821 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/e2327ad3-5d88-4f4b-91ea-91deaf7fc249-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:26 crc kubenswrapper[4821]: I1125 10:50:26.733951 4821 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e2327ad3-5d88-4f4b-91ea-91deaf7fc249-logs\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:26 crc kubenswrapper[4821]: I1125 10:50:26.733961 4821 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e2327ad3-5d88-4f4b-91ea-91deaf7fc249-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:26 crc kubenswrapper[4821]: I1125 10:50:26.733972 4821 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e2327ad3-5d88-4f4b-91ea-91deaf7fc249-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:26 crc kubenswrapper[4821]: I1125 10:50:26.786864 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-869c6f5bfb-l785w"] Nov 25 10:50:26 crc kubenswrapper[4821]: I1125 10:50:26.796017 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-869c6f5bfb-l785w"] Nov 25 10:50:27 crc kubenswrapper[4821]: I1125 10:50:27.473249 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bdacaaac-e3bf-4df9-8c19-e0e1315e5a44","Type":"ContainerStarted","Data":"8515d92114dfb916f0a9bde20543a637bef3e145d0806db5afaf101215ea0308"} Nov 25 10:50:28 crc kubenswrapper[4821]: I1125 10:50:28.126604 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e2327ad3-5d88-4f4b-91ea-91deaf7fc249" path="/var/lib/kubelet/pods/e2327ad3-5d88-4f4b-91ea-91deaf7fc249/volumes" Nov 25 10:50:29 crc kubenswrapper[4821]: I1125 10:50:29.496171 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bdacaaac-e3bf-4df9-8c19-e0e1315e5a44","Type":"ContainerStarted","Data":"2c1795eb47371ee16e40666571a2d9cae9bf0c2bbb9394e5cdcf4394f9d7956d"} Nov 25 10:50:32 crc kubenswrapper[4821]: I1125 10:50:32.911516 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 25 10:50:32 crc kubenswrapper[4821]: I1125 10:50:32.911812 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 25 10:50:32 crc kubenswrapper[4821]: I1125 10:50:32.940585 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 25 10:50:32 crc kubenswrapper[4821]: I1125 10:50:32.951462 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 25 10:50:33 crc kubenswrapper[4821]: I1125 10:50:33.527368 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 25 10:50:33 crc kubenswrapper[4821]: I1125 10:50:33.527517 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 25 10:50:33 crc kubenswrapper[4821]: I1125 10:50:33.778649 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 25 10:50:33 crc kubenswrapper[4821]: I1125 10:50:33.779628 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 25 10:50:33 crc kubenswrapper[4821]: I1125 10:50:33.813941 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 25 10:50:33 crc kubenswrapper[4821]: I1125 10:50:33.838044 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 25 10:50:34 crc kubenswrapper[4821]: I1125 10:50:34.535299 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 25 10:50:34 crc kubenswrapper[4821]: I1125 10:50:34.536261 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 25 10:50:35 crc kubenswrapper[4821]: I1125 10:50:35.530358 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 25 10:50:35 crc kubenswrapper[4821]: I1125 10:50:35.545199 4821 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 10:50:35 crc kubenswrapper[4821]: I1125 10:50:35.605089 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 25 10:50:36 crc kubenswrapper[4821]: I1125 10:50:36.448051 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 25 10:50:36 crc kubenswrapper[4821]: I1125 10:50:36.459990 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 25 10:50:38 crc kubenswrapper[4821]: I1125 10:50:38.581429 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bdacaaac-e3bf-4df9-8c19-e0e1315e5a44" containerName="ceilometer-central-agent" containerID="cri-o://ebde8cc019d2e6bc8b0f56dcc8c74918cc2fd155ea3c89261fce81c770361fff" gracePeriod=30 Nov 25 10:50:38 crc kubenswrapper[4821]: I1125 10:50:38.581551 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bdacaaac-e3bf-4df9-8c19-e0e1315e5a44" containerName="sg-core" containerID="cri-o://8515d92114dfb916f0a9bde20543a637bef3e145d0806db5afaf101215ea0308" gracePeriod=30 Nov 25 10:50:38 crc kubenswrapper[4821]: I1125 10:50:38.581552 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bdacaaac-e3bf-4df9-8c19-e0e1315e5a44" containerName="ceilometer-notification-agent" containerID="cri-o://bff9fe05abde28189abb9c82700233bc4dd30325edad0e27869797b6920c6b54" gracePeriod=30 Nov 25 10:50:38 crc kubenswrapper[4821]: I1125 10:50:38.581572 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 25 10:50:38 crc kubenswrapper[4821]: I1125 10:50:38.581498 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bdacaaac-e3bf-4df9-8c19-e0e1315e5a44" containerName="proxy-httpd" containerID="cri-o://2c1795eb47371ee16e40666571a2d9cae9bf0c2bbb9394e5cdcf4394f9d7956d" gracePeriod=30 Nov 25 10:50:38 crc kubenswrapper[4821]: I1125 10:50:38.606197 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=11.400296222 podStartE2EDuration="16.606134796s" podCreationTimestamp="2025-11-25 10:50:22 +0000 UTC" firstStartedPulling="2025-11-25 10:50:23.317297145 +0000 UTC m=+1093.853616992" lastFinishedPulling="2025-11-25 10:50:28.523135719 +0000 UTC m=+1099.059455566" observedRunningTime="2025-11-25 10:50:38.602914484 +0000 UTC m=+1109.139234331" watchObservedRunningTime="2025-11-25 10:50:38.606134796 +0000 UTC m=+1109.142454643" Nov 25 10:50:39 crc kubenswrapper[4821]: I1125 10:50:39.595219 4821 generic.go:334] "Generic (PLEG): container finished" podID="bdacaaac-e3bf-4df9-8c19-e0e1315e5a44" containerID="2c1795eb47371ee16e40666571a2d9cae9bf0c2bbb9394e5cdcf4394f9d7956d" exitCode=0 Nov 25 10:50:39 crc kubenswrapper[4821]: I1125 10:50:39.595527 4821 generic.go:334] "Generic (PLEG): container finished" podID="bdacaaac-e3bf-4df9-8c19-e0e1315e5a44" containerID="8515d92114dfb916f0a9bde20543a637bef3e145d0806db5afaf101215ea0308" exitCode=2 Nov 25 10:50:39 crc kubenswrapper[4821]: I1125 10:50:39.595539 4821 generic.go:334] "Generic (PLEG): container finished" podID="bdacaaac-e3bf-4df9-8c19-e0e1315e5a44" containerID="bff9fe05abde28189abb9c82700233bc4dd30325edad0e27869797b6920c6b54" exitCode=0 Nov 25 10:50:39 crc kubenswrapper[4821]: I1125 10:50:39.595547 4821 generic.go:334] "Generic (PLEG): container finished" podID="bdacaaac-e3bf-4df9-8c19-e0e1315e5a44" containerID="ebde8cc019d2e6bc8b0f56dcc8c74918cc2fd155ea3c89261fce81c770361fff" exitCode=0 Nov 25 10:50:39 crc kubenswrapper[4821]: I1125 10:50:39.595265 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bdacaaac-e3bf-4df9-8c19-e0e1315e5a44","Type":"ContainerDied","Data":"2c1795eb47371ee16e40666571a2d9cae9bf0c2bbb9394e5cdcf4394f9d7956d"} Nov 25 10:50:39 crc kubenswrapper[4821]: I1125 10:50:39.595581 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bdacaaac-e3bf-4df9-8c19-e0e1315e5a44","Type":"ContainerDied","Data":"8515d92114dfb916f0a9bde20543a637bef3e145d0806db5afaf101215ea0308"} Nov 25 10:50:39 crc kubenswrapper[4821]: I1125 10:50:39.595594 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bdacaaac-e3bf-4df9-8c19-e0e1315e5a44","Type":"ContainerDied","Data":"bff9fe05abde28189abb9c82700233bc4dd30325edad0e27869797b6920c6b54"} Nov 25 10:50:39 crc kubenswrapper[4821]: I1125 10:50:39.595605 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bdacaaac-e3bf-4df9-8c19-e0e1315e5a44","Type":"ContainerDied","Data":"ebde8cc019d2e6bc8b0f56dcc8c74918cc2fd155ea3c89261fce81c770361fff"} Nov 25 10:50:39 crc kubenswrapper[4821]: I1125 10:50:39.647963 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:50:39 crc kubenswrapper[4821]: I1125 10:50:39.689050 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdacaaac-e3bf-4df9-8c19-e0e1315e5a44-combined-ca-bundle\") pod \"bdacaaac-e3bf-4df9-8c19-e0e1315e5a44\" (UID: \"bdacaaac-e3bf-4df9-8c19-e0e1315e5a44\") " Nov 25 10:50:39 crc kubenswrapper[4821]: I1125 10:50:39.689124 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bdacaaac-e3bf-4df9-8c19-e0e1315e5a44-scripts\") pod \"bdacaaac-e3bf-4df9-8c19-e0e1315e5a44\" (UID: \"bdacaaac-e3bf-4df9-8c19-e0e1315e5a44\") " Nov 25 10:50:39 crc kubenswrapper[4821]: I1125 10:50:39.689203 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdacaaac-e3bf-4df9-8c19-e0e1315e5a44-config-data\") pod \"bdacaaac-e3bf-4df9-8c19-e0e1315e5a44\" (UID: \"bdacaaac-e3bf-4df9-8c19-e0e1315e5a44\") " Nov 25 10:50:39 crc kubenswrapper[4821]: I1125 10:50:39.689222 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bdacaaac-e3bf-4df9-8c19-e0e1315e5a44-sg-core-conf-yaml\") pod \"bdacaaac-e3bf-4df9-8c19-e0e1315e5a44\" (UID: \"bdacaaac-e3bf-4df9-8c19-e0e1315e5a44\") " Nov 25 10:50:39 crc kubenswrapper[4821]: I1125 10:50:39.689240 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-59lxj\" (UniqueName: \"kubernetes.io/projected/bdacaaac-e3bf-4df9-8c19-e0e1315e5a44-kube-api-access-59lxj\") pod \"bdacaaac-e3bf-4df9-8c19-e0e1315e5a44\" (UID: \"bdacaaac-e3bf-4df9-8c19-e0e1315e5a44\") " Nov 25 10:50:39 crc kubenswrapper[4821]: I1125 10:50:39.689285 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bdacaaac-e3bf-4df9-8c19-e0e1315e5a44-run-httpd\") pod \"bdacaaac-e3bf-4df9-8c19-e0e1315e5a44\" (UID: \"bdacaaac-e3bf-4df9-8c19-e0e1315e5a44\") " Nov 25 10:50:39 crc kubenswrapper[4821]: I1125 10:50:39.689343 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bdacaaac-e3bf-4df9-8c19-e0e1315e5a44-log-httpd\") pod \"bdacaaac-e3bf-4df9-8c19-e0e1315e5a44\" (UID: \"bdacaaac-e3bf-4df9-8c19-e0e1315e5a44\") " Nov 25 10:50:39 crc kubenswrapper[4821]: I1125 10:50:39.692677 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bdacaaac-e3bf-4df9-8c19-e0e1315e5a44-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "bdacaaac-e3bf-4df9-8c19-e0e1315e5a44" (UID: "bdacaaac-e3bf-4df9-8c19-e0e1315e5a44"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:50:39 crc kubenswrapper[4821]: I1125 10:50:39.695665 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bdacaaac-e3bf-4df9-8c19-e0e1315e5a44-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "bdacaaac-e3bf-4df9-8c19-e0e1315e5a44" (UID: "bdacaaac-e3bf-4df9-8c19-e0e1315e5a44"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:50:39 crc kubenswrapper[4821]: I1125 10:50:39.698683 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bdacaaac-e3bf-4df9-8c19-e0e1315e5a44-kube-api-access-59lxj" (OuterVolumeSpecName: "kube-api-access-59lxj") pod "bdacaaac-e3bf-4df9-8c19-e0e1315e5a44" (UID: "bdacaaac-e3bf-4df9-8c19-e0e1315e5a44"). InnerVolumeSpecName "kube-api-access-59lxj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:50:39 crc kubenswrapper[4821]: I1125 10:50:39.703049 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdacaaac-e3bf-4df9-8c19-e0e1315e5a44-scripts" (OuterVolumeSpecName: "scripts") pod "bdacaaac-e3bf-4df9-8c19-e0e1315e5a44" (UID: "bdacaaac-e3bf-4df9-8c19-e0e1315e5a44"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:50:39 crc kubenswrapper[4821]: I1125 10:50:39.744086 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdacaaac-e3bf-4df9-8c19-e0e1315e5a44-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "bdacaaac-e3bf-4df9-8c19-e0e1315e5a44" (UID: "bdacaaac-e3bf-4df9-8c19-e0e1315e5a44"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:50:39 crc kubenswrapper[4821]: I1125 10:50:39.791779 4821 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bdacaaac-e3bf-4df9-8c19-e0e1315e5a44-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:39 crc kubenswrapper[4821]: I1125 10:50:39.791812 4821 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bdacaaac-e3bf-4df9-8c19-e0e1315e5a44-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:39 crc kubenswrapper[4821]: I1125 10:50:39.791822 4821 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bdacaaac-e3bf-4df9-8c19-e0e1315e5a44-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:39 crc kubenswrapper[4821]: I1125 10:50:39.791831 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-59lxj\" (UniqueName: \"kubernetes.io/projected/bdacaaac-e3bf-4df9-8c19-e0e1315e5a44-kube-api-access-59lxj\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:39 crc kubenswrapper[4821]: I1125 10:50:39.791839 4821 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bdacaaac-e3bf-4df9-8c19-e0e1315e5a44-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:39 crc kubenswrapper[4821]: I1125 10:50:39.796920 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdacaaac-e3bf-4df9-8c19-e0e1315e5a44-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bdacaaac-e3bf-4df9-8c19-e0e1315e5a44" (UID: "bdacaaac-e3bf-4df9-8c19-e0e1315e5a44"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:50:39 crc kubenswrapper[4821]: I1125 10:50:39.815064 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdacaaac-e3bf-4df9-8c19-e0e1315e5a44-config-data" (OuterVolumeSpecName: "config-data") pod "bdacaaac-e3bf-4df9-8c19-e0e1315e5a44" (UID: "bdacaaac-e3bf-4df9-8c19-e0e1315e5a44"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:50:39 crc kubenswrapper[4821]: I1125 10:50:39.893714 4821 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdacaaac-e3bf-4df9-8c19-e0e1315e5a44-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:39 crc kubenswrapper[4821]: I1125 10:50:39.893747 4821 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdacaaac-e3bf-4df9-8c19-e0e1315e5a44-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:40 crc kubenswrapper[4821]: I1125 10:50:40.605971 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bdacaaac-e3bf-4df9-8c19-e0e1315e5a44","Type":"ContainerDied","Data":"f27722b39dc103c305a9fc6f15526c02c0796801f8c195a5e1249dba6a0017d6"} Nov 25 10:50:40 crc kubenswrapper[4821]: I1125 10:50:40.606024 4821 scope.go:117] "RemoveContainer" containerID="2c1795eb47371ee16e40666571a2d9cae9bf0c2bbb9394e5cdcf4394f9d7956d" Nov 25 10:50:40 crc kubenswrapper[4821]: I1125 10:50:40.606047 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:50:40 crc kubenswrapper[4821]: I1125 10:50:40.631528 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:50:40 crc kubenswrapper[4821]: I1125 10:50:40.636382 4821 scope.go:117] "RemoveContainer" containerID="8515d92114dfb916f0a9bde20543a637bef3e145d0806db5afaf101215ea0308" Nov 25 10:50:40 crc kubenswrapper[4821]: I1125 10:50:40.638400 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:50:40 crc kubenswrapper[4821]: I1125 10:50:40.670024 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:50:40 crc kubenswrapper[4821]: E1125 10:50:40.670887 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdacaaac-e3bf-4df9-8c19-e0e1315e5a44" containerName="proxy-httpd" Nov 25 10:50:40 crc kubenswrapper[4821]: I1125 10:50:40.671090 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdacaaac-e3bf-4df9-8c19-e0e1315e5a44" containerName="proxy-httpd" Nov 25 10:50:40 crc kubenswrapper[4821]: E1125 10:50:40.671232 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2327ad3-5d88-4f4b-91ea-91deaf7fc249" containerName="horizon" Nov 25 10:50:40 crc kubenswrapper[4821]: I1125 10:50:40.671432 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2327ad3-5d88-4f4b-91ea-91deaf7fc249" containerName="horizon" Nov 25 10:50:40 crc kubenswrapper[4821]: E1125 10:50:40.671543 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2327ad3-5d88-4f4b-91ea-91deaf7fc249" containerName="horizon-log" Nov 25 10:50:40 crc kubenswrapper[4821]: I1125 10:50:40.671639 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2327ad3-5d88-4f4b-91ea-91deaf7fc249" containerName="horizon-log" Nov 25 10:50:40 crc kubenswrapper[4821]: E1125 10:50:40.671732 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdacaaac-e3bf-4df9-8c19-e0e1315e5a44" containerName="ceilometer-central-agent" Nov 25 10:50:40 crc kubenswrapper[4821]: I1125 10:50:40.671857 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdacaaac-e3bf-4df9-8c19-e0e1315e5a44" containerName="ceilometer-central-agent" Nov 25 10:50:40 crc kubenswrapper[4821]: E1125 10:50:40.671962 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdacaaac-e3bf-4df9-8c19-e0e1315e5a44" containerName="sg-core" Nov 25 10:50:40 crc kubenswrapper[4821]: I1125 10:50:40.672054 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdacaaac-e3bf-4df9-8c19-e0e1315e5a44" containerName="sg-core" Nov 25 10:50:40 crc kubenswrapper[4821]: E1125 10:50:40.672203 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdacaaac-e3bf-4df9-8c19-e0e1315e5a44" containerName="ceilometer-notification-agent" Nov 25 10:50:40 crc kubenswrapper[4821]: I1125 10:50:40.672299 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdacaaac-e3bf-4df9-8c19-e0e1315e5a44" containerName="ceilometer-notification-agent" Nov 25 10:50:40 crc kubenswrapper[4821]: I1125 10:50:40.672724 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="bdacaaac-e3bf-4df9-8c19-e0e1315e5a44" containerName="ceilometer-notification-agent" Nov 25 10:50:40 crc kubenswrapper[4821]: I1125 10:50:40.672869 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="bdacaaac-e3bf-4df9-8c19-e0e1315e5a44" containerName="ceilometer-central-agent" Nov 25 10:50:40 crc kubenswrapper[4821]: I1125 10:50:40.672982 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2327ad3-5d88-4f4b-91ea-91deaf7fc249" containerName="horizon-log" Nov 25 10:50:40 crc kubenswrapper[4821]: I1125 10:50:40.673104 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="bdacaaac-e3bf-4df9-8c19-e0e1315e5a44" containerName="proxy-httpd" Nov 25 10:50:40 crc kubenswrapper[4821]: I1125 10:50:40.675988 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="bdacaaac-e3bf-4df9-8c19-e0e1315e5a44" containerName="sg-core" Nov 25 10:50:40 crc kubenswrapper[4821]: I1125 10:50:40.676818 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2327ad3-5d88-4f4b-91ea-91deaf7fc249" containerName="horizon" Nov 25 10:50:40 crc kubenswrapper[4821]: I1125 10:50:40.679022 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:50:40 crc kubenswrapper[4821]: I1125 10:50:40.679628 4821 scope.go:117] "RemoveContainer" containerID="bff9fe05abde28189abb9c82700233bc4dd30325edad0e27869797b6920c6b54" Nov 25 10:50:40 crc kubenswrapper[4821]: I1125 10:50:40.681191 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 25 10:50:40 crc kubenswrapper[4821]: I1125 10:50:40.681445 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 25 10:50:40 crc kubenswrapper[4821]: I1125 10:50:40.688106 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:50:40 crc kubenswrapper[4821]: I1125 10:50:40.705922 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/420e728b-b00c-4075-b504-7f87402c4838-config-data\") pod \"ceilometer-0\" (UID: \"420e728b-b00c-4075-b504-7f87402c4838\") " pod="openstack/ceilometer-0" Nov 25 10:50:40 crc kubenswrapper[4821]: I1125 10:50:40.706003 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/420e728b-b00c-4075-b504-7f87402c4838-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"420e728b-b00c-4075-b504-7f87402c4838\") " pod="openstack/ceilometer-0" Nov 25 10:50:40 crc kubenswrapper[4821]: I1125 10:50:40.706036 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/420e728b-b00c-4075-b504-7f87402c4838-scripts\") pod \"ceilometer-0\" (UID: \"420e728b-b00c-4075-b504-7f87402c4838\") " pod="openstack/ceilometer-0" Nov 25 10:50:40 crc kubenswrapper[4821]: I1125 10:50:40.706103 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wx2sh\" (UniqueName: \"kubernetes.io/projected/420e728b-b00c-4075-b504-7f87402c4838-kube-api-access-wx2sh\") pod \"ceilometer-0\" (UID: \"420e728b-b00c-4075-b504-7f87402c4838\") " pod="openstack/ceilometer-0" Nov 25 10:50:40 crc kubenswrapper[4821]: I1125 10:50:40.706143 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/420e728b-b00c-4075-b504-7f87402c4838-run-httpd\") pod \"ceilometer-0\" (UID: \"420e728b-b00c-4075-b504-7f87402c4838\") " pod="openstack/ceilometer-0" Nov 25 10:50:40 crc kubenswrapper[4821]: I1125 10:50:40.706210 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/420e728b-b00c-4075-b504-7f87402c4838-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"420e728b-b00c-4075-b504-7f87402c4838\") " pod="openstack/ceilometer-0" Nov 25 10:50:40 crc kubenswrapper[4821]: I1125 10:50:40.706237 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/420e728b-b00c-4075-b504-7f87402c4838-log-httpd\") pod \"ceilometer-0\" (UID: \"420e728b-b00c-4075-b504-7f87402c4838\") " pod="openstack/ceilometer-0" Nov 25 10:50:40 crc kubenswrapper[4821]: I1125 10:50:40.709082 4821 scope.go:117] "RemoveContainer" containerID="ebde8cc019d2e6bc8b0f56dcc8c74918cc2fd155ea3c89261fce81c770361fff" Nov 25 10:50:40 crc kubenswrapper[4821]: I1125 10:50:40.807455 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/420e728b-b00c-4075-b504-7f87402c4838-config-data\") pod \"ceilometer-0\" (UID: \"420e728b-b00c-4075-b504-7f87402c4838\") " pod="openstack/ceilometer-0" Nov 25 10:50:40 crc kubenswrapper[4821]: I1125 10:50:40.807835 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/420e728b-b00c-4075-b504-7f87402c4838-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"420e728b-b00c-4075-b504-7f87402c4838\") " pod="openstack/ceilometer-0" Nov 25 10:50:40 crc kubenswrapper[4821]: I1125 10:50:40.807874 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/420e728b-b00c-4075-b504-7f87402c4838-scripts\") pod \"ceilometer-0\" (UID: \"420e728b-b00c-4075-b504-7f87402c4838\") " pod="openstack/ceilometer-0" Nov 25 10:50:40 crc kubenswrapper[4821]: I1125 10:50:40.807920 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wx2sh\" (UniqueName: \"kubernetes.io/projected/420e728b-b00c-4075-b504-7f87402c4838-kube-api-access-wx2sh\") pod \"ceilometer-0\" (UID: \"420e728b-b00c-4075-b504-7f87402c4838\") " pod="openstack/ceilometer-0" Nov 25 10:50:40 crc kubenswrapper[4821]: I1125 10:50:40.807971 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/420e728b-b00c-4075-b504-7f87402c4838-run-httpd\") pod \"ceilometer-0\" (UID: \"420e728b-b00c-4075-b504-7f87402c4838\") " pod="openstack/ceilometer-0" Nov 25 10:50:40 crc kubenswrapper[4821]: I1125 10:50:40.808018 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/420e728b-b00c-4075-b504-7f87402c4838-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"420e728b-b00c-4075-b504-7f87402c4838\") " pod="openstack/ceilometer-0" Nov 25 10:50:40 crc kubenswrapper[4821]: I1125 10:50:40.808040 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/420e728b-b00c-4075-b504-7f87402c4838-log-httpd\") pod \"ceilometer-0\" (UID: \"420e728b-b00c-4075-b504-7f87402c4838\") " pod="openstack/ceilometer-0" Nov 25 10:50:40 crc kubenswrapper[4821]: I1125 10:50:40.808540 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/420e728b-b00c-4075-b504-7f87402c4838-log-httpd\") pod \"ceilometer-0\" (UID: \"420e728b-b00c-4075-b504-7f87402c4838\") " pod="openstack/ceilometer-0" Nov 25 10:50:40 crc kubenswrapper[4821]: I1125 10:50:40.808606 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/420e728b-b00c-4075-b504-7f87402c4838-run-httpd\") pod \"ceilometer-0\" (UID: \"420e728b-b00c-4075-b504-7f87402c4838\") " pod="openstack/ceilometer-0" Nov 25 10:50:40 crc kubenswrapper[4821]: I1125 10:50:40.822843 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/420e728b-b00c-4075-b504-7f87402c4838-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"420e728b-b00c-4075-b504-7f87402c4838\") " pod="openstack/ceilometer-0" Nov 25 10:50:40 crc kubenswrapper[4821]: I1125 10:50:40.822937 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/420e728b-b00c-4075-b504-7f87402c4838-scripts\") pod \"ceilometer-0\" (UID: \"420e728b-b00c-4075-b504-7f87402c4838\") " pod="openstack/ceilometer-0" Nov 25 10:50:40 crc kubenswrapper[4821]: I1125 10:50:40.823179 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/420e728b-b00c-4075-b504-7f87402c4838-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"420e728b-b00c-4075-b504-7f87402c4838\") " pod="openstack/ceilometer-0" Nov 25 10:50:40 crc kubenswrapper[4821]: I1125 10:50:40.823782 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/420e728b-b00c-4075-b504-7f87402c4838-config-data\") pod \"ceilometer-0\" (UID: \"420e728b-b00c-4075-b504-7f87402c4838\") " pod="openstack/ceilometer-0" Nov 25 10:50:40 crc kubenswrapper[4821]: I1125 10:50:40.826311 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wx2sh\" (UniqueName: \"kubernetes.io/projected/420e728b-b00c-4075-b504-7f87402c4838-kube-api-access-wx2sh\") pod \"ceilometer-0\" (UID: \"420e728b-b00c-4075-b504-7f87402c4838\") " pod="openstack/ceilometer-0" Nov 25 10:50:41 crc kubenswrapper[4821]: I1125 10:50:41.010487 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:50:41 crc kubenswrapper[4821]: I1125 10:50:41.350927 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:50:41 crc kubenswrapper[4821]: I1125 10:50:41.475324 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:50:41 crc kubenswrapper[4821]: W1125 10:50:41.477696 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod420e728b_b00c_4075_b504_7f87402c4838.slice/crio-5e4b32c9290d6f7ac99276e5ad69ff7a3df810b2f477595047fdca4c25896447 WatchSource:0}: Error finding container 5e4b32c9290d6f7ac99276e5ad69ff7a3df810b2f477595047fdca4c25896447: Status 404 returned error can't find the container with id 5e4b32c9290d6f7ac99276e5ad69ff7a3df810b2f477595047fdca4c25896447 Nov 25 10:50:41 crc kubenswrapper[4821]: I1125 10:50:41.615791 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"420e728b-b00c-4075-b504-7f87402c4838","Type":"ContainerStarted","Data":"5e4b32c9290d6f7ac99276e5ad69ff7a3df810b2f477595047fdca4c25896447"} Nov 25 10:50:42 crc kubenswrapper[4821]: I1125 10:50:42.129809 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bdacaaac-e3bf-4df9-8c19-e0e1315e5a44" path="/var/lib/kubelet/pods/bdacaaac-e3bf-4df9-8c19-e0e1315e5a44/volumes" Nov 25 10:50:42 crc kubenswrapper[4821]: I1125 10:50:42.625301 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"420e728b-b00c-4075-b504-7f87402c4838","Type":"ContainerStarted","Data":"774c8bc29c1b4d2314e1ee4be66ba95a1f8f1661f26140696fe62e78d53de57d"} Nov 25 10:50:43 crc kubenswrapper[4821]: I1125 10:50:43.636731 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"420e728b-b00c-4075-b504-7f87402c4838","Type":"ContainerStarted","Data":"363e08aee0b8c87fbe312f4abd6aab12624f6aab82f381b708af2b15ba064830"} Nov 25 10:50:44 crc kubenswrapper[4821]: I1125 10:50:44.636240 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-57l9j"] Nov 25 10:50:44 crc kubenswrapper[4821]: I1125 10:50:44.637300 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-57l9j" Nov 25 10:50:44 crc kubenswrapper[4821]: I1125 10:50:44.655813 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-eaac-account-create-f4xs9"] Nov 25 10:50:44 crc kubenswrapper[4821]: I1125 10:50:44.656978 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-eaac-account-create-f4xs9" Nov 25 10:50:44 crc kubenswrapper[4821]: I1125 10:50:44.665088 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Nov 25 10:50:44 crc kubenswrapper[4821]: I1125 10:50:44.671177 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-57l9j"] Nov 25 10:50:44 crc kubenswrapper[4821]: I1125 10:50:44.683988 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-eaac-account-create-f4xs9"] Nov 25 10:50:44 crc kubenswrapper[4821]: I1125 10:50:44.747079 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-5xnrf"] Nov 25 10:50:44 crc kubenswrapper[4821]: I1125 10:50:44.748675 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-5xnrf" Nov 25 10:50:44 crc kubenswrapper[4821]: I1125 10:50:44.756822 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-5xnrf"] Nov 25 10:50:44 crc kubenswrapper[4821]: I1125 10:50:44.783439 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnp2s\" (UniqueName: \"kubernetes.io/projected/ec0ad62e-84ad-44d8-a8f2-264d374ed70f-kube-api-access-fnp2s\") pod \"nova-api-db-create-57l9j\" (UID: \"ec0ad62e-84ad-44d8-a8f2-264d374ed70f\") " pod="openstack/nova-api-db-create-57l9j" Nov 25 10:50:44 crc kubenswrapper[4821]: I1125 10:50:44.783552 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7wxvt\" (UniqueName: \"kubernetes.io/projected/b2a4919e-4ad6-4cdc-b1cb-5a0200dcc8cc-kube-api-access-7wxvt\") pod \"nova-api-eaac-account-create-f4xs9\" (UID: \"b2a4919e-4ad6-4cdc-b1cb-5a0200dcc8cc\") " pod="openstack/nova-api-eaac-account-create-f4xs9" Nov 25 10:50:44 crc kubenswrapper[4821]: I1125 10:50:44.783599 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ec0ad62e-84ad-44d8-a8f2-264d374ed70f-operator-scripts\") pod \"nova-api-db-create-57l9j\" (UID: \"ec0ad62e-84ad-44d8-a8f2-264d374ed70f\") " pod="openstack/nova-api-db-create-57l9j" Nov 25 10:50:44 crc kubenswrapper[4821]: I1125 10:50:44.783660 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b2a4919e-4ad6-4cdc-b1cb-5a0200dcc8cc-operator-scripts\") pod \"nova-api-eaac-account-create-f4xs9\" (UID: \"b2a4919e-4ad6-4cdc-b1cb-5a0200dcc8cc\") " pod="openstack/nova-api-eaac-account-create-f4xs9" Nov 25 10:50:44 crc kubenswrapper[4821]: I1125 10:50:44.849337 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-4n4fc"] Nov 25 10:50:44 crc kubenswrapper[4821]: I1125 10:50:44.851266 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-4n4fc" Nov 25 10:50:44 crc kubenswrapper[4821]: I1125 10:50:44.860228 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-30f9-account-create-szdhp"] Nov 25 10:50:44 crc kubenswrapper[4821]: I1125 10:50:44.861712 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-30f9-account-create-szdhp" Nov 25 10:50:44 crc kubenswrapper[4821]: I1125 10:50:44.868729 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Nov 25 10:50:44 crc kubenswrapper[4821]: I1125 10:50:44.875278 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-4n4fc"] Nov 25 10:50:44 crc kubenswrapper[4821]: I1125 10:50:44.885184 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7wxvt\" (UniqueName: \"kubernetes.io/projected/b2a4919e-4ad6-4cdc-b1cb-5a0200dcc8cc-kube-api-access-7wxvt\") pod \"nova-api-eaac-account-create-f4xs9\" (UID: \"b2a4919e-4ad6-4cdc-b1cb-5a0200dcc8cc\") " pod="openstack/nova-api-eaac-account-create-f4xs9" Nov 25 10:50:44 crc kubenswrapper[4821]: I1125 10:50:44.885248 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ec0ad62e-84ad-44d8-a8f2-264d374ed70f-operator-scripts\") pod \"nova-api-db-create-57l9j\" (UID: \"ec0ad62e-84ad-44d8-a8f2-264d374ed70f\") " pod="openstack/nova-api-db-create-57l9j" Nov 25 10:50:44 crc kubenswrapper[4821]: I1125 10:50:44.885299 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7a0917a6-232d-43e0-8c9d-44eb9485d0d8-operator-scripts\") pod \"nova-cell0-db-create-5xnrf\" (UID: \"7a0917a6-232d-43e0-8c9d-44eb9485d0d8\") " pod="openstack/nova-cell0-db-create-5xnrf" Nov 25 10:50:44 crc kubenswrapper[4821]: I1125 10:50:44.885321 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b2a4919e-4ad6-4cdc-b1cb-5a0200dcc8cc-operator-scripts\") pod \"nova-api-eaac-account-create-f4xs9\" (UID: \"b2a4919e-4ad6-4cdc-b1cb-5a0200dcc8cc\") " pod="openstack/nova-api-eaac-account-create-f4xs9" Nov 25 10:50:44 crc kubenswrapper[4821]: I1125 10:50:44.885352 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8k9jm\" (UniqueName: \"kubernetes.io/projected/7a0917a6-232d-43e0-8c9d-44eb9485d0d8-kube-api-access-8k9jm\") pod \"nova-cell0-db-create-5xnrf\" (UID: \"7a0917a6-232d-43e0-8c9d-44eb9485d0d8\") " pod="openstack/nova-cell0-db-create-5xnrf" Nov 25 10:50:44 crc kubenswrapper[4821]: I1125 10:50:44.885403 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fnp2s\" (UniqueName: \"kubernetes.io/projected/ec0ad62e-84ad-44d8-a8f2-264d374ed70f-kube-api-access-fnp2s\") pod \"nova-api-db-create-57l9j\" (UID: \"ec0ad62e-84ad-44d8-a8f2-264d374ed70f\") " pod="openstack/nova-api-db-create-57l9j" Nov 25 10:50:44 crc kubenswrapper[4821]: I1125 10:50:44.886730 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ec0ad62e-84ad-44d8-a8f2-264d374ed70f-operator-scripts\") pod \"nova-api-db-create-57l9j\" (UID: \"ec0ad62e-84ad-44d8-a8f2-264d374ed70f\") " pod="openstack/nova-api-db-create-57l9j" Nov 25 10:50:44 crc kubenswrapper[4821]: I1125 10:50:44.887138 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b2a4919e-4ad6-4cdc-b1cb-5a0200dcc8cc-operator-scripts\") pod \"nova-api-eaac-account-create-f4xs9\" (UID: \"b2a4919e-4ad6-4cdc-b1cb-5a0200dcc8cc\") " pod="openstack/nova-api-eaac-account-create-f4xs9" Nov 25 10:50:44 crc kubenswrapper[4821]: I1125 10:50:44.894236 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-30f9-account-create-szdhp"] Nov 25 10:50:44 crc kubenswrapper[4821]: I1125 10:50:44.906921 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnp2s\" (UniqueName: \"kubernetes.io/projected/ec0ad62e-84ad-44d8-a8f2-264d374ed70f-kube-api-access-fnp2s\") pod \"nova-api-db-create-57l9j\" (UID: \"ec0ad62e-84ad-44d8-a8f2-264d374ed70f\") " pod="openstack/nova-api-db-create-57l9j" Nov 25 10:50:44 crc kubenswrapper[4821]: I1125 10:50:44.907319 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7wxvt\" (UniqueName: \"kubernetes.io/projected/b2a4919e-4ad6-4cdc-b1cb-5a0200dcc8cc-kube-api-access-7wxvt\") pod \"nova-api-eaac-account-create-f4xs9\" (UID: \"b2a4919e-4ad6-4cdc-b1cb-5a0200dcc8cc\") " pod="openstack/nova-api-eaac-account-create-f4xs9" Nov 25 10:50:44 crc kubenswrapper[4821]: I1125 10:50:44.956712 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-57l9j" Nov 25 10:50:44 crc kubenswrapper[4821]: I1125 10:50:44.987699 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7a0917a6-232d-43e0-8c9d-44eb9485d0d8-operator-scripts\") pod \"nova-cell0-db-create-5xnrf\" (UID: \"7a0917a6-232d-43e0-8c9d-44eb9485d0d8\") " pod="openstack/nova-cell0-db-create-5xnrf" Nov 25 10:50:44 crc kubenswrapper[4821]: I1125 10:50:44.987803 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8k9jm\" (UniqueName: \"kubernetes.io/projected/7a0917a6-232d-43e0-8c9d-44eb9485d0d8-kube-api-access-8k9jm\") pod \"nova-cell0-db-create-5xnrf\" (UID: \"7a0917a6-232d-43e0-8c9d-44eb9485d0d8\") " pod="openstack/nova-cell0-db-create-5xnrf" Nov 25 10:50:44 crc kubenswrapper[4821]: I1125 10:50:44.987883 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/af27979d-3739-453b-8c9a-1fe2c311fa17-operator-scripts\") pod \"nova-cell0-30f9-account-create-szdhp\" (UID: \"af27979d-3739-453b-8c9a-1fe2c311fa17\") " pod="openstack/nova-cell0-30f9-account-create-szdhp" Nov 25 10:50:44 crc kubenswrapper[4821]: I1125 10:50:44.987925 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9cm7\" (UniqueName: \"kubernetes.io/projected/af27979d-3739-453b-8c9a-1fe2c311fa17-kube-api-access-x9cm7\") pod \"nova-cell0-30f9-account-create-szdhp\" (UID: \"af27979d-3739-453b-8c9a-1fe2c311fa17\") " pod="openstack/nova-cell0-30f9-account-create-szdhp" Nov 25 10:50:44 crc kubenswrapper[4821]: I1125 10:50:44.987952 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/63e2c033-bfd5-4f03-96de-5ab6333b179d-operator-scripts\") pod \"nova-cell1-db-create-4n4fc\" (UID: \"63e2c033-bfd5-4f03-96de-5ab6333b179d\") " pod="openstack/nova-cell1-db-create-4n4fc" Nov 25 10:50:44 crc kubenswrapper[4821]: I1125 10:50:44.987987 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7bjw\" (UniqueName: \"kubernetes.io/projected/63e2c033-bfd5-4f03-96de-5ab6333b179d-kube-api-access-p7bjw\") pod \"nova-cell1-db-create-4n4fc\" (UID: \"63e2c033-bfd5-4f03-96de-5ab6333b179d\") " pod="openstack/nova-cell1-db-create-4n4fc" Nov 25 10:50:44 crc kubenswrapper[4821]: I1125 10:50:44.988382 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7a0917a6-232d-43e0-8c9d-44eb9485d0d8-operator-scripts\") pod \"nova-cell0-db-create-5xnrf\" (UID: \"7a0917a6-232d-43e0-8c9d-44eb9485d0d8\") " pod="openstack/nova-cell0-db-create-5xnrf" Nov 25 10:50:44 crc kubenswrapper[4821]: I1125 10:50:44.991129 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-eaac-account-create-f4xs9" Nov 25 10:50:45 crc kubenswrapper[4821]: I1125 10:50:45.007890 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8k9jm\" (UniqueName: \"kubernetes.io/projected/7a0917a6-232d-43e0-8c9d-44eb9485d0d8-kube-api-access-8k9jm\") pod \"nova-cell0-db-create-5xnrf\" (UID: \"7a0917a6-232d-43e0-8c9d-44eb9485d0d8\") " pod="openstack/nova-cell0-db-create-5xnrf" Nov 25 10:50:45 crc kubenswrapper[4821]: I1125 10:50:45.047656 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-8787-account-create-grx4m"] Nov 25 10:50:45 crc kubenswrapper[4821]: I1125 10:50:45.049236 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-8787-account-create-grx4m" Nov 25 10:50:45 crc kubenswrapper[4821]: I1125 10:50:45.053927 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Nov 25 10:50:45 crc kubenswrapper[4821]: I1125 10:50:45.055873 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-8787-account-create-grx4m"] Nov 25 10:50:45 crc kubenswrapper[4821]: I1125 10:50:45.073481 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-5xnrf" Nov 25 10:50:45 crc kubenswrapper[4821]: I1125 10:50:45.093064 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7bjw\" (UniqueName: \"kubernetes.io/projected/63e2c033-bfd5-4f03-96de-5ab6333b179d-kube-api-access-p7bjw\") pod \"nova-cell1-db-create-4n4fc\" (UID: \"63e2c033-bfd5-4f03-96de-5ab6333b179d\") " pod="openstack/nova-cell1-db-create-4n4fc" Nov 25 10:50:45 crc kubenswrapper[4821]: I1125 10:50:45.093511 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/af27979d-3739-453b-8c9a-1fe2c311fa17-operator-scripts\") pod \"nova-cell0-30f9-account-create-szdhp\" (UID: \"af27979d-3739-453b-8c9a-1fe2c311fa17\") " pod="openstack/nova-cell0-30f9-account-create-szdhp" Nov 25 10:50:45 crc kubenswrapper[4821]: I1125 10:50:45.093547 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9cm7\" (UniqueName: \"kubernetes.io/projected/af27979d-3739-453b-8c9a-1fe2c311fa17-kube-api-access-x9cm7\") pod \"nova-cell0-30f9-account-create-szdhp\" (UID: \"af27979d-3739-453b-8c9a-1fe2c311fa17\") " pod="openstack/nova-cell0-30f9-account-create-szdhp" Nov 25 10:50:45 crc kubenswrapper[4821]: I1125 10:50:45.093565 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/63e2c033-bfd5-4f03-96de-5ab6333b179d-operator-scripts\") pod \"nova-cell1-db-create-4n4fc\" (UID: \"63e2c033-bfd5-4f03-96de-5ab6333b179d\") " pod="openstack/nova-cell1-db-create-4n4fc" Nov 25 10:50:45 crc kubenswrapper[4821]: I1125 10:50:45.094533 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/af27979d-3739-453b-8c9a-1fe2c311fa17-operator-scripts\") pod \"nova-cell0-30f9-account-create-szdhp\" (UID: \"af27979d-3739-453b-8c9a-1fe2c311fa17\") " pod="openstack/nova-cell0-30f9-account-create-szdhp" Nov 25 10:50:45 crc kubenswrapper[4821]: I1125 10:50:45.094725 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/63e2c033-bfd5-4f03-96de-5ab6333b179d-operator-scripts\") pod \"nova-cell1-db-create-4n4fc\" (UID: \"63e2c033-bfd5-4f03-96de-5ab6333b179d\") " pod="openstack/nova-cell1-db-create-4n4fc" Nov 25 10:50:45 crc kubenswrapper[4821]: I1125 10:50:45.124149 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9cm7\" (UniqueName: \"kubernetes.io/projected/af27979d-3739-453b-8c9a-1fe2c311fa17-kube-api-access-x9cm7\") pod \"nova-cell0-30f9-account-create-szdhp\" (UID: \"af27979d-3739-453b-8c9a-1fe2c311fa17\") " pod="openstack/nova-cell0-30f9-account-create-szdhp" Nov 25 10:50:45 crc kubenswrapper[4821]: I1125 10:50:45.132682 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7bjw\" (UniqueName: \"kubernetes.io/projected/63e2c033-bfd5-4f03-96de-5ab6333b179d-kube-api-access-p7bjw\") pod \"nova-cell1-db-create-4n4fc\" (UID: \"63e2c033-bfd5-4f03-96de-5ab6333b179d\") " pod="openstack/nova-cell1-db-create-4n4fc" Nov 25 10:50:45 crc kubenswrapper[4821]: I1125 10:50:45.177011 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-4n4fc" Nov 25 10:50:45 crc kubenswrapper[4821]: I1125 10:50:45.189654 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-30f9-account-create-szdhp" Nov 25 10:50:45 crc kubenswrapper[4821]: I1125 10:50:45.198986 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wmt7h\" (UniqueName: \"kubernetes.io/projected/3305a4a3-6c80-4595-9b1b-9adbbbd3c83a-kube-api-access-wmt7h\") pod \"nova-cell1-8787-account-create-grx4m\" (UID: \"3305a4a3-6c80-4595-9b1b-9adbbbd3c83a\") " pod="openstack/nova-cell1-8787-account-create-grx4m" Nov 25 10:50:45 crc kubenswrapper[4821]: I1125 10:50:45.199400 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3305a4a3-6c80-4595-9b1b-9adbbbd3c83a-operator-scripts\") pod \"nova-cell1-8787-account-create-grx4m\" (UID: \"3305a4a3-6c80-4595-9b1b-9adbbbd3c83a\") " pod="openstack/nova-cell1-8787-account-create-grx4m" Nov 25 10:50:45 crc kubenswrapper[4821]: I1125 10:50:45.301469 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3305a4a3-6c80-4595-9b1b-9adbbbd3c83a-operator-scripts\") pod \"nova-cell1-8787-account-create-grx4m\" (UID: \"3305a4a3-6c80-4595-9b1b-9adbbbd3c83a\") " pod="openstack/nova-cell1-8787-account-create-grx4m" Nov 25 10:50:45 crc kubenswrapper[4821]: I1125 10:50:45.301561 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wmt7h\" (UniqueName: \"kubernetes.io/projected/3305a4a3-6c80-4595-9b1b-9adbbbd3c83a-kube-api-access-wmt7h\") pod \"nova-cell1-8787-account-create-grx4m\" (UID: \"3305a4a3-6c80-4595-9b1b-9adbbbd3c83a\") " pod="openstack/nova-cell1-8787-account-create-grx4m" Nov 25 10:50:45 crc kubenswrapper[4821]: I1125 10:50:45.305210 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3305a4a3-6c80-4595-9b1b-9adbbbd3c83a-operator-scripts\") pod \"nova-cell1-8787-account-create-grx4m\" (UID: \"3305a4a3-6c80-4595-9b1b-9adbbbd3c83a\") " pod="openstack/nova-cell1-8787-account-create-grx4m" Nov 25 10:50:45 crc kubenswrapper[4821]: I1125 10:50:45.330689 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wmt7h\" (UniqueName: \"kubernetes.io/projected/3305a4a3-6c80-4595-9b1b-9adbbbd3c83a-kube-api-access-wmt7h\") pod \"nova-cell1-8787-account-create-grx4m\" (UID: \"3305a4a3-6c80-4595-9b1b-9adbbbd3c83a\") " pod="openstack/nova-cell1-8787-account-create-grx4m" Nov 25 10:50:45 crc kubenswrapper[4821]: I1125 10:50:45.374508 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-8787-account-create-grx4m" Nov 25 10:50:45 crc kubenswrapper[4821]: I1125 10:50:45.510512 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-57l9j"] Nov 25 10:50:45 crc kubenswrapper[4821]: I1125 10:50:45.593239 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-eaac-account-create-f4xs9"] Nov 25 10:50:45 crc kubenswrapper[4821]: I1125 10:50:45.671128 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"420e728b-b00c-4075-b504-7f87402c4838","Type":"ContainerStarted","Data":"aab4176941903138e8e120df18a9e4fcc7827fea07bcbbaaaafd9b53b76a66c3"} Nov 25 10:50:45 crc kubenswrapper[4821]: I1125 10:50:45.675367 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-57l9j" event={"ID":"ec0ad62e-84ad-44d8-a8f2-264d374ed70f","Type":"ContainerStarted","Data":"0729c02bbe821d37dbd8bfbf8802999490f537f586e064ece7df3304889d3f15"} Nov 25 10:50:45 crc kubenswrapper[4821]: I1125 10:50:45.677503 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-5xnrf"] Nov 25 10:50:45 crc kubenswrapper[4821]: I1125 10:50:45.682294 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-eaac-account-create-f4xs9" event={"ID":"b2a4919e-4ad6-4cdc-b1cb-5a0200dcc8cc","Type":"ContainerStarted","Data":"b3429c2dbc4145603b246cea43d1c0de33bb5b73cbd53fd0a017bc102269839b"} Nov 25 10:50:45 crc kubenswrapper[4821]: W1125 10:50:45.689991 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7a0917a6_232d_43e0_8c9d_44eb9485d0d8.slice/crio-f11692ec664c910488176bea6ae7134231b853d051d0c4358eb163c77f4f3733 WatchSource:0}: Error finding container f11692ec664c910488176bea6ae7134231b853d051d0c4358eb163c77f4f3733: Status 404 returned error can't find the container with id f11692ec664c910488176bea6ae7134231b853d051d0c4358eb163c77f4f3733 Nov 25 10:50:45 crc kubenswrapper[4821]: W1125 10:50:45.767829 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod63e2c033_bfd5_4f03_96de_5ab6333b179d.slice/crio-15653a945d1d50ece27cdca23a47b73cedc2207abad33d3d77ae10d1e5f6f34e WatchSource:0}: Error finding container 15653a945d1d50ece27cdca23a47b73cedc2207abad33d3d77ae10d1e5f6f34e: Status 404 returned error can't find the container with id 15653a945d1d50ece27cdca23a47b73cedc2207abad33d3d77ae10d1e5f6f34e Nov 25 10:50:45 crc kubenswrapper[4821]: I1125 10:50:45.771778 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-4n4fc"] Nov 25 10:50:45 crc kubenswrapper[4821]: W1125 10:50:45.923079 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaf27979d_3739_453b_8c9a_1fe2c311fa17.slice/crio-7d24ab8d7ff0cf5856f4b8a53fd4bc27b541d9a32715c8bebde0a389443685d1 WatchSource:0}: Error finding container 7d24ab8d7ff0cf5856f4b8a53fd4bc27b541d9a32715c8bebde0a389443685d1: Status 404 returned error can't find the container with id 7d24ab8d7ff0cf5856f4b8a53fd4bc27b541d9a32715c8bebde0a389443685d1 Nov 25 10:50:45 crc kubenswrapper[4821]: I1125 10:50:45.925304 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-30f9-account-create-szdhp"] Nov 25 10:50:46 crc kubenswrapper[4821]: W1125 10:50:46.027425 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3305a4a3_6c80_4595_9b1b_9adbbbd3c83a.slice/crio-9190ecfc9a6f42154e012259b6d4f468e9ffb8f268a58f6b1d6aa07249059865 WatchSource:0}: Error finding container 9190ecfc9a6f42154e012259b6d4f468e9ffb8f268a58f6b1d6aa07249059865: Status 404 returned error can't find the container with id 9190ecfc9a6f42154e012259b6d4f468e9ffb8f268a58f6b1d6aa07249059865 Nov 25 10:50:46 crc kubenswrapper[4821]: I1125 10:50:46.028311 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-8787-account-create-grx4m"] Nov 25 10:50:46 crc kubenswrapper[4821]: I1125 10:50:46.692301 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-4n4fc" event={"ID":"63e2c033-bfd5-4f03-96de-5ab6333b179d","Type":"ContainerStarted","Data":"56412c7ea7e93aa6cbba9324eaff183a3a64bdbb1869f833b15d246ab3c9e9ae"} Nov 25 10:50:46 crc kubenswrapper[4821]: I1125 10:50:46.692394 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-4n4fc" event={"ID":"63e2c033-bfd5-4f03-96de-5ab6333b179d","Type":"ContainerStarted","Data":"15653a945d1d50ece27cdca23a47b73cedc2207abad33d3d77ae10d1e5f6f34e"} Nov 25 10:50:46 crc kubenswrapper[4821]: I1125 10:50:46.694976 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-eaac-account-create-f4xs9" event={"ID":"b2a4919e-4ad6-4cdc-b1cb-5a0200dcc8cc","Type":"ContainerStarted","Data":"a40830e437f33f17d87382d4727c377ad8dc87a7a26dc8d568de4098842642d9"} Nov 25 10:50:46 crc kubenswrapper[4821]: I1125 10:50:46.699503 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-5xnrf" event={"ID":"7a0917a6-232d-43e0-8c9d-44eb9485d0d8","Type":"ContainerStarted","Data":"04c32b3179869b2701526757c21af3020ccc290479543c3c6f210843407d8ba1"} Nov 25 10:50:46 crc kubenswrapper[4821]: I1125 10:50:46.699541 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-5xnrf" event={"ID":"7a0917a6-232d-43e0-8c9d-44eb9485d0d8","Type":"ContainerStarted","Data":"f11692ec664c910488176bea6ae7134231b853d051d0c4358eb163c77f4f3733"} Nov 25 10:50:46 crc kubenswrapper[4821]: I1125 10:50:46.701994 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-57l9j" event={"ID":"ec0ad62e-84ad-44d8-a8f2-264d374ed70f","Type":"ContainerStarted","Data":"3da8abe15577b506083dffcb7a9f7e50bafc6e525b9d574ed741d769adcfd66c"} Nov 25 10:50:46 crc kubenswrapper[4821]: I1125 10:50:46.705234 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-30f9-account-create-szdhp" event={"ID":"af27979d-3739-453b-8c9a-1fe2c311fa17","Type":"ContainerStarted","Data":"05e371437d8c7a317b3cfd78207ff0ea04163ceb02785bce5e5f1497fbf42e9a"} Nov 25 10:50:46 crc kubenswrapper[4821]: I1125 10:50:46.705269 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-30f9-account-create-szdhp" event={"ID":"af27979d-3739-453b-8c9a-1fe2c311fa17","Type":"ContainerStarted","Data":"7d24ab8d7ff0cf5856f4b8a53fd4bc27b541d9a32715c8bebde0a389443685d1"} Nov 25 10:50:46 crc kubenswrapper[4821]: I1125 10:50:46.706719 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-8787-account-create-grx4m" event={"ID":"3305a4a3-6c80-4595-9b1b-9adbbbd3c83a","Type":"ContainerStarted","Data":"df96d8a8aa5a3a8fcdbb3300533a9784a900a827714f334a3c3c4e608c75c433"} Nov 25 10:50:46 crc kubenswrapper[4821]: I1125 10:50:46.706781 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-8787-account-create-grx4m" event={"ID":"3305a4a3-6c80-4595-9b1b-9adbbbd3c83a","Type":"ContainerStarted","Data":"9190ecfc9a6f42154e012259b6d4f468e9ffb8f268a58f6b1d6aa07249059865"} Nov 25 10:50:46 crc kubenswrapper[4821]: I1125 10:50:46.731586 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-db-create-4n4fc" podStartSLOduration=2.7315701580000002 podStartE2EDuration="2.731570158s" podCreationTimestamp="2025-11-25 10:50:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:50:46.714483631 +0000 UTC m=+1117.250803478" watchObservedRunningTime="2025-11-25 10:50:46.731570158 +0000 UTC m=+1117.267890005" Nov 25 10:50:46 crc kubenswrapper[4821]: I1125 10:50:46.742403 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-30f9-account-create-szdhp" podStartSLOduration=2.742384315 podStartE2EDuration="2.742384315s" podCreationTimestamp="2025-11-25 10:50:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:50:46.741568474 +0000 UTC m=+1117.277888341" watchObservedRunningTime="2025-11-25 10:50:46.742384315 +0000 UTC m=+1117.278704162" Nov 25 10:50:46 crc kubenswrapper[4821]: I1125 10:50:46.776885 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-5xnrf" podStartSLOduration=2.776865959 podStartE2EDuration="2.776865959s" podCreationTimestamp="2025-11-25 10:50:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:50:46.767446377 +0000 UTC m=+1117.303766234" watchObservedRunningTime="2025-11-25 10:50:46.776865959 +0000 UTC m=+1117.313185796" Nov 25 10:50:46 crc kubenswrapper[4821]: I1125 10:50:46.796294 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-8787-account-create-grx4m" podStartSLOduration=1.7962784250000001 podStartE2EDuration="1.796278425s" podCreationTimestamp="2025-11-25 10:50:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:50:46.793714061 +0000 UTC m=+1117.330033908" watchObservedRunningTime="2025-11-25 10:50:46.796278425 +0000 UTC m=+1117.332598272" Nov 25 10:50:46 crc kubenswrapper[4821]: I1125 10:50:46.824385 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-57l9j" podStartSLOduration=2.824358815 podStartE2EDuration="2.824358815s" podCreationTimestamp="2025-11-25 10:50:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:50:46.816460283 +0000 UTC m=+1117.352780140" watchObservedRunningTime="2025-11-25 10:50:46.824358815 +0000 UTC m=+1117.360678662" Nov 25 10:50:46 crc kubenswrapper[4821]: I1125 10:50:46.860914 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-eaac-account-create-f4xs9" podStartSLOduration=2.860890341 podStartE2EDuration="2.860890341s" podCreationTimestamp="2025-11-25 10:50:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:50:46.838808175 +0000 UTC m=+1117.375128032" watchObservedRunningTime="2025-11-25 10:50:46.860890341 +0000 UTC m=+1117.397210188" Nov 25 10:50:47 crc kubenswrapper[4821]: I1125 10:50:47.715543 4821 generic.go:334] "Generic (PLEG): container finished" podID="3305a4a3-6c80-4595-9b1b-9adbbbd3c83a" containerID="df96d8a8aa5a3a8fcdbb3300533a9784a900a827714f334a3c3c4e608c75c433" exitCode=0 Nov 25 10:50:47 crc kubenswrapper[4821]: I1125 10:50:47.715608 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-8787-account-create-grx4m" event={"ID":"3305a4a3-6c80-4595-9b1b-9adbbbd3c83a","Type":"ContainerDied","Data":"df96d8a8aa5a3a8fcdbb3300533a9784a900a827714f334a3c3c4e608c75c433"} Nov 25 10:50:47 crc kubenswrapper[4821]: I1125 10:50:47.717968 4821 generic.go:334] "Generic (PLEG): container finished" podID="63e2c033-bfd5-4f03-96de-5ab6333b179d" containerID="56412c7ea7e93aa6cbba9324eaff183a3a64bdbb1869f833b15d246ab3c9e9ae" exitCode=0 Nov 25 10:50:47 crc kubenswrapper[4821]: I1125 10:50:47.718938 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-4n4fc" event={"ID":"63e2c033-bfd5-4f03-96de-5ab6333b179d","Type":"ContainerDied","Data":"56412c7ea7e93aa6cbba9324eaff183a3a64bdbb1869f833b15d246ab3c9e9ae"} Nov 25 10:50:48 crc kubenswrapper[4821]: I1125 10:50:48.727126 4821 generic.go:334] "Generic (PLEG): container finished" podID="ec0ad62e-84ad-44d8-a8f2-264d374ed70f" containerID="3da8abe15577b506083dffcb7a9f7e50bafc6e525b9d574ed741d769adcfd66c" exitCode=0 Nov 25 10:50:48 crc kubenswrapper[4821]: I1125 10:50:48.727226 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-57l9j" event={"ID":"ec0ad62e-84ad-44d8-a8f2-264d374ed70f","Type":"ContainerDied","Data":"3da8abe15577b506083dffcb7a9f7e50bafc6e525b9d574ed741d769adcfd66c"} Nov 25 10:50:49 crc kubenswrapper[4821]: I1125 10:50:49.170507 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-4n4fc" Nov 25 10:50:49 crc kubenswrapper[4821]: I1125 10:50:49.176693 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-8787-account-create-grx4m" Nov 25 10:50:49 crc kubenswrapper[4821]: I1125 10:50:49.318010 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p7bjw\" (UniqueName: \"kubernetes.io/projected/63e2c033-bfd5-4f03-96de-5ab6333b179d-kube-api-access-p7bjw\") pod \"63e2c033-bfd5-4f03-96de-5ab6333b179d\" (UID: \"63e2c033-bfd5-4f03-96de-5ab6333b179d\") " Nov 25 10:50:49 crc kubenswrapper[4821]: I1125 10:50:49.318248 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wmt7h\" (UniqueName: \"kubernetes.io/projected/3305a4a3-6c80-4595-9b1b-9adbbbd3c83a-kube-api-access-wmt7h\") pod \"3305a4a3-6c80-4595-9b1b-9adbbbd3c83a\" (UID: \"3305a4a3-6c80-4595-9b1b-9adbbbd3c83a\") " Nov 25 10:50:49 crc kubenswrapper[4821]: I1125 10:50:49.318366 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/63e2c033-bfd5-4f03-96de-5ab6333b179d-operator-scripts\") pod \"63e2c033-bfd5-4f03-96de-5ab6333b179d\" (UID: \"63e2c033-bfd5-4f03-96de-5ab6333b179d\") " Nov 25 10:50:49 crc kubenswrapper[4821]: I1125 10:50:49.318434 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3305a4a3-6c80-4595-9b1b-9adbbbd3c83a-operator-scripts\") pod \"3305a4a3-6c80-4595-9b1b-9adbbbd3c83a\" (UID: \"3305a4a3-6c80-4595-9b1b-9adbbbd3c83a\") " Nov 25 10:50:49 crc kubenswrapper[4821]: I1125 10:50:49.319086 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63e2c033-bfd5-4f03-96de-5ab6333b179d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "63e2c033-bfd5-4f03-96de-5ab6333b179d" (UID: "63e2c033-bfd5-4f03-96de-5ab6333b179d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:50:49 crc kubenswrapper[4821]: I1125 10:50:49.319129 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3305a4a3-6c80-4595-9b1b-9adbbbd3c83a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3305a4a3-6c80-4595-9b1b-9adbbbd3c83a" (UID: "3305a4a3-6c80-4595-9b1b-9adbbbd3c83a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:50:49 crc kubenswrapper[4821]: I1125 10:50:49.319694 4821 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/63e2c033-bfd5-4f03-96de-5ab6333b179d-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:49 crc kubenswrapper[4821]: I1125 10:50:49.319718 4821 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3305a4a3-6c80-4595-9b1b-9adbbbd3c83a-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:49 crc kubenswrapper[4821]: I1125 10:50:49.325591 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3305a4a3-6c80-4595-9b1b-9adbbbd3c83a-kube-api-access-wmt7h" (OuterVolumeSpecName: "kube-api-access-wmt7h") pod "3305a4a3-6c80-4595-9b1b-9adbbbd3c83a" (UID: "3305a4a3-6c80-4595-9b1b-9adbbbd3c83a"). InnerVolumeSpecName "kube-api-access-wmt7h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:50:49 crc kubenswrapper[4821]: I1125 10:50:49.421094 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wmt7h\" (UniqueName: \"kubernetes.io/projected/3305a4a3-6c80-4595-9b1b-9adbbbd3c83a-kube-api-access-wmt7h\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:49 crc kubenswrapper[4821]: I1125 10:50:49.422734 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63e2c033-bfd5-4f03-96de-5ab6333b179d-kube-api-access-p7bjw" (OuterVolumeSpecName: "kube-api-access-p7bjw") pod "63e2c033-bfd5-4f03-96de-5ab6333b179d" (UID: "63e2c033-bfd5-4f03-96de-5ab6333b179d"). InnerVolumeSpecName "kube-api-access-p7bjw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:50:49 crc kubenswrapper[4821]: I1125 10:50:49.523168 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p7bjw\" (UniqueName: \"kubernetes.io/projected/63e2c033-bfd5-4f03-96de-5ab6333b179d-kube-api-access-p7bjw\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:49 crc kubenswrapper[4821]: I1125 10:50:49.737731 4821 generic.go:334] "Generic (PLEG): container finished" podID="b2a4919e-4ad6-4cdc-b1cb-5a0200dcc8cc" containerID="a40830e437f33f17d87382d4727c377ad8dc87a7a26dc8d568de4098842642d9" exitCode=0 Nov 25 10:50:49 crc kubenswrapper[4821]: I1125 10:50:49.737771 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-eaac-account-create-f4xs9" event={"ID":"b2a4919e-4ad6-4cdc-b1cb-5a0200dcc8cc","Type":"ContainerDied","Data":"a40830e437f33f17d87382d4727c377ad8dc87a7a26dc8d568de4098842642d9"} Nov 25 10:50:49 crc kubenswrapper[4821]: I1125 10:50:49.740428 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"420e728b-b00c-4075-b504-7f87402c4838","Type":"ContainerStarted","Data":"abe129552a4e5ed8b81b88045f58cb71647d2741181d8499dbb4b94a651a0e2d"} Nov 25 10:50:49 crc kubenswrapper[4821]: I1125 10:50:49.742002 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-8787-account-create-grx4m" event={"ID":"3305a4a3-6c80-4595-9b1b-9adbbbd3c83a","Type":"ContainerDied","Data":"9190ecfc9a6f42154e012259b6d4f468e9ffb8f268a58f6b1d6aa07249059865"} Nov 25 10:50:49 crc kubenswrapper[4821]: I1125 10:50:49.742059 4821 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9190ecfc9a6f42154e012259b6d4f468e9ffb8f268a58f6b1d6aa07249059865" Nov 25 10:50:49 crc kubenswrapper[4821]: I1125 10:50:49.742029 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-8787-account-create-grx4m" Nov 25 10:50:49 crc kubenswrapper[4821]: I1125 10:50:49.743882 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-4n4fc" Nov 25 10:50:49 crc kubenswrapper[4821]: I1125 10:50:49.749729 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-4n4fc" event={"ID":"63e2c033-bfd5-4f03-96de-5ab6333b179d","Type":"ContainerDied","Data":"15653a945d1d50ece27cdca23a47b73cedc2207abad33d3d77ae10d1e5f6f34e"} Nov 25 10:50:49 crc kubenswrapper[4821]: I1125 10:50:49.752247 4821 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="15653a945d1d50ece27cdca23a47b73cedc2207abad33d3d77ae10d1e5f6f34e" Nov 25 10:50:50 crc kubenswrapper[4821]: I1125 10:50:50.011584 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-57l9j" Nov 25 10:50:50 crc kubenswrapper[4821]: I1125 10:50:50.052845 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fnp2s\" (UniqueName: \"kubernetes.io/projected/ec0ad62e-84ad-44d8-a8f2-264d374ed70f-kube-api-access-fnp2s\") pod \"ec0ad62e-84ad-44d8-a8f2-264d374ed70f\" (UID: \"ec0ad62e-84ad-44d8-a8f2-264d374ed70f\") " Nov 25 10:50:50 crc kubenswrapper[4821]: I1125 10:50:50.052886 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ec0ad62e-84ad-44d8-a8f2-264d374ed70f-operator-scripts\") pod \"ec0ad62e-84ad-44d8-a8f2-264d374ed70f\" (UID: \"ec0ad62e-84ad-44d8-a8f2-264d374ed70f\") " Nov 25 10:50:50 crc kubenswrapper[4821]: I1125 10:50:50.054227 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ec0ad62e-84ad-44d8-a8f2-264d374ed70f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ec0ad62e-84ad-44d8-a8f2-264d374ed70f" (UID: "ec0ad62e-84ad-44d8-a8f2-264d374ed70f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:50:50 crc kubenswrapper[4821]: I1125 10:50:50.060326 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec0ad62e-84ad-44d8-a8f2-264d374ed70f-kube-api-access-fnp2s" (OuterVolumeSpecName: "kube-api-access-fnp2s") pod "ec0ad62e-84ad-44d8-a8f2-264d374ed70f" (UID: "ec0ad62e-84ad-44d8-a8f2-264d374ed70f"). InnerVolumeSpecName "kube-api-access-fnp2s". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:50:50 crc kubenswrapper[4821]: I1125 10:50:50.158985 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fnp2s\" (UniqueName: \"kubernetes.io/projected/ec0ad62e-84ad-44d8-a8f2-264d374ed70f-kube-api-access-fnp2s\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:50 crc kubenswrapper[4821]: I1125 10:50:50.159321 4821 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ec0ad62e-84ad-44d8-a8f2-264d374ed70f-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:50 crc kubenswrapper[4821]: I1125 10:50:50.753352 4821 generic.go:334] "Generic (PLEG): container finished" podID="7a0917a6-232d-43e0-8c9d-44eb9485d0d8" containerID="04c32b3179869b2701526757c21af3020ccc290479543c3c6f210843407d8ba1" exitCode=0 Nov 25 10:50:50 crc kubenswrapper[4821]: I1125 10:50:50.753432 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-5xnrf" event={"ID":"7a0917a6-232d-43e0-8c9d-44eb9485d0d8","Type":"ContainerDied","Data":"04c32b3179869b2701526757c21af3020ccc290479543c3c6f210843407d8ba1"} Nov 25 10:50:50 crc kubenswrapper[4821]: I1125 10:50:50.754853 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-57l9j" event={"ID":"ec0ad62e-84ad-44d8-a8f2-264d374ed70f","Type":"ContainerDied","Data":"0729c02bbe821d37dbd8bfbf8802999490f537f586e064ece7df3304889d3f15"} Nov 25 10:50:50 crc kubenswrapper[4821]: I1125 10:50:50.754895 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-57l9j" Nov 25 10:50:50 crc kubenswrapper[4821]: I1125 10:50:50.754900 4821 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0729c02bbe821d37dbd8bfbf8802999490f537f586e064ece7df3304889d3f15" Nov 25 10:50:50 crc kubenswrapper[4821]: I1125 10:50:50.756211 4821 generic.go:334] "Generic (PLEG): container finished" podID="af27979d-3739-453b-8c9a-1fe2c311fa17" containerID="05e371437d8c7a317b3cfd78207ff0ea04163ceb02785bce5e5f1497fbf42e9a" exitCode=0 Nov 25 10:50:50 crc kubenswrapper[4821]: I1125 10:50:50.756255 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-30f9-account-create-szdhp" event={"ID":"af27979d-3739-453b-8c9a-1fe2c311fa17","Type":"ContainerDied","Data":"05e371437d8c7a317b3cfd78207ff0ea04163ceb02785bce5e5f1497fbf42e9a"} Nov 25 10:50:50 crc kubenswrapper[4821]: I1125 10:50:50.756449 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="420e728b-b00c-4075-b504-7f87402c4838" containerName="ceilometer-central-agent" containerID="cri-o://774c8bc29c1b4d2314e1ee4be66ba95a1f8f1661f26140696fe62e78d53de57d" gracePeriod=30 Nov 25 10:50:50 crc kubenswrapper[4821]: I1125 10:50:50.756497 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="420e728b-b00c-4075-b504-7f87402c4838" containerName="proxy-httpd" containerID="cri-o://abe129552a4e5ed8b81b88045f58cb71647d2741181d8499dbb4b94a651a0e2d" gracePeriod=30 Nov 25 10:50:50 crc kubenswrapper[4821]: I1125 10:50:50.756504 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="420e728b-b00c-4075-b504-7f87402c4838" containerName="sg-core" containerID="cri-o://aab4176941903138e8e120df18a9e4fcc7827fea07bcbbaaaafd9b53b76a66c3" gracePeriod=30 Nov 25 10:50:50 crc kubenswrapper[4821]: I1125 10:50:50.756499 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="420e728b-b00c-4075-b504-7f87402c4838" containerName="ceilometer-notification-agent" containerID="cri-o://363e08aee0b8c87fbe312f4abd6aab12624f6aab82f381b708af2b15ba064830" gracePeriod=30 Nov 25 10:50:50 crc kubenswrapper[4821]: I1125 10:50:50.818947 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=4.600765955 podStartE2EDuration="10.81892595s" podCreationTimestamp="2025-11-25 10:50:40 +0000 UTC" firstStartedPulling="2025-11-25 10:50:41.480677151 +0000 UTC m=+1112.016996998" lastFinishedPulling="2025-11-25 10:50:47.698837156 +0000 UTC m=+1118.235156993" observedRunningTime="2025-11-25 10:50:50.790950074 +0000 UTC m=+1121.327269921" watchObservedRunningTime="2025-11-25 10:50:50.81892595 +0000 UTC m=+1121.355245797" Nov 25 10:50:51 crc kubenswrapper[4821]: I1125 10:50:51.157509 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-eaac-account-create-f4xs9" Nov 25 10:50:51 crc kubenswrapper[4821]: I1125 10:50:51.277689 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b2a4919e-4ad6-4cdc-b1cb-5a0200dcc8cc-operator-scripts\") pod \"b2a4919e-4ad6-4cdc-b1cb-5a0200dcc8cc\" (UID: \"b2a4919e-4ad6-4cdc-b1cb-5a0200dcc8cc\") " Nov 25 10:50:51 crc kubenswrapper[4821]: I1125 10:50:51.277779 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7wxvt\" (UniqueName: \"kubernetes.io/projected/b2a4919e-4ad6-4cdc-b1cb-5a0200dcc8cc-kube-api-access-7wxvt\") pod \"b2a4919e-4ad6-4cdc-b1cb-5a0200dcc8cc\" (UID: \"b2a4919e-4ad6-4cdc-b1cb-5a0200dcc8cc\") " Nov 25 10:50:51 crc kubenswrapper[4821]: I1125 10:50:51.278934 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b2a4919e-4ad6-4cdc-b1cb-5a0200dcc8cc-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b2a4919e-4ad6-4cdc-b1cb-5a0200dcc8cc" (UID: "b2a4919e-4ad6-4cdc-b1cb-5a0200dcc8cc"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:50:51 crc kubenswrapper[4821]: I1125 10:50:51.280412 4821 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b2a4919e-4ad6-4cdc-b1cb-5a0200dcc8cc-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:51 crc kubenswrapper[4821]: I1125 10:50:51.283947 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2a4919e-4ad6-4cdc-b1cb-5a0200dcc8cc-kube-api-access-7wxvt" (OuterVolumeSpecName: "kube-api-access-7wxvt") pod "b2a4919e-4ad6-4cdc-b1cb-5a0200dcc8cc" (UID: "b2a4919e-4ad6-4cdc-b1cb-5a0200dcc8cc"). InnerVolumeSpecName "kube-api-access-7wxvt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:50:51 crc kubenswrapper[4821]: I1125 10:50:51.382005 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7wxvt\" (UniqueName: \"kubernetes.io/projected/b2a4919e-4ad6-4cdc-b1cb-5a0200dcc8cc-kube-api-access-7wxvt\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:51 crc kubenswrapper[4821]: I1125 10:50:51.765595 4821 generic.go:334] "Generic (PLEG): container finished" podID="420e728b-b00c-4075-b504-7f87402c4838" containerID="abe129552a4e5ed8b81b88045f58cb71647d2741181d8499dbb4b94a651a0e2d" exitCode=0 Nov 25 10:50:51 crc kubenswrapper[4821]: I1125 10:50:51.765610 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"420e728b-b00c-4075-b504-7f87402c4838","Type":"ContainerDied","Data":"abe129552a4e5ed8b81b88045f58cb71647d2741181d8499dbb4b94a651a0e2d"} Nov 25 10:50:51 crc kubenswrapper[4821]: I1125 10:50:51.781459 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-eaac-account-create-f4xs9" Nov 25 10:50:51 crc kubenswrapper[4821]: I1125 10:50:51.781681 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-eaac-account-create-f4xs9" event={"ID":"b2a4919e-4ad6-4cdc-b1cb-5a0200dcc8cc","Type":"ContainerDied","Data":"b3429c2dbc4145603b246cea43d1c0de33bb5b73cbd53fd0a017bc102269839b"} Nov 25 10:50:51 crc kubenswrapper[4821]: I1125 10:50:51.781723 4821 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b3429c2dbc4145603b246cea43d1c0de33bb5b73cbd53fd0a017bc102269839b" Nov 25 10:50:52 crc kubenswrapper[4821]: I1125 10:50:52.190649 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-5xnrf" Nov 25 10:50:52 crc kubenswrapper[4821]: I1125 10:50:52.196395 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-30f9-account-create-szdhp" Nov 25 10:50:52 crc kubenswrapper[4821]: I1125 10:50:52.306415 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x9cm7\" (UniqueName: \"kubernetes.io/projected/af27979d-3739-453b-8c9a-1fe2c311fa17-kube-api-access-x9cm7\") pod \"af27979d-3739-453b-8c9a-1fe2c311fa17\" (UID: \"af27979d-3739-453b-8c9a-1fe2c311fa17\") " Nov 25 10:50:52 crc kubenswrapper[4821]: I1125 10:50:52.306480 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8k9jm\" (UniqueName: \"kubernetes.io/projected/7a0917a6-232d-43e0-8c9d-44eb9485d0d8-kube-api-access-8k9jm\") pod \"7a0917a6-232d-43e0-8c9d-44eb9485d0d8\" (UID: \"7a0917a6-232d-43e0-8c9d-44eb9485d0d8\") " Nov 25 10:50:52 crc kubenswrapper[4821]: I1125 10:50:52.306530 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7a0917a6-232d-43e0-8c9d-44eb9485d0d8-operator-scripts\") pod \"7a0917a6-232d-43e0-8c9d-44eb9485d0d8\" (UID: \"7a0917a6-232d-43e0-8c9d-44eb9485d0d8\") " Nov 25 10:50:52 crc kubenswrapper[4821]: I1125 10:50:52.306577 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/af27979d-3739-453b-8c9a-1fe2c311fa17-operator-scripts\") pod \"af27979d-3739-453b-8c9a-1fe2c311fa17\" (UID: \"af27979d-3739-453b-8c9a-1fe2c311fa17\") " Nov 25 10:50:52 crc kubenswrapper[4821]: I1125 10:50:52.307341 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af27979d-3739-453b-8c9a-1fe2c311fa17-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "af27979d-3739-453b-8c9a-1fe2c311fa17" (UID: "af27979d-3739-453b-8c9a-1fe2c311fa17"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:50:52 crc kubenswrapper[4821]: I1125 10:50:52.307600 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a0917a6-232d-43e0-8c9d-44eb9485d0d8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7a0917a6-232d-43e0-8c9d-44eb9485d0d8" (UID: "7a0917a6-232d-43e0-8c9d-44eb9485d0d8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:50:52 crc kubenswrapper[4821]: I1125 10:50:52.311248 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a0917a6-232d-43e0-8c9d-44eb9485d0d8-kube-api-access-8k9jm" (OuterVolumeSpecName: "kube-api-access-8k9jm") pod "7a0917a6-232d-43e0-8c9d-44eb9485d0d8" (UID: "7a0917a6-232d-43e0-8c9d-44eb9485d0d8"). InnerVolumeSpecName "kube-api-access-8k9jm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:50:52 crc kubenswrapper[4821]: I1125 10:50:52.311579 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af27979d-3739-453b-8c9a-1fe2c311fa17-kube-api-access-x9cm7" (OuterVolumeSpecName: "kube-api-access-x9cm7") pod "af27979d-3739-453b-8c9a-1fe2c311fa17" (UID: "af27979d-3739-453b-8c9a-1fe2c311fa17"). InnerVolumeSpecName "kube-api-access-x9cm7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:50:52 crc kubenswrapper[4821]: I1125 10:50:52.408417 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x9cm7\" (UniqueName: \"kubernetes.io/projected/af27979d-3739-453b-8c9a-1fe2c311fa17-kube-api-access-x9cm7\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:52 crc kubenswrapper[4821]: I1125 10:50:52.408755 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8k9jm\" (UniqueName: \"kubernetes.io/projected/7a0917a6-232d-43e0-8c9d-44eb9485d0d8-kube-api-access-8k9jm\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:52 crc kubenswrapper[4821]: I1125 10:50:52.408765 4821 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7a0917a6-232d-43e0-8c9d-44eb9485d0d8-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:52 crc kubenswrapper[4821]: I1125 10:50:52.408774 4821 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/af27979d-3739-453b-8c9a-1fe2c311fa17-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:50:52 crc kubenswrapper[4821]: I1125 10:50:52.797072 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-30f9-account-create-szdhp" event={"ID":"af27979d-3739-453b-8c9a-1fe2c311fa17","Type":"ContainerDied","Data":"7d24ab8d7ff0cf5856f4b8a53fd4bc27b541d9a32715c8bebde0a389443685d1"} Nov 25 10:50:52 crc kubenswrapper[4821]: I1125 10:50:52.797122 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-30f9-account-create-szdhp" Nov 25 10:50:52 crc kubenswrapper[4821]: I1125 10:50:52.797138 4821 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7d24ab8d7ff0cf5856f4b8a53fd4bc27b541d9a32715c8bebde0a389443685d1" Nov 25 10:50:52 crc kubenswrapper[4821]: I1125 10:50:52.803066 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-5xnrf" Nov 25 10:50:52 crc kubenswrapper[4821]: I1125 10:50:52.803071 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-5xnrf" event={"ID":"7a0917a6-232d-43e0-8c9d-44eb9485d0d8","Type":"ContainerDied","Data":"f11692ec664c910488176bea6ae7134231b853d051d0c4358eb163c77f4f3733"} Nov 25 10:50:52 crc kubenswrapper[4821]: I1125 10:50:52.803127 4821 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f11692ec664c910488176bea6ae7134231b853d051d0c4358eb163c77f4f3733" Nov 25 10:50:52 crc kubenswrapper[4821]: I1125 10:50:52.806501 4821 generic.go:334] "Generic (PLEG): container finished" podID="420e728b-b00c-4075-b504-7f87402c4838" containerID="aab4176941903138e8e120df18a9e4fcc7827fea07bcbbaaaafd9b53b76a66c3" exitCode=2 Nov 25 10:50:52 crc kubenswrapper[4821]: I1125 10:50:52.806536 4821 generic.go:334] "Generic (PLEG): container finished" podID="420e728b-b00c-4075-b504-7f87402c4838" containerID="363e08aee0b8c87fbe312f4abd6aab12624f6aab82f381b708af2b15ba064830" exitCode=0 Nov 25 10:50:52 crc kubenswrapper[4821]: I1125 10:50:52.806559 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"420e728b-b00c-4075-b504-7f87402c4838","Type":"ContainerDied","Data":"aab4176941903138e8e120df18a9e4fcc7827fea07bcbbaaaafd9b53b76a66c3"} Nov 25 10:50:52 crc kubenswrapper[4821]: I1125 10:50:52.806586 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"420e728b-b00c-4075-b504-7f87402c4838","Type":"ContainerDied","Data":"363e08aee0b8c87fbe312f4abd6aab12624f6aab82f381b708af2b15ba064830"} Nov 25 10:50:55 crc kubenswrapper[4821]: I1125 10:50:55.222518 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-42f5v"] Nov 25 10:50:55 crc kubenswrapper[4821]: E1125 10:50:55.223833 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec0ad62e-84ad-44d8-a8f2-264d374ed70f" containerName="mariadb-database-create" Nov 25 10:50:55 crc kubenswrapper[4821]: I1125 10:50:55.223858 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec0ad62e-84ad-44d8-a8f2-264d374ed70f" containerName="mariadb-database-create" Nov 25 10:50:55 crc kubenswrapper[4821]: E1125 10:50:55.223891 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a0917a6-232d-43e0-8c9d-44eb9485d0d8" containerName="mariadb-database-create" Nov 25 10:50:55 crc kubenswrapper[4821]: I1125 10:50:55.223900 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a0917a6-232d-43e0-8c9d-44eb9485d0d8" containerName="mariadb-database-create" Nov 25 10:50:55 crc kubenswrapper[4821]: E1125 10:50:55.223916 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af27979d-3739-453b-8c9a-1fe2c311fa17" containerName="mariadb-account-create" Nov 25 10:50:55 crc kubenswrapper[4821]: I1125 10:50:55.223924 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="af27979d-3739-453b-8c9a-1fe2c311fa17" containerName="mariadb-account-create" Nov 25 10:50:55 crc kubenswrapper[4821]: E1125 10:50:55.223932 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63e2c033-bfd5-4f03-96de-5ab6333b179d" containerName="mariadb-database-create" Nov 25 10:50:55 crc kubenswrapper[4821]: I1125 10:50:55.223939 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="63e2c033-bfd5-4f03-96de-5ab6333b179d" containerName="mariadb-database-create" Nov 25 10:50:55 crc kubenswrapper[4821]: E1125 10:50:55.223965 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3305a4a3-6c80-4595-9b1b-9adbbbd3c83a" containerName="mariadb-account-create" Nov 25 10:50:55 crc kubenswrapper[4821]: I1125 10:50:55.223973 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="3305a4a3-6c80-4595-9b1b-9adbbbd3c83a" containerName="mariadb-account-create" Nov 25 10:50:55 crc kubenswrapper[4821]: E1125 10:50:55.223988 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2a4919e-4ad6-4cdc-b1cb-5a0200dcc8cc" containerName="mariadb-account-create" Nov 25 10:50:55 crc kubenswrapper[4821]: I1125 10:50:55.223997 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2a4919e-4ad6-4cdc-b1cb-5a0200dcc8cc" containerName="mariadb-account-create" Nov 25 10:50:55 crc kubenswrapper[4821]: I1125 10:50:55.224223 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="af27979d-3739-453b-8c9a-1fe2c311fa17" containerName="mariadb-account-create" Nov 25 10:50:55 crc kubenswrapper[4821]: I1125 10:50:55.224248 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2a4919e-4ad6-4cdc-b1cb-5a0200dcc8cc" containerName="mariadb-account-create" Nov 25 10:50:55 crc kubenswrapper[4821]: I1125 10:50:55.224262 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a0917a6-232d-43e0-8c9d-44eb9485d0d8" containerName="mariadb-database-create" Nov 25 10:50:55 crc kubenswrapper[4821]: I1125 10:50:55.224273 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="63e2c033-bfd5-4f03-96de-5ab6333b179d" containerName="mariadb-database-create" Nov 25 10:50:55 crc kubenswrapper[4821]: I1125 10:50:55.224283 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec0ad62e-84ad-44d8-a8f2-264d374ed70f" containerName="mariadb-database-create" Nov 25 10:50:55 crc kubenswrapper[4821]: I1125 10:50:55.224296 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="3305a4a3-6c80-4595-9b1b-9adbbbd3c83a" containerName="mariadb-account-create" Nov 25 10:50:55 crc kubenswrapper[4821]: I1125 10:50:55.225058 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-42f5v" Nov 25 10:50:55 crc kubenswrapper[4821]: I1125 10:50:55.228986 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 25 10:50:55 crc kubenswrapper[4821]: I1125 10:50:55.229336 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-865zj" Nov 25 10:50:55 crc kubenswrapper[4821]: I1125 10:50:55.233138 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-42f5v"] Nov 25 10:50:55 crc kubenswrapper[4821]: I1125 10:50:55.233633 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Nov 25 10:50:55 crc kubenswrapper[4821]: I1125 10:50:55.358925 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7745217e-45ba-403f-bf4f-1cbced054784-scripts\") pod \"nova-cell0-conductor-db-sync-42f5v\" (UID: \"7745217e-45ba-403f-bf4f-1cbced054784\") " pod="openstack/nova-cell0-conductor-db-sync-42f5v" Nov 25 10:50:55 crc kubenswrapper[4821]: I1125 10:50:55.358998 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7745217e-45ba-403f-bf4f-1cbced054784-config-data\") pod \"nova-cell0-conductor-db-sync-42f5v\" (UID: \"7745217e-45ba-403f-bf4f-1cbced054784\") " pod="openstack/nova-cell0-conductor-db-sync-42f5v" Nov 25 10:50:55 crc kubenswrapper[4821]: I1125 10:50:55.359020 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7745217e-45ba-403f-bf4f-1cbced054784-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-42f5v\" (UID: \"7745217e-45ba-403f-bf4f-1cbced054784\") " pod="openstack/nova-cell0-conductor-db-sync-42f5v" Nov 25 10:50:55 crc kubenswrapper[4821]: I1125 10:50:55.359062 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7klvp\" (UniqueName: \"kubernetes.io/projected/7745217e-45ba-403f-bf4f-1cbced054784-kube-api-access-7klvp\") pod \"nova-cell0-conductor-db-sync-42f5v\" (UID: \"7745217e-45ba-403f-bf4f-1cbced054784\") " pod="openstack/nova-cell0-conductor-db-sync-42f5v" Nov 25 10:50:55 crc kubenswrapper[4821]: I1125 10:50:55.462195 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7745217e-45ba-403f-bf4f-1cbced054784-scripts\") pod \"nova-cell0-conductor-db-sync-42f5v\" (UID: \"7745217e-45ba-403f-bf4f-1cbced054784\") " pod="openstack/nova-cell0-conductor-db-sync-42f5v" Nov 25 10:50:55 crc kubenswrapper[4821]: I1125 10:50:55.462582 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7745217e-45ba-403f-bf4f-1cbced054784-config-data\") pod \"nova-cell0-conductor-db-sync-42f5v\" (UID: \"7745217e-45ba-403f-bf4f-1cbced054784\") " pod="openstack/nova-cell0-conductor-db-sync-42f5v" Nov 25 10:50:55 crc kubenswrapper[4821]: I1125 10:50:55.462614 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7745217e-45ba-403f-bf4f-1cbced054784-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-42f5v\" (UID: \"7745217e-45ba-403f-bf4f-1cbced054784\") " pod="openstack/nova-cell0-conductor-db-sync-42f5v" Nov 25 10:50:55 crc kubenswrapper[4821]: I1125 10:50:55.462637 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7klvp\" (UniqueName: \"kubernetes.io/projected/7745217e-45ba-403f-bf4f-1cbced054784-kube-api-access-7klvp\") pod \"nova-cell0-conductor-db-sync-42f5v\" (UID: \"7745217e-45ba-403f-bf4f-1cbced054784\") " pod="openstack/nova-cell0-conductor-db-sync-42f5v" Nov 25 10:50:55 crc kubenswrapper[4821]: I1125 10:50:55.470233 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7745217e-45ba-403f-bf4f-1cbced054784-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-42f5v\" (UID: \"7745217e-45ba-403f-bf4f-1cbced054784\") " pod="openstack/nova-cell0-conductor-db-sync-42f5v" Nov 25 10:50:55 crc kubenswrapper[4821]: I1125 10:50:55.471727 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7745217e-45ba-403f-bf4f-1cbced054784-scripts\") pod \"nova-cell0-conductor-db-sync-42f5v\" (UID: \"7745217e-45ba-403f-bf4f-1cbced054784\") " pod="openstack/nova-cell0-conductor-db-sync-42f5v" Nov 25 10:50:55 crc kubenswrapper[4821]: I1125 10:50:55.477929 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7745217e-45ba-403f-bf4f-1cbced054784-config-data\") pod \"nova-cell0-conductor-db-sync-42f5v\" (UID: \"7745217e-45ba-403f-bf4f-1cbced054784\") " pod="openstack/nova-cell0-conductor-db-sync-42f5v" Nov 25 10:50:55 crc kubenswrapper[4821]: I1125 10:50:55.482632 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7klvp\" (UniqueName: \"kubernetes.io/projected/7745217e-45ba-403f-bf4f-1cbced054784-kube-api-access-7klvp\") pod \"nova-cell0-conductor-db-sync-42f5v\" (UID: \"7745217e-45ba-403f-bf4f-1cbced054784\") " pod="openstack/nova-cell0-conductor-db-sync-42f5v" Nov 25 10:50:55 crc kubenswrapper[4821]: I1125 10:50:55.553658 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-42f5v" Nov 25 10:50:56 crc kubenswrapper[4821]: I1125 10:50:56.019987 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-42f5v"] Nov 25 10:50:56 crc kubenswrapper[4821]: I1125 10:50:56.842221 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-42f5v" event={"ID":"7745217e-45ba-403f-bf4f-1cbced054784","Type":"ContainerStarted","Data":"933af00ac731f452a40fb942d646576e0d483ceeb14bb6145afa58ed1cf7498f"} Nov 25 10:51:01 crc kubenswrapper[4821]: I1125 10:51:01.893951 4821 generic.go:334] "Generic (PLEG): container finished" podID="420e728b-b00c-4075-b504-7f87402c4838" containerID="774c8bc29c1b4d2314e1ee4be66ba95a1f8f1661f26140696fe62e78d53de57d" exitCode=0 Nov 25 10:51:01 crc kubenswrapper[4821]: I1125 10:51:01.894028 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"420e728b-b00c-4075-b504-7f87402c4838","Type":"ContainerDied","Data":"774c8bc29c1b4d2314e1ee4be66ba95a1f8f1661f26140696fe62e78d53de57d"} Nov 25 10:51:04 crc kubenswrapper[4821]: I1125 10:51:04.718826 4821 patch_prober.go:28] interesting pod/machine-config-daemon-2krbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:51:04 crc kubenswrapper[4821]: I1125 10:51:04.720604 4821 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:51:05 crc kubenswrapper[4821]: I1125 10:51:05.414785 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:51:05 crc kubenswrapper[4821]: I1125 10:51:05.548457 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/420e728b-b00c-4075-b504-7f87402c4838-config-data\") pod \"420e728b-b00c-4075-b504-7f87402c4838\" (UID: \"420e728b-b00c-4075-b504-7f87402c4838\") " Nov 25 10:51:05 crc kubenswrapper[4821]: I1125 10:51:05.549036 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/420e728b-b00c-4075-b504-7f87402c4838-log-httpd\") pod \"420e728b-b00c-4075-b504-7f87402c4838\" (UID: \"420e728b-b00c-4075-b504-7f87402c4838\") " Nov 25 10:51:05 crc kubenswrapper[4821]: I1125 10:51:05.549230 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/420e728b-b00c-4075-b504-7f87402c4838-combined-ca-bundle\") pod \"420e728b-b00c-4075-b504-7f87402c4838\" (UID: \"420e728b-b00c-4075-b504-7f87402c4838\") " Nov 25 10:51:05 crc kubenswrapper[4821]: I1125 10:51:05.549401 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/420e728b-b00c-4075-b504-7f87402c4838-scripts\") pod \"420e728b-b00c-4075-b504-7f87402c4838\" (UID: \"420e728b-b00c-4075-b504-7f87402c4838\") " Nov 25 10:51:05 crc kubenswrapper[4821]: I1125 10:51:05.549550 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/420e728b-b00c-4075-b504-7f87402c4838-sg-core-conf-yaml\") pod \"420e728b-b00c-4075-b504-7f87402c4838\" (UID: \"420e728b-b00c-4075-b504-7f87402c4838\") " Nov 25 10:51:05 crc kubenswrapper[4821]: I1125 10:51:05.549698 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wx2sh\" (UniqueName: \"kubernetes.io/projected/420e728b-b00c-4075-b504-7f87402c4838-kube-api-access-wx2sh\") pod \"420e728b-b00c-4075-b504-7f87402c4838\" (UID: \"420e728b-b00c-4075-b504-7f87402c4838\") " Nov 25 10:51:05 crc kubenswrapper[4821]: I1125 10:51:05.549564 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/420e728b-b00c-4075-b504-7f87402c4838-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "420e728b-b00c-4075-b504-7f87402c4838" (UID: "420e728b-b00c-4075-b504-7f87402c4838"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:51:05 crc kubenswrapper[4821]: I1125 10:51:05.549945 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/420e728b-b00c-4075-b504-7f87402c4838-run-httpd\") pod \"420e728b-b00c-4075-b504-7f87402c4838\" (UID: \"420e728b-b00c-4075-b504-7f87402c4838\") " Nov 25 10:51:05 crc kubenswrapper[4821]: I1125 10:51:05.550272 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/420e728b-b00c-4075-b504-7f87402c4838-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "420e728b-b00c-4075-b504-7f87402c4838" (UID: "420e728b-b00c-4075-b504-7f87402c4838"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:51:05 crc kubenswrapper[4821]: I1125 10:51:05.550749 4821 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/420e728b-b00c-4075-b504-7f87402c4838-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 10:51:05 crc kubenswrapper[4821]: I1125 10:51:05.550837 4821 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/420e728b-b00c-4075-b504-7f87402c4838-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 10:51:05 crc kubenswrapper[4821]: I1125 10:51:05.554266 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/420e728b-b00c-4075-b504-7f87402c4838-kube-api-access-wx2sh" (OuterVolumeSpecName: "kube-api-access-wx2sh") pod "420e728b-b00c-4075-b504-7f87402c4838" (UID: "420e728b-b00c-4075-b504-7f87402c4838"). InnerVolumeSpecName "kube-api-access-wx2sh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:51:05 crc kubenswrapper[4821]: I1125 10:51:05.554616 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/420e728b-b00c-4075-b504-7f87402c4838-scripts" (OuterVolumeSpecName: "scripts") pod "420e728b-b00c-4075-b504-7f87402c4838" (UID: "420e728b-b00c-4075-b504-7f87402c4838"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:51:05 crc kubenswrapper[4821]: I1125 10:51:05.580388 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/420e728b-b00c-4075-b504-7f87402c4838-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "420e728b-b00c-4075-b504-7f87402c4838" (UID: "420e728b-b00c-4075-b504-7f87402c4838"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:51:05 crc kubenswrapper[4821]: I1125 10:51:05.635300 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/420e728b-b00c-4075-b504-7f87402c4838-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "420e728b-b00c-4075-b504-7f87402c4838" (UID: "420e728b-b00c-4075-b504-7f87402c4838"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:51:05 crc kubenswrapper[4821]: I1125 10:51:05.652313 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wx2sh\" (UniqueName: \"kubernetes.io/projected/420e728b-b00c-4075-b504-7f87402c4838-kube-api-access-wx2sh\") on node \"crc\" DevicePath \"\"" Nov 25 10:51:05 crc kubenswrapper[4821]: I1125 10:51:05.652349 4821 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/420e728b-b00c-4075-b504-7f87402c4838-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:51:05 crc kubenswrapper[4821]: I1125 10:51:05.652370 4821 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/420e728b-b00c-4075-b504-7f87402c4838-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:51:05 crc kubenswrapper[4821]: I1125 10:51:05.652380 4821 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/420e728b-b00c-4075-b504-7f87402c4838-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 25 10:51:05 crc kubenswrapper[4821]: I1125 10:51:05.659117 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/420e728b-b00c-4075-b504-7f87402c4838-config-data" (OuterVolumeSpecName: "config-data") pod "420e728b-b00c-4075-b504-7f87402c4838" (UID: "420e728b-b00c-4075-b504-7f87402c4838"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:51:05 crc kubenswrapper[4821]: I1125 10:51:05.754636 4821 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/420e728b-b00c-4075-b504-7f87402c4838-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:51:05 crc kubenswrapper[4821]: I1125 10:51:05.929790 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-42f5v" event={"ID":"7745217e-45ba-403f-bf4f-1cbced054784","Type":"ContainerStarted","Data":"4148d293fc86b2ec5132a58cee28e829308cf0fbcd9ab995a4217bc85f1c0c50"} Nov 25 10:51:05 crc kubenswrapper[4821]: I1125 10:51:05.934472 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"420e728b-b00c-4075-b504-7f87402c4838","Type":"ContainerDied","Data":"5e4b32c9290d6f7ac99276e5ad69ff7a3df810b2f477595047fdca4c25896447"} Nov 25 10:51:05 crc kubenswrapper[4821]: I1125 10:51:05.934719 4821 scope.go:117] "RemoveContainer" containerID="abe129552a4e5ed8b81b88045f58cb71647d2741181d8499dbb4b94a651a0e2d" Nov 25 10:51:05 crc kubenswrapper[4821]: I1125 10:51:05.934541 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:51:05 crc kubenswrapper[4821]: I1125 10:51:05.948149 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-42f5v" podStartSLOduration=1.844930374 podStartE2EDuration="10.948132663s" podCreationTimestamp="2025-11-25 10:50:55 +0000 UTC" firstStartedPulling="2025-11-25 10:50:56.021675265 +0000 UTC m=+1126.557995112" lastFinishedPulling="2025-11-25 10:51:05.124877554 +0000 UTC m=+1135.661197401" observedRunningTime="2025-11-25 10:51:05.946671456 +0000 UTC m=+1136.482991303" watchObservedRunningTime="2025-11-25 10:51:05.948132663 +0000 UTC m=+1136.484452510" Nov 25 10:51:05 crc kubenswrapper[4821]: I1125 10:51:05.968681 4821 scope.go:117] "RemoveContainer" containerID="aab4176941903138e8e120df18a9e4fcc7827fea07bcbbaaaafd9b53b76a66c3" Nov 25 10:51:05 crc kubenswrapper[4821]: I1125 10:51:05.980717 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:51:05 crc kubenswrapper[4821]: I1125 10:51:05.993924 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:51:06 crc kubenswrapper[4821]: I1125 10:51:06.005926 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:51:06 crc kubenswrapper[4821]: E1125 10:51:06.007459 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="420e728b-b00c-4075-b504-7f87402c4838" containerName="proxy-httpd" Nov 25 10:51:06 crc kubenswrapper[4821]: I1125 10:51:06.009029 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="420e728b-b00c-4075-b504-7f87402c4838" containerName="proxy-httpd" Nov 25 10:51:06 crc kubenswrapper[4821]: E1125 10:51:06.009203 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="420e728b-b00c-4075-b504-7f87402c4838" containerName="sg-core" Nov 25 10:51:06 crc kubenswrapper[4821]: I1125 10:51:06.009278 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="420e728b-b00c-4075-b504-7f87402c4838" containerName="sg-core" Nov 25 10:51:06 crc kubenswrapper[4821]: E1125 10:51:06.009397 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="420e728b-b00c-4075-b504-7f87402c4838" containerName="ceilometer-central-agent" Nov 25 10:51:06 crc kubenswrapper[4821]: I1125 10:51:06.009474 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="420e728b-b00c-4075-b504-7f87402c4838" containerName="ceilometer-central-agent" Nov 25 10:51:06 crc kubenswrapper[4821]: E1125 10:51:06.009565 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="420e728b-b00c-4075-b504-7f87402c4838" containerName="ceilometer-notification-agent" Nov 25 10:51:06 crc kubenswrapper[4821]: I1125 10:51:06.009650 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="420e728b-b00c-4075-b504-7f87402c4838" containerName="ceilometer-notification-agent" Nov 25 10:51:06 crc kubenswrapper[4821]: I1125 10:51:06.010094 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="420e728b-b00c-4075-b504-7f87402c4838" containerName="sg-core" Nov 25 10:51:06 crc kubenswrapper[4821]: I1125 10:51:06.010242 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="420e728b-b00c-4075-b504-7f87402c4838" containerName="ceilometer-notification-agent" Nov 25 10:51:06 crc kubenswrapper[4821]: I1125 10:51:06.010322 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="420e728b-b00c-4075-b504-7f87402c4838" containerName="proxy-httpd" Nov 25 10:51:06 crc kubenswrapper[4821]: I1125 10:51:06.010403 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="420e728b-b00c-4075-b504-7f87402c4838" containerName="ceilometer-central-agent" Nov 25 10:51:06 crc kubenswrapper[4821]: I1125 10:51:06.012856 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:51:06 crc kubenswrapper[4821]: I1125 10:51:06.017278 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 25 10:51:06 crc kubenswrapper[4821]: I1125 10:51:06.018721 4821 scope.go:117] "RemoveContainer" containerID="363e08aee0b8c87fbe312f4abd6aab12624f6aab82f381b708af2b15ba064830" Nov 25 10:51:06 crc kubenswrapper[4821]: I1125 10:51:06.021417 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 25 10:51:06 crc kubenswrapper[4821]: I1125 10:51:06.021600 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:51:06 crc kubenswrapper[4821]: I1125 10:51:06.104015 4821 scope.go:117] "RemoveContainer" containerID="774c8bc29c1b4d2314e1ee4be66ba95a1f8f1661f26140696fe62e78d53de57d" Nov 25 10:51:06 crc kubenswrapper[4821]: I1125 10:51:06.126488 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="420e728b-b00c-4075-b504-7f87402c4838" path="/var/lib/kubelet/pods/420e728b-b00c-4075-b504-7f87402c4838/volumes" Nov 25 10:51:06 crc kubenswrapper[4821]: I1125 10:51:06.161487 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1eb58e31-397e-4e03-8e61-a643644e01a6-run-httpd\") pod \"ceilometer-0\" (UID: \"1eb58e31-397e-4e03-8e61-a643644e01a6\") " pod="openstack/ceilometer-0" Nov 25 10:51:06 crc kubenswrapper[4821]: I1125 10:51:06.161777 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1eb58e31-397e-4e03-8e61-a643644e01a6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1eb58e31-397e-4e03-8e61-a643644e01a6\") " pod="openstack/ceilometer-0" Nov 25 10:51:06 crc kubenswrapper[4821]: I1125 10:51:06.161891 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1eb58e31-397e-4e03-8e61-a643644e01a6-config-data\") pod \"ceilometer-0\" (UID: \"1eb58e31-397e-4e03-8e61-a643644e01a6\") " pod="openstack/ceilometer-0" Nov 25 10:51:06 crc kubenswrapper[4821]: I1125 10:51:06.161967 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1eb58e31-397e-4e03-8e61-a643644e01a6-scripts\") pod \"ceilometer-0\" (UID: \"1eb58e31-397e-4e03-8e61-a643644e01a6\") " pod="openstack/ceilometer-0" Nov 25 10:51:06 crc kubenswrapper[4821]: I1125 10:51:06.162037 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eb58e31-397e-4e03-8e61-a643644e01a6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1eb58e31-397e-4e03-8e61-a643644e01a6\") " pod="openstack/ceilometer-0" Nov 25 10:51:06 crc kubenswrapper[4821]: I1125 10:51:06.162194 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1eb58e31-397e-4e03-8e61-a643644e01a6-log-httpd\") pod \"ceilometer-0\" (UID: \"1eb58e31-397e-4e03-8e61-a643644e01a6\") " pod="openstack/ceilometer-0" Nov 25 10:51:06 crc kubenswrapper[4821]: I1125 10:51:06.162292 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zblt\" (UniqueName: \"kubernetes.io/projected/1eb58e31-397e-4e03-8e61-a643644e01a6-kube-api-access-4zblt\") pod \"ceilometer-0\" (UID: \"1eb58e31-397e-4e03-8e61-a643644e01a6\") " pod="openstack/ceilometer-0" Nov 25 10:51:06 crc kubenswrapper[4821]: I1125 10:51:06.264819 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1eb58e31-397e-4e03-8e61-a643644e01a6-log-httpd\") pod \"ceilometer-0\" (UID: \"1eb58e31-397e-4e03-8e61-a643644e01a6\") " pod="openstack/ceilometer-0" Nov 25 10:51:06 crc kubenswrapper[4821]: I1125 10:51:06.264897 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zblt\" (UniqueName: \"kubernetes.io/projected/1eb58e31-397e-4e03-8e61-a643644e01a6-kube-api-access-4zblt\") pod \"ceilometer-0\" (UID: \"1eb58e31-397e-4e03-8e61-a643644e01a6\") " pod="openstack/ceilometer-0" Nov 25 10:51:06 crc kubenswrapper[4821]: I1125 10:51:06.264950 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1eb58e31-397e-4e03-8e61-a643644e01a6-run-httpd\") pod \"ceilometer-0\" (UID: \"1eb58e31-397e-4e03-8e61-a643644e01a6\") " pod="openstack/ceilometer-0" Nov 25 10:51:06 crc kubenswrapper[4821]: I1125 10:51:06.265002 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1eb58e31-397e-4e03-8e61-a643644e01a6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1eb58e31-397e-4e03-8e61-a643644e01a6\") " pod="openstack/ceilometer-0" Nov 25 10:51:06 crc kubenswrapper[4821]: I1125 10:51:06.265099 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1eb58e31-397e-4e03-8e61-a643644e01a6-config-data\") pod \"ceilometer-0\" (UID: \"1eb58e31-397e-4e03-8e61-a643644e01a6\") " pod="openstack/ceilometer-0" Nov 25 10:51:06 crc kubenswrapper[4821]: I1125 10:51:06.265138 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1eb58e31-397e-4e03-8e61-a643644e01a6-scripts\") pod \"ceilometer-0\" (UID: \"1eb58e31-397e-4e03-8e61-a643644e01a6\") " pod="openstack/ceilometer-0" Nov 25 10:51:06 crc kubenswrapper[4821]: I1125 10:51:06.265193 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eb58e31-397e-4e03-8e61-a643644e01a6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1eb58e31-397e-4e03-8e61-a643644e01a6\") " pod="openstack/ceilometer-0" Nov 25 10:51:06 crc kubenswrapper[4821]: I1125 10:51:06.265644 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1eb58e31-397e-4e03-8e61-a643644e01a6-log-httpd\") pod \"ceilometer-0\" (UID: \"1eb58e31-397e-4e03-8e61-a643644e01a6\") " pod="openstack/ceilometer-0" Nov 25 10:51:06 crc kubenswrapper[4821]: I1125 10:51:06.267331 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1eb58e31-397e-4e03-8e61-a643644e01a6-run-httpd\") pod \"ceilometer-0\" (UID: \"1eb58e31-397e-4e03-8e61-a643644e01a6\") " pod="openstack/ceilometer-0" Nov 25 10:51:06 crc kubenswrapper[4821]: I1125 10:51:06.269621 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1eb58e31-397e-4e03-8e61-a643644e01a6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1eb58e31-397e-4e03-8e61-a643644e01a6\") " pod="openstack/ceilometer-0" Nov 25 10:51:06 crc kubenswrapper[4821]: I1125 10:51:06.269665 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1eb58e31-397e-4e03-8e61-a643644e01a6-scripts\") pod \"ceilometer-0\" (UID: \"1eb58e31-397e-4e03-8e61-a643644e01a6\") " pod="openstack/ceilometer-0" Nov 25 10:51:06 crc kubenswrapper[4821]: I1125 10:51:06.271417 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eb58e31-397e-4e03-8e61-a643644e01a6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1eb58e31-397e-4e03-8e61-a643644e01a6\") " pod="openstack/ceilometer-0" Nov 25 10:51:06 crc kubenswrapper[4821]: I1125 10:51:06.271673 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1eb58e31-397e-4e03-8e61-a643644e01a6-config-data\") pod \"ceilometer-0\" (UID: \"1eb58e31-397e-4e03-8e61-a643644e01a6\") " pod="openstack/ceilometer-0" Nov 25 10:51:06 crc kubenswrapper[4821]: I1125 10:51:06.306726 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zblt\" (UniqueName: \"kubernetes.io/projected/1eb58e31-397e-4e03-8e61-a643644e01a6-kube-api-access-4zblt\") pod \"ceilometer-0\" (UID: \"1eb58e31-397e-4e03-8e61-a643644e01a6\") " pod="openstack/ceilometer-0" Nov 25 10:51:06 crc kubenswrapper[4821]: I1125 10:51:06.392204 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:51:06 crc kubenswrapper[4821]: I1125 10:51:06.831800 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:51:06 crc kubenswrapper[4821]: W1125 10:51:06.840493 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1eb58e31_397e_4e03_8e61_a643644e01a6.slice/crio-1eb0badfcf8baab1edb54795e8051f8e5918a6a4870d5f7e9dded85d1bb1fbc8 WatchSource:0}: Error finding container 1eb0badfcf8baab1edb54795e8051f8e5918a6a4870d5f7e9dded85d1bb1fbc8: Status 404 returned error can't find the container with id 1eb0badfcf8baab1edb54795e8051f8e5918a6a4870d5f7e9dded85d1bb1fbc8 Nov 25 10:51:06 crc kubenswrapper[4821]: I1125 10:51:06.851397 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:51:06 crc kubenswrapper[4821]: I1125 10:51:06.960770 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1eb58e31-397e-4e03-8e61-a643644e01a6","Type":"ContainerStarted","Data":"1eb0badfcf8baab1edb54795e8051f8e5918a6a4870d5f7e9dded85d1bb1fbc8"} Nov 25 10:51:07 crc kubenswrapper[4821]: I1125 10:51:07.974375 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1eb58e31-397e-4e03-8e61-a643644e01a6","Type":"ContainerStarted","Data":"9f30c9daf69a4ceace013d827e3c99558f37346d8b7c0d51e47075ff8d63a573"} Nov 25 10:51:08 crc kubenswrapper[4821]: I1125 10:51:08.985145 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1eb58e31-397e-4e03-8e61-a643644e01a6","Type":"ContainerStarted","Data":"bead096b46e5a62241d7459e95a00012b6d00a413aeadd8a12ac8bdd40537fbb"} Nov 25 10:51:09 crc kubenswrapper[4821]: I1125 10:51:09.998830 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1eb58e31-397e-4e03-8e61-a643644e01a6","Type":"ContainerStarted","Data":"ab3f320ab3a0711ad2f94159f670b1611e24a6591847f7775632dcaabbc1194b"} Nov 25 10:51:12 crc kubenswrapper[4821]: I1125 10:51:12.018778 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1eb58e31-397e-4e03-8e61-a643644e01a6","Type":"ContainerStarted","Data":"8bfdd140d5558ddf2697e88ccd5aefcd6fbf01cb566994d495ba0694edafd838"} Nov 25 10:51:12 crc kubenswrapper[4821]: I1125 10:51:12.019884 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 25 10:51:12 crc kubenswrapper[4821]: I1125 10:51:12.018967 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1eb58e31-397e-4e03-8e61-a643644e01a6" containerName="proxy-httpd" containerID="cri-o://8bfdd140d5558ddf2697e88ccd5aefcd6fbf01cb566994d495ba0694edafd838" gracePeriod=30 Nov 25 10:51:12 crc kubenswrapper[4821]: I1125 10:51:12.018918 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1eb58e31-397e-4e03-8e61-a643644e01a6" containerName="ceilometer-central-agent" containerID="cri-o://9f30c9daf69a4ceace013d827e3c99558f37346d8b7c0d51e47075ff8d63a573" gracePeriod=30 Nov 25 10:51:12 crc kubenswrapper[4821]: I1125 10:51:12.019011 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1eb58e31-397e-4e03-8e61-a643644e01a6" containerName="ceilometer-notification-agent" containerID="cri-o://bead096b46e5a62241d7459e95a00012b6d00a413aeadd8a12ac8bdd40537fbb" gracePeriod=30 Nov 25 10:51:12 crc kubenswrapper[4821]: I1125 10:51:12.019037 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1eb58e31-397e-4e03-8e61-a643644e01a6" containerName="sg-core" containerID="cri-o://ab3f320ab3a0711ad2f94159f670b1611e24a6591847f7775632dcaabbc1194b" gracePeriod=30 Nov 25 10:51:12 crc kubenswrapper[4821]: I1125 10:51:12.049236 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.795839064 podStartE2EDuration="7.049212089s" podCreationTimestamp="2025-11-25 10:51:05 +0000 UTC" firstStartedPulling="2025-11-25 10:51:06.844084264 +0000 UTC m=+1137.380404111" lastFinishedPulling="2025-11-25 10:51:11.097457289 +0000 UTC m=+1141.633777136" observedRunningTime="2025-11-25 10:51:12.038142435 +0000 UTC m=+1142.574462292" watchObservedRunningTime="2025-11-25 10:51:12.049212089 +0000 UTC m=+1142.585531936" Nov 25 10:51:13 crc kubenswrapper[4821]: I1125 10:51:13.842432 4821 generic.go:334] "Generic (PLEG): container finished" podID="1eb58e31-397e-4e03-8e61-a643644e01a6" containerID="8bfdd140d5558ddf2697e88ccd5aefcd6fbf01cb566994d495ba0694edafd838" exitCode=0 Nov 25 10:51:13 crc kubenswrapper[4821]: I1125 10:51:13.842700 4821 generic.go:334] "Generic (PLEG): container finished" podID="1eb58e31-397e-4e03-8e61-a643644e01a6" containerID="ab3f320ab3a0711ad2f94159f670b1611e24a6591847f7775632dcaabbc1194b" exitCode=2 Nov 25 10:51:13 crc kubenswrapper[4821]: I1125 10:51:13.842709 4821 generic.go:334] "Generic (PLEG): container finished" podID="1eb58e31-397e-4e03-8e61-a643644e01a6" containerID="bead096b46e5a62241d7459e95a00012b6d00a413aeadd8a12ac8bdd40537fbb" exitCode=0 Nov 25 10:51:13 crc kubenswrapper[4821]: I1125 10:51:13.842507 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1eb58e31-397e-4e03-8e61-a643644e01a6","Type":"ContainerDied","Data":"8bfdd140d5558ddf2697e88ccd5aefcd6fbf01cb566994d495ba0694edafd838"} Nov 25 10:51:13 crc kubenswrapper[4821]: I1125 10:51:13.842742 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1eb58e31-397e-4e03-8e61-a643644e01a6","Type":"ContainerDied","Data":"ab3f320ab3a0711ad2f94159f670b1611e24a6591847f7775632dcaabbc1194b"} Nov 25 10:51:13 crc kubenswrapper[4821]: I1125 10:51:13.842755 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1eb58e31-397e-4e03-8e61-a643644e01a6","Type":"ContainerDied","Data":"bead096b46e5a62241d7459e95a00012b6d00a413aeadd8a12ac8bdd40537fbb"} Nov 25 10:51:14 crc kubenswrapper[4821]: I1125 10:51:14.854697 4821 generic.go:334] "Generic (PLEG): container finished" podID="1eb58e31-397e-4e03-8e61-a643644e01a6" containerID="9f30c9daf69a4ceace013d827e3c99558f37346d8b7c0d51e47075ff8d63a573" exitCode=0 Nov 25 10:51:14 crc kubenswrapper[4821]: I1125 10:51:14.854737 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1eb58e31-397e-4e03-8e61-a643644e01a6","Type":"ContainerDied","Data":"9f30c9daf69a4ceace013d827e3c99558f37346d8b7c0d51e47075ff8d63a573"} Nov 25 10:51:15 crc kubenswrapper[4821]: I1125 10:51:15.171613 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:51:15 crc kubenswrapper[4821]: I1125 10:51:15.333907 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4zblt\" (UniqueName: \"kubernetes.io/projected/1eb58e31-397e-4e03-8e61-a643644e01a6-kube-api-access-4zblt\") pod \"1eb58e31-397e-4e03-8e61-a643644e01a6\" (UID: \"1eb58e31-397e-4e03-8e61-a643644e01a6\") " Nov 25 10:51:15 crc kubenswrapper[4821]: I1125 10:51:15.334489 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1eb58e31-397e-4e03-8e61-a643644e01a6-log-httpd\") pod \"1eb58e31-397e-4e03-8e61-a643644e01a6\" (UID: \"1eb58e31-397e-4e03-8e61-a643644e01a6\") " Nov 25 10:51:15 crc kubenswrapper[4821]: I1125 10:51:15.334520 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eb58e31-397e-4e03-8e61-a643644e01a6-combined-ca-bundle\") pod \"1eb58e31-397e-4e03-8e61-a643644e01a6\" (UID: \"1eb58e31-397e-4e03-8e61-a643644e01a6\") " Nov 25 10:51:15 crc kubenswrapper[4821]: I1125 10:51:15.334602 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1eb58e31-397e-4e03-8e61-a643644e01a6-config-data\") pod \"1eb58e31-397e-4e03-8e61-a643644e01a6\" (UID: \"1eb58e31-397e-4e03-8e61-a643644e01a6\") " Nov 25 10:51:15 crc kubenswrapper[4821]: I1125 10:51:15.334630 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1eb58e31-397e-4e03-8e61-a643644e01a6-run-httpd\") pod \"1eb58e31-397e-4e03-8e61-a643644e01a6\" (UID: \"1eb58e31-397e-4e03-8e61-a643644e01a6\") " Nov 25 10:51:15 crc kubenswrapper[4821]: I1125 10:51:15.334671 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1eb58e31-397e-4e03-8e61-a643644e01a6-sg-core-conf-yaml\") pod \"1eb58e31-397e-4e03-8e61-a643644e01a6\" (UID: \"1eb58e31-397e-4e03-8e61-a643644e01a6\") " Nov 25 10:51:15 crc kubenswrapper[4821]: I1125 10:51:15.334764 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1eb58e31-397e-4e03-8e61-a643644e01a6-scripts\") pod \"1eb58e31-397e-4e03-8e61-a643644e01a6\" (UID: \"1eb58e31-397e-4e03-8e61-a643644e01a6\") " Nov 25 10:51:15 crc kubenswrapper[4821]: I1125 10:51:15.335192 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1eb58e31-397e-4e03-8e61-a643644e01a6-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "1eb58e31-397e-4e03-8e61-a643644e01a6" (UID: "1eb58e31-397e-4e03-8e61-a643644e01a6"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:51:15 crc kubenswrapper[4821]: I1125 10:51:15.335225 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1eb58e31-397e-4e03-8e61-a643644e01a6-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "1eb58e31-397e-4e03-8e61-a643644e01a6" (UID: "1eb58e31-397e-4e03-8e61-a643644e01a6"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:51:15 crc kubenswrapper[4821]: I1125 10:51:15.335909 4821 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1eb58e31-397e-4e03-8e61-a643644e01a6-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 10:51:15 crc kubenswrapper[4821]: I1125 10:51:15.335937 4821 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1eb58e31-397e-4e03-8e61-a643644e01a6-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 10:51:15 crc kubenswrapper[4821]: I1125 10:51:15.341035 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1eb58e31-397e-4e03-8e61-a643644e01a6-scripts" (OuterVolumeSpecName: "scripts") pod "1eb58e31-397e-4e03-8e61-a643644e01a6" (UID: "1eb58e31-397e-4e03-8e61-a643644e01a6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:51:15 crc kubenswrapper[4821]: I1125 10:51:15.341491 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1eb58e31-397e-4e03-8e61-a643644e01a6-kube-api-access-4zblt" (OuterVolumeSpecName: "kube-api-access-4zblt") pod "1eb58e31-397e-4e03-8e61-a643644e01a6" (UID: "1eb58e31-397e-4e03-8e61-a643644e01a6"). InnerVolumeSpecName "kube-api-access-4zblt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:51:15 crc kubenswrapper[4821]: I1125 10:51:15.362763 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1eb58e31-397e-4e03-8e61-a643644e01a6-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "1eb58e31-397e-4e03-8e61-a643644e01a6" (UID: "1eb58e31-397e-4e03-8e61-a643644e01a6"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:51:15 crc kubenswrapper[4821]: I1125 10:51:15.414743 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1eb58e31-397e-4e03-8e61-a643644e01a6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1eb58e31-397e-4e03-8e61-a643644e01a6" (UID: "1eb58e31-397e-4e03-8e61-a643644e01a6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:51:15 crc kubenswrapper[4821]: I1125 10:51:15.437109 4821 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eb58e31-397e-4e03-8e61-a643644e01a6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:51:15 crc kubenswrapper[4821]: I1125 10:51:15.437150 4821 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1eb58e31-397e-4e03-8e61-a643644e01a6-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 25 10:51:15 crc kubenswrapper[4821]: I1125 10:51:15.437172 4821 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1eb58e31-397e-4e03-8e61-a643644e01a6-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:51:15 crc kubenswrapper[4821]: I1125 10:51:15.437184 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4zblt\" (UniqueName: \"kubernetes.io/projected/1eb58e31-397e-4e03-8e61-a643644e01a6-kube-api-access-4zblt\") on node \"crc\" DevicePath \"\"" Nov 25 10:51:15 crc kubenswrapper[4821]: I1125 10:51:15.438848 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1eb58e31-397e-4e03-8e61-a643644e01a6-config-data" (OuterVolumeSpecName: "config-data") pod "1eb58e31-397e-4e03-8e61-a643644e01a6" (UID: "1eb58e31-397e-4e03-8e61-a643644e01a6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:51:15 crc kubenswrapper[4821]: I1125 10:51:15.548938 4821 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1eb58e31-397e-4e03-8e61-a643644e01a6-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:51:15 crc kubenswrapper[4821]: I1125 10:51:15.874300 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1eb58e31-397e-4e03-8e61-a643644e01a6","Type":"ContainerDied","Data":"1eb0badfcf8baab1edb54795e8051f8e5918a6a4870d5f7e9dded85d1bb1fbc8"} Nov 25 10:51:15 crc kubenswrapper[4821]: I1125 10:51:15.874339 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:51:15 crc kubenswrapper[4821]: I1125 10:51:15.874380 4821 scope.go:117] "RemoveContainer" containerID="8bfdd140d5558ddf2697e88ccd5aefcd6fbf01cb566994d495ba0694edafd838" Nov 25 10:51:15 crc kubenswrapper[4821]: I1125 10:51:15.897708 4821 scope.go:117] "RemoveContainer" containerID="ab3f320ab3a0711ad2f94159f670b1611e24a6591847f7775632dcaabbc1194b" Nov 25 10:51:15 crc kubenswrapper[4821]: I1125 10:51:15.911020 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:51:15 crc kubenswrapper[4821]: I1125 10:51:15.921042 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:51:15 crc kubenswrapper[4821]: I1125 10:51:15.943034 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:51:15 crc kubenswrapper[4821]: I1125 10:51:15.943334 4821 scope.go:117] "RemoveContainer" containerID="bead096b46e5a62241d7459e95a00012b6d00a413aeadd8a12ac8bdd40537fbb" Nov 25 10:51:15 crc kubenswrapper[4821]: E1125 10:51:15.943420 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1eb58e31-397e-4e03-8e61-a643644e01a6" containerName="sg-core" Nov 25 10:51:15 crc kubenswrapper[4821]: I1125 10:51:15.943433 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="1eb58e31-397e-4e03-8e61-a643644e01a6" containerName="sg-core" Nov 25 10:51:15 crc kubenswrapper[4821]: E1125 10:51:15.943446 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1eb58e31-397e-4e03-8e61-a643644e01a6" containerName="ceilometer-notification-agent" Nov 25 10:51:15 crc kubenswrapper[4821]: I1125 10:51:15.943452 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="1eb58e31-397e-4e03-8e61-a643644e01a6" containerName="ceilometer-notification-agent" Nov 25 10:51:15 crc kubenswrapper[4821]: E1125 10:51:15.943466 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1eb58e31-397e-4e03-8e61-a643644e01a6" containerName="ceilometer-central-agent" Nov 25 10:51:15 crc kubenswrapper[4821]: I1125 10:51:15.943472 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="1eb58e31-397e-4e03-8e61-a643644e01a6" containerName="ceilometer-central-agent" Nov 25 10:51:15 crc kubenswrapper[4821]: E1125 10:51:15.943489 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1eb58e31-397e-4e03-8e61-a643644e01a6" containerName="proxy-httpd" Nov 25 10:51:15 crc kubenswrapper[4821]: I1125 10:51:15.943494 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="1eb58e31-397e-4e03-8e61-a643644e01a6" containerName="proxy-httpd" Nov 25 10:51:15 crc kubenswrapper[4821]: I1125 10:51:15.943655 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="1eb58e31-397e-4e03-8e61-a643644e01a6" containerName="ceilometer-notification-agent" Nov 25 10:51:15 crc kubenswrapper[4821]: I1125 10:51:15.943668 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="1eb58e31-397e-4e03-8e61-a643644e01a6" containerName="ceilometer-central-agent" Nov 25 10:51:15 crc kubenswrapper[4821]: I1125 10:51:15.943678 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="1eb58e31-397e-4e03-8e61-a643644e01a6" containerName="proxy-httpd" Nov 25 10:51:15 crc kubenswrapper[4821]: I1125 10:51:15.943689 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="1eb58e31-397e-4e03-8e61-a643644e01a6" containerName="sg-core" Nov 25 10:51:15 crc kubenswrapper[4821]: I1125 10:51:15.945769 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:51:15 crc kubenswrapper[4821]: I1125 10:51:15.950536 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 25 10:51:15 crc kubenswrapper[4821]: I1125 10:51:15.950531 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 25 10:51:15 crc kubenswrapper[4821]: I1125 10:51:15.959100 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:51:15 crc kubenswrapper[4821]: I1125 10:51:15.985121 4821 scope.go:117] "RemoveContainer" containerID="9f30c9daf69a4ceace013d827e3c99558f37346d8b7c0d51e47075ff8d63a573" Nov 25 10:51:16 crc kubenswrapper[4821]: I1125 10:51:16.058400 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93ebfd94-d02a-4a94-bd08-2fa977ece0c2-scripts\") pod \"ceilometer-0\" (UID: \"93ebfd94-d02a-4a94-bd08-2fa977ece0c2\") " pod="openstack/ceilometer-0" Nov 25 10:51:16 crc kubenswrapper[4821]: I1125 10:51:16.058461 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/93ebfd94-d02a-4a94-bd08-2fa977ece0c2-run-httpd\") pod \"ceilometer-0\" (UID: \"93ebfd94-d02a-4a94-bd08-2fa977ece0c2\") " pod="openstack/ceilometer-0" Nov 25 10:51:16 crc kubenswrapper[4821]: I1125 10:51:16.058660 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/93ebfd94-d02a-4a94-bd08-2fa977ece0c2-log-httpd\") pod \"ceilometer-0\" (UID: \"93ebfd94-d02a-4a94-bd08-2fa977ece0c2\") " pod="openstack/ceilometer-0" Nov 25 10:51:16 crc kubenswrapper[4821]: I1125 10:51:16.058747 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g5fnr\" (UniqueName: \"kubernetes.io/projected/93ebfd94-d02a-4a94-bd08-2fa977ece0c2-kube-api-access-g5fnr\") pod \"ceilometer-0\" (UID: \"93ebfd94-d02a-4a94-bd08-2fa977ece0c2\") " pod="openstack/ceilometer-0" Nov 25 10:51:16 crc kubenswrapper[4821]: I1125 10:51:16.058769 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93ebfd94-d02a-4a94-bd08-2fa977ece0c2-config-data\") pod \"ceilometer-0\" (UID: \"93ebfd94-d02a-4a94-bd08-2fa977ece0c2\") " pod="openstack/ceilometer-0" Nov 25 10:51:16 crc kubenswrapper[4821]: I1125 10:51:16.058788 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93ebfd94-d02a-4a94-bd08-2fa977ece0c2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"93ebfd94-d02a-4a94-bd08-2fa977ece0c2\") " pod="openstack/ceilometer-0" Nov 25 10:51:16 crc kubenswrapper[4821]: I1125 10:51:16.058935 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/93ebfd94-d02a-4a94-bd08-2fa977ece0c2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"93ebfd94-d02a-4a94-bd08-2fa977ece0c2\") " pod="openstack/ceilometer-0" Nov 25 10:51:16 crc kubenswrapper[4821]: I1125 10:51:16.126761 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1eb58e31-397e-4e03-8e61-a643644e01a6" path="/var/lib/kubelet/pods/1eb58e31-397e-4e03-8e61-a643644e01a6/volumes" Nov 25 10:51:16 crc kubenswrapper[4821]: I1125 10:51:16.160995 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93ebfd94-d02a-4a94-bd08-2fa977ece0c2-scripts\") pod \"ceilometer-0\" (UID: \"93ebfd94-d02a-4a94-bd08-2fa977ece0c2\") " pod="openstack/ceilometer-0" Nov 25 10:51:16 crc kubenswrapper[4821]: I1125 10:51:16.161059 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/93ebfd94-d02a-4a94-bd08-2fa977ece0c2-run-httpd\") pod \"ceilometer-0\" (UID: \"93ebfd94-d02a-4a94-bd08-2fa977ece0c2\") " pod="openstack/ceilometer-0" Nov 25 10:51:16 crc kubenswrapper[4821]: I1125 10:51:16.161116 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/93ebfd94-d02a-4a94-bd08-2fa977ece0c2-log-httpd\") pod \"ceilometer-0\" (UID: \"93ebfd94-d02a-4a94-bd08-2fa977ece0c2\") " pod="openstack/ceilometer-0" Nov 25 10:51:16 crc kubenswrapper[4821]: I1125 10:51:16.161155 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g5fnr\" (UniqueName: \"kubernetes.io/projected/93ebfd94-d02a-4a94-bd08-2fa977ece0c2-kube-api-access-g5fnr\") pod \"ceilometer-0\" (UID: \"93ebfd94-d02a-4a94-bd08-2fa977ece0c2\") " pod="openstack/ceilometer-0" Nov 25 10:51:16 crc kubenswrapper[4821]: I1125 10:51:16.161189 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93ebfd94-d02a-4a94-bd08-2fa977ece0c2-config-data\") pod \"ceilometer-0\" (UID: \"93ebfd94-d02a-4a94-bd08-2fa977ece0c2\") " pod="openstack/ceilometer-0" Nov 25 10:51:16 crc kubenswrapper[4821]: I1125 10:51:16.161206 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93ebfd94-d02a-4a94-bd08-2fa977ece0c2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"93ebfd94-d02a-4a94-bd08-2fa977ece0c2\") " pod="openstack/ceilometer-0" Nov 25 10:51:16 crc kubenswrapper[4821]: I1125 10:51:16.161268 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/93ebfd94-d02a-4a94-bd08-2fa977ece0c2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"93ebfd94-d02a-4a94-bd08-2fa977ece0c2\") " pod="openstack/ceilometer-0" Nov 25 10:51:16 crc kubenswrapper[4821]: I1125 10:51:16.162195 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/93ebfd94-d02a-4a94-bd08-2fa977ece0c2-log-httpd\") pod \"ceilometer-0\" (UID: \"93ebfd94-d02a-4a94-bd08-2fa977ece0c2\") " pod="openstack/ceilometer-0" Nov 25 10:51:16 crc kubenswrapper[4821]: I1125 10:51:16.162870 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/93ebfd94-d02a-4a94-bd08-2fa977ece0c2-run-httpd\") pod \"ceilometer-0\" (UID: \"93ebfd94-d02a-4a94-bd08-2fa977ece0c2\") " pod="openstack/ceilometer-0" Nov 25 10:51:16 crc kubenswrapper[4821]: I1125 10:51:16.166343 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93ebfd94-d02a-4a94-bd08-2fa977ece0c2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"93ebfd94-d02a-4a94-bd08-2fa977ece0c2\") " pod="openstack/ceilometer-0" Nov 25 10:51:16 crc kubenswrapper[4821]: I1125 10:51:16.166534 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/93ebfd94-d02a-4a94-bd08-2fa977ece0c2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"93ebfd94-d02a-4a94-bd08-2fa977ece0c2\") " pod="openstack/ceilometer-0" Nov 25 10:51:16 crc kubenswrapper[4821]: I1125 10:51:16.171358 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93ebfd94-d02a-4a94-bd08-2fa977ece0c2-config-data\") pod \"ceilometer-0\" (UID: \"93ebfd94-d02a-4a94-bd08-2fa977ece0c2\") " pod="openstack/ceilometer-0" Nov 25 10:51:16 crc kubenswrapper[4821]: I1125 10:51:16.182705 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93ebfd94-d02a-4a94-bd08-2fa977ece0c2-scripts\") pod \"ceilometer-0\" (UID: \"93ebfd94-d02a-4a94-bd08-2fa977ece0c2\") " pod="openstack/ceilometer-0" Nov 25 10:51:16 crc kubenswrapper[4821]: I1125 10:51:16.182840 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g5fnr\" (UniqueName: \"kubernetes.io/projected/93ebfd94-d02a-4a94-bd08-2fa977ece0c2-kube-api-access-g5fnr\") pod \"ceilometer-0\" (UID: \"93ebfd94-d02a-4a94-bd08-2fa977ece0c2\") " pod="openstack/ceilometer-0" Nov 25 10:51:16 crc kubenswrapper[4821]: I1125 10:51:16.265476 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:51:16 crc kubenswrapper[4821]: I1125 10:51:16.736570 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:51:16 crc kubenswrapper[4821]: I1125 10:51:16.884961 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"93ebfd94-d02a-4a94-bd08-2fa977ece0c2","Type":"ContainerStarted","Data":"02ce8cb24bca19c9c9402a011d797633829934b981dcce8b74b6a27133ad4607"} Nov 25 10:51:17 crc kubenswrapper[4821]: I1125 10:51:17.900674 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"93ebfd94-d02a-4a94-bd08-2fa977ece0c2","Type":"ContainerStarted","Data":"58854775c62d04df46d082b0b3e917a617f69ac9824ab4befff224cb93919dfa"} Nov 25 10:51:18 crc kubenswrapper[4821]: I1125 10:51:18.910615 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"93ebfd94-d02a-4a94-bd08-2fa977ece0c2","Type":"ContainerStarted","Data":"4524ca39e676a4c2d2a0a5cd6233d1a9da964e9de331e24b23cf829c337da27a"} Nov 25 10:51:19 crc kubenswrapper[4821]: I1125 10:51:19.923098 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"93ebfd94-d02a-4a94-bd08-2fa977ece0c2","Type":"ContainerStarted","Data":"5651bfa9d025b88a03615fdbcd8d8ab5c0354dcbfa711ef2309fa1144334340c"} Nov 25 10:51:19 crc kubenswrapper[4821]: I1125 10:51:19.924717 4821 generic.go:334] "Generic (PLEG): container finished" podID="7745217e-45ba-403f-bf4f-1cbced054784" containerID="4148d293fc86b2ec5132a58cee28e829308cf0fbcd9ab995a4217bc85f1c0c50" exitCode=0 Nov 25 10:51:19 crc kubenswrapper[4821]: I1125 10:51:19.924766 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-42f5v" event={"ID":"7745217e-45ba-403f-bf4f-1cbced054784","Type":"ContainerDied","Data":"4148d293fc86b2ec5132a58cee28e829308cf0fbcd9ab995a4217bc85f1c0c50"} Nov 25 10:51:21 crc kubenswrapper[4821]: I1125 10:51:21.376834 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-42f5v" Nov 25 10:51:21 crc kubenswrapper[4821]: I1125 10:51:21.490146 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7745217e-45ba-403f-bf4f-1cbced054784-combined-ca-bundle\") pod \"7745217e-45ba-403f-bf4f-1cbced054784\" (UID: \"7745217e-45ba-403f-bf4f-1cbced054784\") " Nov 25 10:51:21 crc kubenswrapper[4821]: I1125 10:51:21.490487 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7klvp\" (UniqueName: \"kubernetes.io/projected/7745217e-45ba-403f-bf4f-1cbced054784-kube-api-access-7klvp\") pod \"7745217e-45ba-403f-bf4f-1cbced054784\" (UID: \"7745217e-45ba-403f-bf4f-1cbced054784\") " Nov 25 10:51:21 crc kubenswrapper[4821]: I1125 10:51:21.490516 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7745217e-45ba-403f-bf4f-1cbced054784-scripts\") pod \"7745217e-45ba-403f-bf4f-1cbced054784\" (UID: \"7745217e-45ba-403f-bf4f-1cbced054784\") " Nov 25 10:51:21 crc kubenswrapper[4821]: I1125 10:51:21.490609 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7745217e-45ba-403f-bf4f-1cbced054784-config-data\") pod \"7745217e-45ba-403f-bf4f-1cbced054784\" (UID: \"7745217e-45ba-403f-bf4f-1cbced054784\") " Nov 25 10:51:21 crc kubenswrapper[4821]: I1125 10:51:21.494793 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7745217e-45ba-403f-bf4f-1cbced054784-scripts" (OuterVolumeSpecName: "scripts") pod "7745217e-45ba-403f-bf4f-1cbced054784" (UID: "7745217e-45ba-403f-bf4f-1cbced054784"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:51:21 crc kubenswrapper[4821]: I1125 10:51:21.495497 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7745217e-45ba-403f-bf4f-1cbced054784-kube-api-access-7klvp" (OuterVolumeSpecName: "kube-api-access-7klvp") pod "7745217e-45ba-403f-bf4f-1cbced054784" (UID: "7745217e-45ba-403f-bf4f-1cbced054784"). InnerVolumeSpecName "kube-api-access-7klvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:51:21 crc kubenswrapper[4821]: I1125 10:51:21.518424 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7745217e-45ba-403f-bf4f-1cbced054784-config-data" (OuterVolumeSpecName: "config-data") pod "7745217e-45ba-403f-bf4f-1cbced054784" (UID: "7745217e-45ba-403f-bf4f-1cbced054784"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:51:21 crc kubenswrapper[4821]: I1125 10:51:21.521130 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7745217e-45ba-403f-bf4f-1cbced054784-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7745217e-45ba-403f-bf4f-1cbced054784" (UID: "7745217e-45ba-403f-bf4f-1cbced054784"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:51:21 crc kubenswrapper[4821]: I1125 10:51:21.592824 4821 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7745217e-45ba-403f-bf4f-1cbced054784-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:51:21 crc kubenswrapper[4821]: I1125 10:51:21.592860 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7klvp\" (UniqueName: \"kubernetes.io/projected/7745217e-45ba-403f-bf4f-1cbced054784-kube-api-access-7klvp\") on node \"crc\" DevicePath \"\"" Nov 25 10:51:21 crc kubenswrapper[4821]: I1125 10:51:21.592875 4821 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7745217e-45ba-403f-bf4f-1cbced054784-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:51:21 crc kubenswrapper[4821]: I1125 10:51:21.592886 4821 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7745217e-45ba-403f-bf4f-1cbced054784-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:51:21 crc kubenswrapper[4821]: I1125 10:51:21.946860 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-42f5v" Nov 25 10:51:21 crc kubenswrapper[4821]: I1125 10:51:21.951225 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-42f5v" event={"ID":"7745217e-45ba-403f-bf4f-1cbced054784","Type":"ContainerDied","Data":"933af00ac731f452a40fb942d646576e0d483ceeb14bb6145afa58ed1cf7498f"} Nov 25 10:51:21 crc kubenswrapper[4821]: I1125 10:51:21.951281 4821 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="933af00ac731f452a40fb942d646576e0d483ceeb14bb6145afa58ed1cf7498f" Nov 25 10:51:21 crc kubenswrapper[4821]: I1125 10:51:21.954201 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"93ebfd94-d02a-4a94-bd08-2fa977ece0c2","Type":"ContainerStarted","Data":"7f092fdc0f9a31f25a3a369e3b7e07011c4b9033d12d7a977abf4cf3161f86a1"} Nov 25 10:51:21 crc kubenswrapper[4821]: I1125 10:51:21.955644 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 25 10:51:21 crc kubenswrapper[4821]: I1125 10:51:21.993735 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.532223082 podStartE2EDuration="6.993710709s" podCreationTimestamp="2025-11-25 10:51:15 +0000 UTC" firstStartedPulling="2025-11-25 10:51:16.750625471 +0000 UTC m=+1147.286945318" lastFinishedPulling="2025-11-25 10:51:21.212113098 +0000 UTC m=+1151.748432945" observedRunningTime="2025-11-25 10:51:21.988047284 +0000 UTC m=+1152.524367161" watchObservedRunningTime="2025-11-25 10:51:21.993710709 +0000 UTC m=+1152.530030566" Nov 25 10:51:22 crc kubenswrapper[4821]: I1125 10:51:22.044264 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 25 10:51:22 crc kubenswrapper[4821]: E1125 10:51:22.060283 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7745217e-45ba-403f-bf4f-1cbced054784" containerName="nova-cell0-conductor-db-sync" Nov 25 10:51:22 crc kubenswrapper[4821]: I1125 10:51:22.060341 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="7745217e-45ba-403f-bf4f-1cbced054784" containerName="nova-cell0-conductor-db-sync" Nov 25 10:51:22 crc kubenswrapper[4821]: I1125 10:51:22.060995 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="7745217e-45ba-403f-bf4f-1cbced054784" containerName="nova-cell0-conductor-db-sync" Nov 25 10:51:22 crc kubenswrapper[4821]: I1125 10:51:22.062061 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 25 10:51:22 crc kubenswrapper[4821]: I1125 10:51:22.065208 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-865zj" Nov 25 10:51:22 crc kubenswrapper[4821]: I1125 10:51:22.065435 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 25 10:51:22 crc kubenswrapper[4821]: I1125 10:51:22.080306 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 25 10:51:22 crc kubenswrapper[4821]: I1125 10:51:22.103973 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wg5l\" (UniqueName: \"kubernetes.io/projected/e5dfcd15-af77-4be7-93db-e26f3da1d7d2-kube-api-access-4wg5l\") pod \"nova-cell0-conductor-0\" (UID: \"e5dfcd15-af77-4be7-93db-e26f3da1d7d2\") " pod="openstack/nova-cell0-conductor-0" Nov 25 10:51:22 crc kubenswrapper[4821]: I1125 10:51:22.104118 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5dfcd15-af77-4be7-93db-e26f3da1d7d2-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"e5dfcd15-af77-4be7-93db-e26f3da1d7d2\") " pod="openstack/nova-cell0-conductor-0" Nov 25 10:51:22 crc kubenswrapper[4821]: I1125 10:51:22.104156 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5dfcd15-af77-4be7-93db-e26f3da1d7d2-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"e5dfcd15-af77-4be7-93db-e26f3da1d7d2\") " pod="openstack/nova-cell0-conductor-0" Nov 25 10:51:22 crc kubenswrapper[4821]: I1125 10:51:22.206401 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wg5l\" (UniqueName: \"kubernetes.io/projected/e5dfcd15-af77-4be7-93db-e26f3da1d7d2-kube-api-access-4wg5l\") pod \"nova-cell0-conductor-0\" (UID: \"e5dfcd15-af77-4be7-93db-e26f3da1d7d2\") " pod="openstack/nova-cell0-conductor-0" Nov 25 10:51:22 crc kubenswrapper[4821]: I1125 10:51:22.207790 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5dfcd15-af77-4be7-93db-e26f3da1d7d2-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"e5dfcd15-af77-4be7-93db-e26f3da1d7d2\") " pod="openstack/nova-cell0-conductor-0" Nov 25 10:51:22 crc kubenswrapper[4821]: I1125 10:51:22.208213 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5dfcd15-af77-4be7-93db-e26f3da1d7d2-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"e5dfcd15-af77-4be7-93db-e26f3da1d7d2\") " pod="openstack/nova-cell0-conductor-0" Nov 25 10:51:22 crc kubenswrapper[4821]: I1125 10:51:22.212293 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5dfcd15-af77-4be7-93db-e26f3da1d7d2-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"e5dfcd15-af77-4be7-93db-e26f3da1d7d2\") " pod="openstack/nova-cell0-conductor-0" Nov 25 10:51:22 crc kubenswrapper[4821]: I1125 10:51:22.215504 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5dfcd15-af77-4be7-93db-e26f3da1d7d2-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"e5dfcd15-af77-4be7-93db-e26f3da1d7d2\") " pod="openstack/nova-cell0-conductor-0" Nov 25 10:51:22 crc kubenswrapper[4821]: I1125 10:51:22.220331 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wg5l\" (UniqueName: \"kubernetes.io/projected/e5dfcd15-af77-4be7-93db-e26f3da1d7d2-kube-api-access-4wg5l\") pod \"nova-cell0-conductor-0\" (UID: \"e5dfcd15-af77-4be7-93db-e26f3da1d7d2\") " pod="openstack/nova-cell0-conductor-0" Nov 25 10:51:22 crc kubenswrapper[4821]: I1125 10:51:22.400499 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 25 10:51:22 crc kubenswrapper[4821]: I1125 10:51:22.834996 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 25 10:51:22 crc kubenswrapper[4821]: W1125 10:51:22.838375 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode5dfcd15_af77_4be7_93db_e26f3da1d7d2.slice/crio-023c4aad10bdfd3a1ddf16e15f42bee6c5b7801b7a4751ed88961bb3eb80dfd2 WatchSource:0}: Error finding container 023c4aad10bdfd3a1ddf16e15f42bee6c5b7801b7a4751ed88961bb3eb80dfd2: Status 404 returned error can't find the container with id 023c4aad10bdfd3a1ddf16e15f42bee6c5b7801b7a4751ed88961bb3eb80dfd2 Nov 25 10:51:22 crc kubenswrapper[4821]: I1125 10:51:22.964587 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"e5dfcd15-af77-4be7-93db-e26f3da1d7d2","Type":"ContainerStarted","Data":"023c4aad10bdfd3a1ddf16e15f42bee6c5b7801b7a4751ed88961bb3eb80dfd2"} Nov 25 10:51:23 crc kubenswrapper[4821]: I1125 10:51:23.975017 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"e5dfcd15-af77-4be7-93db-e26f3da1d7d2","Type":"ContainerStarted","Data":"e677a033bae3ba2c9c61358595db1f2ad05f6cd3b13302d36b81434d8bab5fbf"} Nov 25 10:51:23 crc kubenswrapper[4821]: I1125 10:51:23.975532 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Nov 25 10:51:24 crc kubenswrapper[4821]: I1125 10:51:24.003560 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.003532812 podStartE2EDuration="2.003532812s" podCreationTimestamp="2025-11-25 10:51:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:51:23.988718554 +0000 UTC m=+1154.525038421" watchObservedRunningTime="2025-11-25 10:51:24.003532812 +0000 UTC m=+1154.539852689" Nov 25 10:51:32 crc kubenswrapper[4821]: I1125 10:51:32.431305 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Nov 25 10:51:32 crc kubenswrapper[4821]: I1125 10:51:32.867627 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-qwwdq"] Nov 25 10:51:32 crc kubenswrapper[4821]: I1125 10:51:32.869400 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-qwwdq" Nov 25 10:51:32 crc kubenswrapper[4821]: I1125 10:51:32.871472 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Nov 25 10:51:32 crc kubenswrapper[4821]: I1125 10:51:32.871898 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Nov 25 10:51:32 crc kubenswrapper[4821]: I1125 10:51:32.879355 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-qwwdq"] Nov 25 10:51:32 crc kubenswrapper[4821]: I1125 10:51:32.926074 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vntwq\" (UniqueName: \"kubernetes.io/projected/06c32eab-f3e4-4ebb-bb0e-0a887c0da0de-kube-api-access-vntwq\") pod \"nova-cell0-cell-mapping-qwwdq\" (UID: \"06c32eab-f3e4-4ebb-bb0e-0a887c0da0de\") " pod="openstack/nova-cell0-cell-mapping-qwwdq" Nov 25 10:51:32 crc kubenswrapper[4821]: I1125 10:51:32.926195 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06c32eab-f3e4-4ebb-bb0e-0a887c0da0de-scripts\") pod \"nova-cell0-cell-mapping-qwwdq\" (UID: \"06c32eab-f3e4-4ebb-bb0e-0a887c0da0de\") " pod="openstack/nova-cell0-cell-mapping-qwwdq" Nov 25 10:51:32 crc kubenswrapper[4821]: I1125 10:51:32.926314 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06c32eab-f3e4-4ebb-bb0e-0a887c0da0de-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-qwwdq\" (UID: \"06c32eab-f3e4-4ebb-bb0e-0a887c0da0de\") " pod="openstack/nova-cell0-cell-mapping-qwwdq" Nov 25 10:51:32 crc kubenswrapper[4821]: I1125 10:51:32.926342 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06c32eab-f3e4-4ebb-bb0e-0a887c0da0de-config-data\") pod \"nova-cell0-cell-mapping-qwwdq\" (UID: \"06c32eab-f3e4-4ebb-bb0e-0a887c0da0de\") " pod="openstack/nova-cell0-cell-mapping-qwwdq" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.036226 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06c32eab-f3e4-4ebb-bb0e-0a887c0da0de-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-qwwdq\" (UID: \"06c32eab-f3e4-4ebb-bb0e-0a887c0da0de\") " pod="openstack/nova-cell0-cell-mapping-qwwdq" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.036269 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06c32eab-f3e4-4ebb-bb0e-0a887c0da0de-config-data\") pod \"nova-cell0-cell-mapping-qwwdq\" (UID: \"06c32eab-f3e4-4ebb-bb0e-0a887c0da0de\") " pod="openstack/nova-cell0-cell-mapping-qwwdq" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.036364 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vntwq\" (UniqueName: \"kubernetes.io/projected/06c32eab-f3e4-4ebb-bb0e-0a887c0da0de-kube-api-access-vntwq\") pod \"nova-cell0-cell-mapping-qwwdq\" (UID: \"06c32eab-f3e4-4ebb-bb0e-0a887c0da0de\") " pod="openstack/nova-cell0-cell-mapping-qwwdq" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.036401 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06c32eab-f3e4-4ebb-bb0e-0a887c0da0de-scripts\") pod \"nova-cell0-cell-mapping-qwwdq\" (UID: \"06c32eab-f3e4-4ebb-bb0e-0a887c0da0de\") " pod="openstack/nova-cell0-cell-mapping-qwwdq" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.045121 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06c32eab-f3e4-4ebb-bb0e-0a887c0da0de-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-qwwdq\" (UID: \"06c32eab-f3e4-4ebb-bb0e-0a887c0da0de\") " pod="openstack/nova-cell0-cell-mapping-qwwdq" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.045957 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06c32eab-f3e4-4ebb-bb0e-0a887c0da0de-scripts\") pod \"nova-cell0-cell-mapping-qwwdq\" (UID: \"06c32eab-f3e4-4ebb-bb0e-0a887c0da0de\") " pod="openstack/nova-cell0-cell-mapping-qwwdq" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.062272 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06c32eab-f3e4-4ebb-bb0e-0a887c0da0de-config-data\") pod \"nova-cell0-cell-mapping-qwwdq\" (UID: \"06c32eab-f3e4-4ebb-bb0e-0a887c0da0de\") " pod="openstack/nova-cell0-cell-mapping-qwwdq" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.082809 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vntwq\" (UniqueName: \"kubernetes.io/projected/06c32eab-f3e4-4ebb-bb0e-0a887c0da0de-kube-api-access-vntwq\") pod \"nova-cell0-cell-mapping-qwwdq\" (UID: \"06c32eab-f3e4-4ebb-bb0e-0a887c0da0de\") " pod="openstack/nova-cell0-cell-mapping-qwwdq" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.106011 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.108988 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.111431 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.136032 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.145563 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.150402 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.187232 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.191983 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-qwwdq" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.220895 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.242115 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87987062-8fb1-4174-829c-cb35c4629905-config-data\") pod \"nova-api-0\" (UID: \"87987062-8fb1-4174-829c-cb35c4629905\") " pod="openstack/nova-api-0" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.242406 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8khc\" (UniqueName: \"kubernetes.io/projected/2bd16d9e-1992-4537-8376-c8d0da2f4a31-kube-api-access-f8khc\") pod \"nova-scheduler-0\" (UID: \"2bd16d9e-1992-4537-8376-c8d0da2f4a31\") " pod="openstack/nova-scheduler-0" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.242575 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bd16d9e-1992-4537-8376-c8d0da2f4a31-config-data\") pod \"nova-scheduler-0\" (UID: \"2bd16d9e-1992-4537-8376-c8d0da2f4a31\") " pod="openstack/nova-scheduler-0" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.242688 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2djw\" (UniqueName: \"kubernetes.io/projected/87987062-8fb1-4174-829c-cb35c4629905-kube-api-access-n2djw\") pod \"nova-api-0\" (UID: \"87987062-8fb1-4174-829c-cb35c4629905\") " pod="openstack/nova-api-0" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.242767 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bd16d9e-1992-4537-8376-c8d0da2f4a31-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"2bd16d9e-1992-4537-8376-c8d0da2f4a31\") " pod="openstack/nova-scheduler-0" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.242868 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87987062-8fb1-4174-829c-cb35c4629905-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"87987062-8fb1-4174-829c-cb35c4629905\") " pod="openstack/nova-api-0" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.242943 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87987062-8fb1-4174-829c-cb35c4629905-logs\") pod \"nova-api-0\" (UID: \"87987062-8fb1-4174-829c-cb35c4629905\") " pod="openstack/nova-api-0" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.354200 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bd16d9e-1992-4537-8376-c8d0da2f4a31-config-data\") pod \"nova-scheduler-0\" (UID: \"2bd16d9e-1992-4537-8376-c8d0da2f4a31\") " pod="openstack/nova-scheduler-0" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.354290 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n2djw\" (UniqueName: \"kubernetes.io/projected/87987062-8fb1-4174-829c-cb35c4629905-kube-api-access-n2djw\") pod \"nova-api-0\" (UID: \"87987062-8fb1-4174-829c-cb35c4629905\") " pod="openstack/nova-api-0" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.366222 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.369460 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.375313 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bd16d9e-1992-4537-8376-c8d0da2f4a31-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"2bd16d9e-1992-4537-8376-c8d0da2f4a31\") " pod="openstack/nova-scheduler-0" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.375489 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87987062-8fb1-4174-829c-cb35c4629905-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"87987062-8fb1-4174-829c-cb35c4629905\") " pod="openstack/nova-api-0" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.375533 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87987062-8fb1-4174-829c-cb35c4629905-logs\") pod \"nova-api-0\" (UID: \"87987062-8fb1-4174-829c-cb35c4629905\") " pod="openstack/nova-api-0" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.375583 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87987062-8fb1-4174-829c-cb35c4629905-config-data\") pod \"nova-api-0\" (UID: \"87987062-8fb1-4174-829c-cb35c4629905\") " pod="openstack/nova-api-0" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.375610 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8khc\" (UniqueName: \"kubernetes.io/projected/2bd16d9e-1992-4537-8376-c8d0da2f4a31-kube-api-access-f8khc\") pod \"nova-scheduler-0\" (UID: \"2bd16d9e-1992-4537-8376-c8d0da2f4a31\") " pod="openstack/nova-scheduler-0" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.378428 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87987062-8fb1-4174-829c-cb35c4629905-logs\") pod \"nova-api-0\" (UID: \"87987062-8fb1-4174-829c-cb35c4629905\") " pod="openstack/nova-api-0" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.392927 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.401202 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.407854 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87987062-8fb1-4174-829c-cb35c4629905-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"87987062-8fb1-4174-829c-cb35c4629905\") " pod="openstack/nova-api-0" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.410372 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8khc\" (UniqueName: \"kubernetes.io/projected/2bd16d9e-1992-4537-8376-c8d0da2f4a31-kube-api-access-f8khc\") pod \"nova-scheduler-0\" (UID: \"2bd16d9e-1992-4537-8376-c8d0da2f4a31\") " pod="openstack/nova-scheduler-0" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.432247 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bd16d9e-1992-4537-8376-c8d0da2f4a31-config-data\") pod \"nova-scheduler-0\" (UID: \"2bd16d9e-1992-4537-8376-c8d0da2f4a31\") " pod="openstack/nova-scheduler-0" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.437668 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bd16d9e-1992-4537-8376-c8d0da2f4a31-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"2bd16d9e-1992-4537-8376-c8d0da2f4a31\") " pod="openstack/nova-scheduler-0" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.438476 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87987062-8fb1-4174-829c-cb35c4629905-config-data\") pod \"nova-api-0\" (UID: \"87987062-8fb1-4174-829c-cb35c4629905\") " pod="openstack/nova-api-0" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.439344 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.440234 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.448866 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.453019 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.453141 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2djw\" (UniqueName: \"kubernetes.io/projected/87987062-8fb1-4174-829c-cb35c4629905-kube-api-access-n2djw\") pod \"nova-api-0\" (UID: \"87987062-8fb1-4174-829c-cb35c4629905\") " pod="openstack/nova-api-0" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.490332 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b044e354-e28e-43c0-ba43-bb51e8a21761-logs\") pod \"nova-metadata-0\" (UID: \"b044e354-e28e-43c0-ba43-bb51e8a21761\") " pod="openstack/nova-metadata-0" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.490523 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b044e354-e28e-43c0-ba43-bb51e8a21761-config-data\") pod \"nova-metadata-0\" (UID: \"b044e354-e28e-43c0-ba43-bb51e8a21761\") " pod="openstack/nova-metadata-0" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.500332 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b044e354-e28e-43c0-ba43-bb51e8a21761-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b044e354-e28e-43c0-ba43-bb51e8a21761\") " pod="openstack/nova-metadata-0" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.500585 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79qdb\" (UniqueName: \"kubernetes.io/projected/b044e354-e28e-43c0-ba43-bb51e8a21761-kube-api-access-79qdb\") pod \"nova-metadata-0\" (UID: \"b044e354-e28e-43c0-ba43-bb51e8a21761\") " pod="openstack/nova-metadata-0" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.510449 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.513840 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-wsvhq"] Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.517495 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-wsvhq" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.535080 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-wsvhq"] Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.564898 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.602538 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b044e354-e28e-43c0-ba43-bb51e8a21761-config-data\") pod \"nova-metadata-0\" (UID: \"b044e354-e28e-43c0-ba43-bb51e8a21761\") " pod="openstack/nova-metadata-0" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.602623 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b044e354-e28e-43c0-ba43-bb51e8a21761-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b044e354-e28e-43c0-ba43-bb51e8a21761\") " pod="openstack/nova-metadata-0" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.602658 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98e78fd5-b460-436b-990a-fdf91c8d95c9-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"98e78fd5-b460-436b-990a-fdf91c8d95c9\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.602735 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98e78fd5-b460-436b-990a-fdf91c8d95c9-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"98e78fd5-b460-436b-990a-fdf91c8d95c9\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.602773 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79qdb\" (UniqueName: \"kubernetes.io/projected/b044e354-e28e-43c0-ba43-bb51e8a21761-kube-api-access-79qdb\") pod \"nova-metadata-0\" (UID: \"b044e354-e28e-43c0-ba43-bb51e8a21761\") " pod="openstack/nova-metadata-0" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.602802 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tvnnv\" (UniqueName: \"kubernetes.io/projected/98e78fd5-b460-436b-990a-fdf91c8d95c9-kube-api-access-tvnnv\") pod \"nova-cell1-novncproxy-0\" (UID: \"98e78fd5-b460-436b-990a-fdf91c8d95c9\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.602836 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b044e354-e28e-43c0-ba43-bb51e8a21761-logs\") pod \"nova-metadata-0\" (UID: \"b044e354-e28e-43c0-ba43-bb51e8a21761\") " pod="openstack/nova-metadata-0" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.603377 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b044e354-e28e-43c0-ba43-bb51e8a21761-logs\") pod \"nova-metadata-0\" (UID: \"b044e354-e28e-43c0-ba43-bb51e8a21761\") " pod="openstack/nova-metadata-0" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.608790 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b044e354-e28e-43c0-ba43-bb51e8a21761-config-data\") pod \"nova-metadata-0\" (UID: \"b044e354-e28e-43c0-ba43-bb51e8a21761\") " pod="openstack/nova-metadata-0" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.610725 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b044e354-e28e-43c0-ba43-bb51e8a21761-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b044e354-e28e-43c0-ba43-bb51e8a21761\") " pod="openstack/nova-metadata-0" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.704827 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/77c23b7d-3861-4de3-b70f-d109bd949332-config\") pod \"dnsmasq-dns-bccf8f775-wsvhq\" (UID: \"77c23b7d-3861-4de3-b70f-d109bd949332\") " pod="openstack/dnsmasq-dns-bccf8f775-wsvhq" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.704899 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98e78fd5-b460-436b-990a-fdf91c8d95c9-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"98e78fd5-b460-436b-990a-fdf91c8d95c9\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.704925 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/77c23b7d-3861-4de3-b70f-d109bd949332-ovsdbserver-sb\") pod \"dnsmasq-dns-bccf8f775-wsvhq\" (UID: \"77c23b7d-3861-4de3-b70f-d109bd949332\") " pod="openstack/dnsmasq-dns-bccf8f775-wsvhq" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.704974 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59dq7\" (UniqueName: \"kubernetes.io/projected/77c23b7d-3861-4de3-b70f-d109bd949332-kube-api-access-59dq7\") pod \"dnsmasq-dns-bccf8f775-wsvhq\" (UID: \"77c23b7d-3861-4de3-b70f-d109bd949332\") " pod="openstack/dnsmasq-dns-bccf8f775-wsvhq" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.705022 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98e78fd5-b460-436b-990a-fdf91c8d95c9-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"98e78fd5-b460-436b-990a-fdf91c8d95c9\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.705068 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tvnnv\" (UniqueName: \"kubernetes.io/projected/98e78fd5-b460-436b-990a-fdf91c8d95c9-kube-api-access-tvnnv\") pod \"nova-cell1-novncproxy-0\" (UID: \"98e78fd5-b460-436b-990a-fdf91c8d95c9\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.705144 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/77c23b7d-3861-4de3-b70f-d109bd949332-dns-swift-storage-0\") pod \"dnsmasq-dns-bccf8f775-wsvhq\" (UID: \"77c23b7d-3861-4de3-b70f-d109bd949332\") " pod="openstack/dnsmasq-dns-bccf8f775-wsvhq" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.705202 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/77c23b7d-3861-4de3-b70f-d109bd949332-ovsdbserver-nb\") pod \"dnsmasq-dns-bccf8f775-wsvhq\" (UID: \"77c23b7d-3861-4de3-b70f-d109bd949332\") " pod="openstack/dnsmasq-dns-bccf8f775-wsvhq" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.705238 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/77c23b7d-3861-4de3-b70f-d109bd949332-dns-svc\") pod \"dnsmasq-dns-bccf8f775-wsvhq\" (UID: \"77c23b7d-3861-4de3-b70f-d109bd949332\") " pod="openstack/dnsmasq-dns-bccf8f775-wsvhq" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.806730 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/77c23b7d-3861-4de3-b70f-d109bd949332-dns-swift-storage-0\") pod \"dnsmasq-dns-bccf8f775-wsvhq\" (UID: \"77c23b7d-3861-4de3-b70f-d109bd949332\") " pod="openstack/dnsmasq-dns-bccf8f775-wsvhq" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.806798 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/77c23b7d-3861-4de3-b70f-d109bd949332-ovsdbserver-nb\") pod \"dnsmasq-dns-bccf8f775-wsvhq\" (UID: \"77c23b7d-3861-4de3-b70f-d109bd949332\") " pod="openstack/dnsmasq-dns-bccf8f775-wsvhq" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.806839 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/77c23b7d-3861-4de3-b70f-d109bd949332-dns-svc\") pod \"dnsmasq-dns-bccf8f775-wsvhq\" (UID: \"77c23b7d-3861-4de3-b70f-d109bd949332\") " pod="openstack/dnsmasq-dns-bccf8f775-wsvhq" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.807857 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/77c23b7d-3861-4de3-b70f-d109bd949332-dns-swift-storage-0\") pod \"dnsmasq-dns-bccf8f775-wsvhq\" (UID: \"77c23b7d-3861-4de3-b70f-d109bd949332\") " pod="openstack/dnsmasq-dns-bccf8f775-wsvhq" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.808234 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/77c23b7d-3861-4de3-b70f-d109bd949332-dns-svc\") pod \"dnsmasq-dns-bccf8f775-wsvhq\" (UID: \"77c23b7d-3861-4de3-b70f-d109bd949332\") " pod="openstack/dnsmasq-dns-bccf8f775-wsvhq" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.808348 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/77c23b7d-3861-4de3-b70f-d109bd949332-config\") pod \"dnsmasq-dns-bccf8f775-wsvhq\" (UID: \"77c23b7d-3861-4de3-b70f-d109bd949332\") " pod="openstack/dnsmasq-dns-bccf8f775-wsvhq" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.808527 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/77c23b7d-3861-4de3-b70f-d109bd949332-ovsdbserver-sb\") pod \"dnsmasq-dns-bccf8f775-wsvhq\" (UID: \"77c23b7d-3861-4de3-b70f-d109bd949332\") " pod="openstack/dnsmasq-dns-bccf8f775-wsvhq" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.808649 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59dq7\" (UniqueName: \"kubernetes.io/projected/77c23b7d-3861-4de3-b70f-d109bd949332-kube-api-access-59dq7\") pod \"dnsmasq-dns-bccf8f775-wsvhq\" (UID: \"77c23b7d-3861-4de3-b70f-d109bd949332\") " pod="openstack/dnsmasq-dns-bccf8f775-wsvhq" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.849311 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79qdb\" (UniqueName: \"kubernetes.io/projected/b044e354-e28e-43c0-ba43-bb51e8a21761-kube-api-access-79qdb\") pod \"nova-metadata-0\" (UID: \"b044e354-e28e-43c0-ba43-bb51e8a21761\") " pod="openstack/nova-metadata-0" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.891353 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-qwwdq"] Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.906005 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/77c23b7d-3861-4de3-b70f-d109bd949332-config\") pod \"dnsmasq-dns-bccf8f775-wsvhq\" (UID: \"77c23b7d-3861-4de3-b70f-d109bd949332\") " pod="openstack/dnsmasq-dns-bccf8f775-wsvhq" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.906011 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98e78fd5-b460-436b-990a-fdf91c8d95c9-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"98e78fd5-b460-436b-990a-fdf91c8d95c9\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.906199 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98e78fd5-b460-436b-990a-fdf91c8d95c9-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"98e78fd5-b460-436b-990a-fdf91c8d95c9\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.909035 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tvnnv\" (UniqueName: \"kubernetes.io/projected/98e78fd5-b460-436b-990a-fdf91c8d95c9-kube-api-access-tvnnv\") pod \"nova-cell1-novncproxy-0\" (UID: \"98e78fd5-b460-436b-990a-fdf91c8d95c9\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.911297 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/77c23b7d-3861-4de3-b70f-d109bd949332-ovsdbserver-nb\") pod \"dnsmasq-dns-bccf8f775-wsvhq\" (UID: \"77c23b7d-3861-4de3-b70f-d109bd949332\") " pod="openstack/dnsmasq-dns-bccf8f775-wsvhq" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.915095 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/77c23b7d-3861-4de3-b70f-d109bd949332-ovsdbserver-sb\") pod \"dnsmasq-dns-bccf8f775-wsvhq\" (UID: \"77c23b7d-3861-4de3-b70f-d109bd949332\") " pod="openstack/dnsmasq-dns-bccf8f775-wsvhq" Nov 25 10:51:33 crc kubenswrapper[4821]: I1125 10:51:33.920298 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59dq7\" (UniqueName: \"kubernetes.io/projected/77c23b7d-3861-4de3-b70f-d109bd949332-kube-api-access-59dq7\") pod \"dnsmasq-dns-bccf8f775-wsvhq\" (UID: \"77c23b7d-3861-4de3-b70f-d109bd949332\") " pod="openstack/dnsmasq-dns-bccf8f775-wsvhq" Nov 25 10:51:34 crc kubenswrapper[4821]: I1125 10:51:34.044334 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 10:51:34 crc kubenswrapper[4821]: W1125 10:51:34.058629 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod87987062_8fb1_4174_829c_cb35c4629905.slice/crio-efb6210df717a08597b9986d49cbfcc2066330ec7e37ba6f2f136d0fd85eff68 WatchSource:0}: Error finding container efb6210df717a08597b9986d49cbfcc2066330ec7e37ba6f2f136d0fd85eff68: Status 404 returned error can't find the container with id efb6210df717a08597b9986d49cbfcc2066330ec7e37ba6f2f136d0fd85eff68 Nov 25 10:51:34 crc kubenswrapper[4821]: I1125 10:51:34.060473 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 25 10:51:34 crc kubenswrapper[4821]: I1125 10:51:34.068621 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-vwzcd"] Nov 25 10:51:34 crc kubenswrapper[4821]: I1125 10:51:34.070583 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-vwzcd" Nov 25 10:51:34 crc kubenswrapper[4821]: I1125 10:51:34.073790 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Nov 25 10:51:34 crc kubenswrapper[4821]: I1125 10:51:34.074216 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 25 10:51:34 crc kubenswrapper[4821]: I1125 10:51:34.083728 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-vwzcd"] Nov 25 10:51:34 crc kubenswrapper[4821]: I1125 10:51:34.099245 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2bd16d9e-1992-4537-8376-c8d0da2f4a31","Type":"ContainerStarted","Data":"aa15114ab525d62b1ea2b5ff34acc1545c341fa46702abcf7db24a155e6e0438"} Nov 25 10:51:34 crc kubenswrapper[4821]: I1125 10:51:34.102758 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-qwwdq" event={"ID":"06c32eab-f3e4-4ebb-bb0e-0a887c0da0de","Type":"ContainerStarted","Data":"dda66e472de86b2c228de028bbf2d6e82bd8bc2c20000428e352be8a20b1f0c3"} Nov 25 10:51:34 crc kubenswrapper[4821]: I1125 10:51:34.104710 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"87987062-8fb1-4174-829c-cb35c4629905","Type":"ContainerStarted","Data":"efb6210df717a08597b9986d49cbfcc2066330ec7e37ba6f2f136d0fd85eff68"} Nov 25 10:51:34 crc kubenswrapper[4821]: I1125 10:51:34.126637 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 10:51:34 crc kubenswrapper[4821]: I1125 10:51:34.156082 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-wsvhq" Nov 25 10:51:34 crc kubenswrapper[4821]: I1125 10:51:34.159589 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:51:34 crc kubenswrapper[4821]: I1125 10:51:34.216289 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e22e60b5-5698-493b-9a3f-8cfdf289c6d9-config-data\") pod \"nova-cell1-conductor-db-sync-vwzcd\" (UID: \"e22e60b5-5698-493b-9a3f-8cfdf289c6d9\") " pod="openstack/nova-cell1-conductor-db-sync-vwzcd" Nov 25 10:51:34 crc kubenswrapper[4821]: I1125 10:51:34.216326 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e22e60b5-5698-493b-9a3f-8cfdf289c6d9-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-vwzcd\" (UID: \"e22e60b5-5698-493b-9a3f-8cfdf289c6d9\") " pod="openstack/nova-cell1-conductor-db-sync-vwzcd" Nov 25 10:51:34 crc kubenswrapper[4821]: I1125 10:51:34.216421 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9h9w\" (UniqueName: \"kubernetes.io/projected/e22e60b5-5698-493b-9a3f-8cfdf289c6d9-kube-api-access-f9h9w\") pod \"nova-cell1-conductor-db-sync-vwzcd\" (UID: \"e22e60b5-5698-493b-9a3f-8cfdf289c6d9\") " pod="openstack/nova-cell1-conductor-db-sync-vwzcd" Nov 25 10:51:34 crc kubenswrapper[4821]: I1125 10:51:34.216458 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e22e60b5-5698-493b-9a3f-8cfdf289c6d9-scripts\") pod \"nova-cell1-conductor-db-sync-vwzcd\" (UID: \"e22e60b5-5698-493b-9a3f-8cfdf289c6d9\") " pod="openstack/nova-cell1-conductor-db-sync-vwzcd" Nov 25 10:51:34 crc kubenswrapper[4821]: I1125 10:51:34.318402 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e22e60b5-5698-493b-9a3f-8cfdf289c6d9-config-data\") pod \"nova-cell1-conductor-db-sync-vwzcd\" (UID: \"e22e60b5-5698-493b-9a3f-8cfdf289c6d9\") " pod="openstack/nova-cell1-conductor-db-sync-vwzcd" Nov 25 10:51:34 crc kubenswrapper[4821]: I1125 10:51:34.318822 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e22e60b5-5698-493b-9a3f-8cfdf289c6d9-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-vwzcd\" (UID: \"e22e60b5-5698-493b-9a3f-8cfdf289c6d9\") " pod="openstack/nova-cell1-conductor-db-sync-vwzcd" Nov 25 10:51:34 crc kubenswrapper[4821]: I1125 10:51:34.318934 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9h9w\" (UniqueName: \"kubernetes.io/projected/e22e60b5-5698-493b-9a3f-8cfdf289c6d9-kube-api-access-f9h9w\") pod \"nova-cell1-conductor-db-sync-vwzcd\" (UID: \"e22e60b5-5698-493b-9a3f-8cfdf289c6d9\") " pod="openstack/nova-cell1-conductor-db-sync-vwzcd" Nov 25 10:51:34 crc kubenswrapper[4821]: I1125 10:51:34.318976 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e22e60b5-5698-493b-9a3f-8cfdf289c6d9-scripts\") pod \"nova-cell1-conductor-db-sync-vwzcd\" (UID: \"e22e60b5-5698-493b-9a3f-8cfdf289c6d9\") " pod="openstack/nova-cell1-conductor-db-sync-vwzcd" Nov 25 10:51:34 crc kubenswrapper[4821]: I1125 10:51:34.329014 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e22e60b5-5698-493b-9a3f-8cfdf289c6d9-scripts\") pod \"nova-cell1-conductor-db-sync-vwzcd\" (UID: \"e22e60b5-5698-493b-9a3f-8cfdf289c6d9\") " pod="openstack/nova-cell1-conductor-db-sync-vwzcd" Nov 25 10:51:34 crc kubenswrapper[4821]: I1125 10:51:34.329468 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e22e60b5-5698-493b-9a3f-8cfdf289c6d9-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-vwzcd\" (UID: \"e22e60b5-5698-493b-9a3f-8cfdf289c6d9\") " pod="openstack/nova-cell1-conductor-db-sync-vwzcd" Nov 25 10:51:34 crc kubenswrapper[4821]: I1125 10:51:34.331191 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e22e60b5-5698-493b-9a3f-8cfdf289c6d9-config-data\") pod \"nova-cell1-conductor-db-sync-vwzcd\" (UID: \"e22e60b5-5698-493b-9a3f-8cfdf289c6d9\") " pod="openstack/nova-cell1-conductor-db-sync-vwzcd" Nov 25 10:51:34 crc kubenswrapper[4821]: I1125 10:51:34.339154 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9h9w\" (UniqueName: \"kubernetes.io/projected/e22e60b5-5698-493b-9a3f-8cfdf289c6d9-kube-api-access-f9h9w\") pod \"nova-cell1-conductor-db-sync-vwzcd\" (UID: \"e22e60b5-5698-493b-9a3f-8cfdf289c6d9\") " pod="openstack/nova-cell1-conductor-db-sync-vwzcd" Nov 25 10:51:34 crc kubenswrapper[4821]: I1125 10:51:34.587243 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-vwzcd" Nov 25 10:51:34 crc kubenswrapper[4821]: I1125 10:51:34.601251 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 10:51:34 crc kubenswrapper[4821]: I1125 10:51:34.689723 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 25 10:51:34 crc kubenswrapper[4821]: I1125 10:51:34.702070 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-wsvhq"] Nov 25 10:51:34 crc kubenswrapper[4821]: I1125 10:51:34.715774 4821 patch_prober.go:28] interesting pod/machine-config-daemon-2krbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:51:34 crc kubenswrapper[4821]: I1125 10:51:34.715829 4821 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:51:35 crc kubenswrapper[4821]: I1125 10:51:35.063726 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-vwzcd"] Nov 25 10:51:35 crc kubenswrapper[4821]: W1125 10:51:35.068191 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode22e60b5_5698_493b_9a3f_8cfdf289c6d9.slice/crio-511a20d2922d8228a6ff5ae595c75af89e038e1d037dead5400c48d1c09b4edb WatchSource:0}: Error finding container 511a20d2922d8228a6ff5ae595c75af89e038e1d037dead5400c48d1c09b4edb: Status 404 returned error can't find the container with id 511a20d2922d8228a6ff5ae595c75af89e038e1d037dead5400c48d1c09b4edb Nov 25 10:51:35 crc kubenswrapper[4821]: I1125 10:51:35.121886 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-qwwdq" event={"ID":"06c32eab-f3e4-4ebb-bb0e-0a887c0da0de","Type":"ContainerStarted","Data":"7cf37449c9b799d7e943c3e4748fcbac868542b3396da366d9c1cfbfb5ba0a26"} Nov 25 10:51:35 crc kubenswrapper[4821]: I1125 10:51:35.123104 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b044e354-e28e-43c0-ba43-bb51e8a21761","Type":"ContainerStarted","Data":"526a6ad2288f544d51b53c64077bbf3622a1a7297731400ef71216736743a1cd"} Nov 25 10:51:35 crc kubenswrapper[4821]: I1125 10:51:35.124309 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-vwzcd" event={"ID":"e22e60b5-5698-493b-9a3f-8cfdf289c6d9","Type":"ContainerStarted","Data":"511a20d2922d8228a6ff5ae595c75af89e038e1d037dead5400c48d1c09b4edb"} Nov 25 10:51:35 crc kubenswrapper[4821]: I1125 10:51:35.126632 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-wsvhq" event={"ID":"77c23b7d-3861-4de3-b70f-d109bd949332","Type":"ContainerStarted","Data":"487c366b60dacb4ec493cea4a8993ad585a343d876f19bb13ce7f26915f559e1"} Nov 25 10:51:35 crc kubenswrapper[4821]: I1125 10:51:35.126662 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-wsvhq" event={"ID":"77c23b7d-3861-4de3-b70f-d109bd949332","Type":"ContainerStarted","Data":"f8273fbd9ec7b7cbec8d0714eb38beea4a7567dfd2c5c6998b30a3df8b54e6f2"} Nov 25 10:51:35 crc kubenswrapper[4821]: I1125 10:51:35.127998 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"98e78fd5-b460-436b-990a-fdf91c8d95c9","Type":"ContainerStarted","Data":"f41e01f314cebee4e04c54d4906c22432d231b7630752977b6aa0bf358721afa"} Nov 25 10:51:36 crc kubenswrapper[4821]: I1125 10:51:36.153443 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-vwzcd" event={"ID":"e22e60b5-5698-493b-9a3f-8cfdf289c6d9","Type":"ContainerStarted","Data":"0c2fa815c3e17265e814630a7030173763f041b2d9166a413539261379e799be"} Nov 25 10:51:36 crc kubenswrapper[4821]: I1125 10:51:36.160391 4821 generic.go:334] "Generic (PLEG): container finished" podID="77c23b7d-3861-4de3-b70f-d109bd949332" containerID="487c366b60dacb4ec493cea4a8993ad585a343d876f19bb13ce7f26915f559e1" exitCode=0 Nov 25 10:51:36 crc kubenswrapper[4821]: I1125 10:51:36.160487 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-wsvhq" event={"ID":"77c23b7d-3861-4de3-b70f-d109bd949332","Type":"ContainerDied","Data":"487c366b60dacb4ec493cea4a8993ad585a343d876f19bb13ce7f26915f559e1"} Nov 25 10:51:36 crc kubenswrapper[4821]: I1125 10:51:36.178053 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-vwzcd" podStartSLOduration=2.178036557 podStartE2EDuration="2.178036557s" podCreationTimestamp="2025-11-25 10:51:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:51:36.171651983 +0000 UTC m=+1166.707971830" watchObservedRunningTime="2025-11-25 10:51:36.178036557 +0000 UTC m=+1166.714356404" Nov 25 10:51:36 crc kubenswrapper[4821]: I1125 10:51:36.214454 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-qwwdq" podStartSLOduration=4.21443683 podStartE2EDuration="4.21443683s" podCreationTimestamp="2025-11-25 10:51:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:51:36.212555411 +0000 UTC m=+1166.748875288" watchObservedRunningTime="2025-11-25 10:51:36.21443683 +0000 UTC m=+1166.750756677" Nov 25 10:51:36 crc kubenswrapper[4821]: I1125 10:51:36.767039 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 25 10:51:36 crc kubenswrapper[4821]: I1125 10:51:36.779295 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 10:51:43 crc kubenswrapper[4821]: I1125 10:51:43.228882 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2bd16d9e-1992-4537-8376-c8d0da2f4a31","Type":"ContainerStarted","Data":"69bd9ea9d864923a312714582044157d74603cd071907d70637996049c8c2904"} Nov 25 10:51:43 crc kubenswrapper[4821]: I1125 10:51:43.231893 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b044e354-e28e-43c0-ba43-bb51e8a21761","Type":"ContainerStarted","Data":"96b7eb0d2e5a8a8aa0bfb1dcf6c3663a69b3ad48e3c9efc121471d2b781c58e3"} Nov 25 10:51:43 crc kubenswrapper[4821]: I1125 10:51:43.233720 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"87987062-8fb1-4174-829c-cb35c4629905","Type":"ContainerStarted","Data":"150ef83d8ae7669f561e8aff63be4c4ade17d7f79c71d5a4bbfd49ed6bd39a1b"} Nov 25 10:51:43 crc kubenswrapper[4821]: I1125 10:51:43.238465 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-wsvhq" event={"ID":"77c23b7d-3861-4de3-b70f-d109bd949332","Type":"ContainerStarted","Data":"751a176c0aa45d407c93fde3d89d4d41fa1f9c1889e805cc20cb48e901ba8d91"} Nov 25 10:51:43 crc kubenswrapper[4821]: I1125 10:51:43.238581 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-bccf8f775-wsvhq" Nov 25 10:51:43 crc kubenswrapper[4821]: I1125 10:51:43.242376 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"98e78fd5-b460-436b-990a-fdf91c8d95c9","Type":"ContainerStarted","Data":"a1f49d02914e3f07788cfd9a71fcc19a3a39bc6a75bc5c0637a3e999d698bac2"} Nov 25 10:51:43 crc kubenswrapper[4821]: I1125 10:51:43.242469 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="98e78fd5-b460-436b-990a-fdf91c8d95c9" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://a1f49d02914e3f07788cfd9a71fcc19a3a39bc6a75bc5c0637a3e999d698bac2" gracePeriod=30 Nov 25 10:51:43 crc kubenswrapper[4821]: I1125 10:51:43.252833 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=1.920425583 podStartE2EDuration="10.252815035s" podCreationTimestamp="2025-11-25 10:51:33 +0000 UTC" firstStartedPulling="2025-11-25 10:51:34.081387599 +0000 UTC m=+1164.617707446" lastFinishedPulling="2025-11-25 10:51:42.413777061 +0000 UTC m=+1172.950096898" observedRunningTime="2025-11-25 10:51:43.247973698 +0000 UTC m=+1173.784293545" watchObservedRunningTime="2025-11-25 10:51:43.252815035 +0000 UTC m=+1173.789134882" Nov 25 10:51:43 crc kubenswrapper[4821]: I1125 10:51:43.273509 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.585061692 podStartE2EDuration="10.273488369s" podCreationTimestamp="2025-11-25 10:51:33 +0000 UTC" firstStartedPulling="2025-11-25 10:51:34.725024166 +0000 UTC m=+1165.261344013" lastFinishedPulling="2025-11-25 10:51:42.413450803 +0000 UTC m=+1172.949770690" observedRunningTime="2025-11-25 10:51:43.261516925 +0000 UTC m=+1173.797836772" watchObservedRunningTime="2025-11-25 10:51:43.273488369 +0000 UTC m=+1173.809808216" Nov 25 10:51:43 crc kubenswrapper[4821]: I1125 10:51:43.299010 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-bccf8f775-wsvhq" podStartSLOduration=10.298992049 podStartE2EDuration="10.298992049s" podCreationTimestamp="2025-11-25 10:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:51:43.29369532 +0000 UTC m=+1173.830015177" watchObservedRunningTime="2025-11-25 10:51:43.298992049 +0000 UTC m=+1173.835311896" Nov 25 10:51:43 crc kubenswrapper[4821]: I1125 10:51:43.566411 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 25 10:51:43 crc kubenswrapper[4821]: I1125 10:51:43.566461 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 25 10:51:43 crc kubenswrapper[4821]: I1125 10:51:43.594673 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 25 10:51:44 crc kubenswrapper[4821]: I1125 10:51:44.160917 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:51:44 crc kubenswrapper[4821]: I1125 10:51:44.260946 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b044e354-e28e-43c0-ba43-bb51e8a21761","Type":"ContainerStarted","Data":"9dc278880a147a8959c8dae238461baf2c4e42d54c0ed542c1d08aa498941b9a"} Nov 25 10:51:44 crc kubenswrapper[4821]: I1125 10:51:44.261404 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="b044e354-e28e-43c0-ba43-bb51e8a21761" containerName="nova-metadata-log" containerID="cri-o://96b7eb0d2e5a8a8aa0bfb1dcf6c3663a69b3ad48e3c9efc121471d2b781c58e3" gracePeriod=30 Nov 25 10:51:44 crc kubenswrapper[4821]: I1125 10:51:44.262078 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="b044e354-e28e-43c0-ba43-bb51e8a21761" containerName="nova-metadata-metadata" containerID="cri-o://9dc278880a147a8959c8dae238461baf2c4e42d54c0ed542c1d08aa498941b9a" gracePeriod=30 Nov 25 10:51:44 crc kubenswrapper[4821]: I1125 10:51:44.271022 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"87987062-8fb1-4174-829c-cb35c4629905","Type":"ContainerStarted","Data":"1e9369c555a202efe04554970bfa90dde3e0bce21d2143f068818cf4ab84f8ff"} Nov 25 10:51:44 crc kubenswrapper[4821]: I1125 10:51:44.300830 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.499242435 podStartE2EDuration="11.300808682s" podCreationTimestamp="2025-11-25 10:51:33 +0000 UTC" firstStartedPulling="2025-11-25 10:51:34.614191677 +0000 UTC m=+1165.150511524" lastFinishedPulling="2025-11-25 10:51:42.415757924 +0000 UTC m=+1172.952077771" observedRunningTime="2025-11-25 10:51:44.290623454 +0000 UTC m=+1174.826943301" watchObservedRunningTime="2025-11-25 10:51:44.300808682 +0000 UTC m=+1174.837128529" Nov 25 10:51:44 crc kubenswrapper[4821]: I1125 10:51:44.314452 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.961245915 podStartE2EDuration="11.31443268s" podCreationTimestamp="2025-11-25 10:51:33 +0000 UTC" firstStartedPulling="2025-11-25 10:51:34.060223157 +0000 UTC m=+1164.596543004" lastFinishedPulling="2025-11-25 10:51:42.413409922 +0000 UTC m=+1172.949729769" observedRunningTime="2025-11-25 10:51:44.312362205 +0000 UTC m=+1174.848682072" watchObservedRunningTime="2025-11-25 10:51:44.31443268 +0000 UTC m=+1174.850752527" Nov 25 10:51:44 crc kubenswrapper[4821]: I1125 10:51:44.324301 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 25 10:51:45 crc kubenswrapper[4821]: I1125 10:51:45.278811 4821 generic.go:334] "Generic (PLEG): container finished" podID="b044e354-e28e-43c0-ba43-bb51e8a21761" containerID="96b7eb0d2e5a8a8aa0bfb1dcf6c3663a69b3ad48e3c9efc121471d2b781c58e3" exitCode=143 Nov 25 10:51:45 crc kubenswrapper[4821]: I1125 10:51:45.278889 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b044e354-e28e-43c0-ba43-bb51e8a21761","Type":"ContainerDied","Data":"96b7eb0d2e5a8a8aa0bfb1dcf6c3663a69b3ad48e3c9efc121471d2b781c58e3"} Nov 25 10:51:46 crc kubenswrapper[4821]: I1125 10:51:46.286916 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 25 10:51:48 crc kubenswrapper[4821]: I1125 10:51:48.307599 4821 generic.go:334] "Generic (PLEG): container finished" podID="06c32eab-f3e4-4ebb-bb0e-0a887c0da0de" containerID="7cf37449c9b799d7e943c3e4748fcbac868542b3396da366d9c1cfbfb5ba0a26" exitCode=0 Nov 25 10:51:48 crc kubenswrapper[4821]: I1125 10:51:48.307823 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-qwwdq" event={"ID":"06c32eab-f3e4-4ebb-bb0e-0a887c0da0de","Type":"ContainerDied","Data":"7cf37449c9b799d7e943c3e4748fcbac868542b3396da366d9c1cfbfb5ba0a26"} Nov 25 10:51:49 crc kubenswrapper[4821]: I1125 10:51:49.127726 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 25 10:51:49 crc kubenswrapper[4821]: I1125 10:51:49.127781 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 25 10:51:49 crc kubenswrapper[4821]: I1125 10:51:49.158397 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-bccf8f775-wsvhq" Nov 25 10:51:49 crc kubenswrapper[4821]: I1125 10:51:49.231371 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-svh8j"] Nov 25 10:51:49 crc kubenswrapper[4821]: I1125 10:51:49.231656 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6578955fd5-svh8j" podUID="3d48a46a-3ae2-433a-a70b-cdda6d764e53" containerName="dnsmasq-dns" containerID="cri-o://66efde77ebbf8d3d6233842f84c3ca3fba12aabb2793cf543e112fd2340ca7a2" gracePeriod=10 Nov 25 10:51:49 crc kubenswrapper[4821]: I1125 10:51:49.754276 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-qwwdq" Nov 25 10:51:49 crc kubenswrapper[4821]: I1125 10:51:49.836707 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vntwq\" (UniqueName: \"kubernetes.io/projected/06c32eab-f3e4-4ebb-bb0e-0a887c0da0de-kube-api-access-vntwq\") pod \"06c32eab-f3e4-4ebb-bb0e-0a887c0da0de\" (UID: \"06c32eab-f3e4-4ebb-bb0e-0a887c0da0de\") " Nov 25 10:51:49 crc kubenswrapper[4821]: I1125 10:51:49.836817 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06c32eab-f3e4-4ebb-bb0e-0a887c0da0de-scripts\") pod \"06c32eab-f3e4-4ebb-bb0e-0a887c0da0de\" (UID: \"06c32eab-f3e4-4ebb-bb0e-0a887c0da0de\") " Nov 25 10:51:49 crc kubenswrapper[4821]: I1125 10:51:49.836878 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06c32eab-f3e4-4ebb-bb0e-0a887c0da0de-config-data\") pod \"06c32eab-f3e4-4ebb-bb0e-0a887c0da0de\" (UID: \"06c32eab-f3e4-4ebb-bb0e-0a887c0da0de\") " Nov 25 10:51:49 crc kubenswrapper[4821]: I1125 10:51:49.836948 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06c32eab-f3e4-4ebb-bb0e-0a887c0da0de-combined-ca-bundle\") pod \"06c32eab-f3e4-4ebb-bb0e-0a887c0da0de\" (UID: \"06c32eab-f3e4-4ebb-bb0e-0a887c0da0de\") " Nov 25 10:51:49 crc kubenswrapper[4821]: I1125 10:51:49.844003 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06c32eab-f3e4-4ebb-bb0e-0a887c0da0de-scripts" (OuterVolumeSpecName: "scripts") pod "06c32eab-f3e4-4ebb-bb0e-0a887c0da0de" (UID: "06c32eab-f3e4-4ebb-bb0e-0a887c0da0de"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:51:49 crc kubenswrapper[4821]: I1125 10:51:49.846334 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06c32eab-f3e4-4ebb-bb0e-0a887c0da0de-kube-api-access-vntwq" (OuterVolumeSpecName: "kube-api-access-vntwq") pod "06c32eab-f3e4-4ebb-bb0e-0a887c0da0de" (UID: "06c32eab-f3e4-4ebb-bb0e-0a887c0da0de"). InnerVolumeSpecName "kube-api-access-vntwq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:51:49 crc kubenswrapper[4821]: I1125 10:51:49.863235 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-svh8j" Nov 25 10:51:49 crc kubenswrapper[4821]: I1125 10:51:49.869865 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06c32eab-f3e4-4ebb-bb0e-0a887c0da0de-config-data" (OuterVolumeSpecName: "config-data") pod "06c32eab-f3e4-4ebb-bb0e-0a887c0da0de" (UID: "06c32eab-f3e4-4ebb-bb0e-0a887c0da0de"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:51:49 crc kubenswrapper[4821]: I1125 10:51:49.882549 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06c32eab-f3e4-4ebb-bb0e-0a887c0da0de-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "06c32eab-f3e4-4ebb-bb0e-0a887c0da0de" (UID: "06c32eab-f3e4-4ebb-bb0e-0a887c0da0de"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:51:49 crc kubenswrapper[4821]: I1125 10:51:49.939036 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v86b8\" (UniqueName: \"kubernetes.io/projected/3d48a46a-3ae2-433a-a70b-cdda6d764e53-kube-api-access-v86b8\") pod \"3d48a46a-3ae2-433a-a70b-cdda6d764e53\" (UID: \"3d48a46a-3ae2-433a-a70b-cdda6d764e53\") " Nov 25 10:51:49 crc kubenswrapper[4821]: I1125 10:51:49.939562 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3d48a46a-3ae2-433a-a70b-cdda6d764e53-ovsdbserver-nb\") pod \"3d48a46a-3ae2-433a-a70b-cdda6d764e53\" (UID: \"3d48a46a-3ae2-433a-a70b-cdda6d764e53\") " Nov 25 10:51:49 crc kubenswrapper[4821]: I1125 10:51:49.939712 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d48a46a-3ae2-433a-a70b-cdda6d764e53-config\") pod \"3d48a46a-3ae2-433a-a70b-cdda6d764e53\" (UID: \"3d48a46a-3ae2-433a-a70b-cdda6d764e53\") " Nov 25 10:51:49 crc kubenswrapper[4821]: I1125 10:51:49.939831 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3d48a46a-3ae2-433a-a70b-cdda6d764e53-ovsdbserver-sb\") pod \"3d48a46a-3ae2-433a-a70b-cdda6d764e53\" (UID: \"3d48a46a-3ae2-433a-a70b-cdda6d764e53\") " Nov 25 10:51:49 crc kubenswrapper[4821]: I1125 10:51:49.939996 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3d48a46a-3ae2-433a-a70b-cdda6d764e53-dns-svc\") pod \"3d48a46a-3ae2-433a-a70b-cdda6d764e53\" (UID: \"3d48a46a-3ae2-433a-a70b-cdda6d764e53\") " Nov 25 10:51:49 crc kubenswrapper[4821]: I1125 10:51:49.940177 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3d48a46a-3ae2-433a-a70b-cdda6d764e53-dns-swift-storage-0\") pod \"3d48a46a-3ae2-433a-a70b-cdda6d764e53\" (UID: \"3d48a46a-3ae2-433a-a70b-cdda6d764e53\") " Nov 25 10:51:49 crc kubenswrapper[4821]: I1125 10:51:49.940817 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vntwq\" (UniqueName: \"kubernetes.io/projected/06c32eab-f3e4-4ebb-bb0e-0a887c0da0de-kube-api-access-vntwq\") on node \"crc\" DevicePath \"\"" Nov 25 10:51:49 crc kubenswrapper[4821]: I1125 10:51:49.940945 4821 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06c32eab-f3e4-4ebb-bb0e-0a887c0da0de-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:51:49 crc kubenswrapper[4821]: I1125 10:51:49.941027 4821 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06c32eab-f3e4-4ebb-bb0e-0a887c0da0de-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:51:49 crc kubenswrapper[4821]: I1125 10:51:49.941106 4821 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06c32eab-f3e4-4ebb-bb0e-0a887c0da0de-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:51:49 crc kubenswrapper[4821]: I1125 10:51:49.944678 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d48a46a-3ae2-433a-a70b-cdda6d764e53-kube-api-access-v86b8" (OuterVolumeSpecName: "kube-api-access-v86b8") pod "3d48a46a-3ae2-433a-a70b-cdda6d764e53" (UID: "3d48a46a-3ae2-433a-a70b-cdda6d764e53"). InnerVolumeSpecName "kube-api-access-v86b8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:51:49 crc kubenswrapper[4821]: I1125 10:51:49.998867 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3d48a46a-3ae2-433a-a70b-cdda6d764e53-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "3d48a46a-3ae2-433a-a70b-cdda6d764e53" (UID: "3d48a46a-3ae2-433a-a70b-cdda6d764e53"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:51:50 crc kubenswrapper[4821]: I1125 10:51:50.000774 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3d48a46a-3ae2-433a-a70b-cdda6d764e53-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3d48a46a-3ae2-433a-a70b-cdda6d764e53" (UID: "3d48a46a-3ae2-433a-a70b-cdda6d764e53"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:51:50 crc kubenswrapper[4821]: I1125 10:51:50.001693 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3d48a46a-3ae2-433a-a70b-cdda6d764e53-config" (OuterVolumeSpecName: "config") pod "3d48a46a-3ae2-433a-a70b-cdda6d764e53" (UID: "3d48a46a-3ae2-433a-a70b-cdda6d764e53"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:51:50 crc kubenswrapper[4821]: I1125 10:51:50.003029 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3d48a46a-3ae2-433a-a70b-cdda6d764e53-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3d48a46a-3ae2-433a-a70b-cdda6d764e53" (UID: "3d48a46a-3ae2-433a-a70b-cdda6d764e53"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:51:50 crc kubenswrapper[4821]: I1125 10:51:50.007416 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3d48a46a-3ae2-433a-a70b-cdda6d764e53-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3d48a46a-3ae2-433a-a70b-cdda6d764e53" (UID: "3d48a46a-3ae2-433a-a70b-cdda6d764e53"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:51:50 crc kubenswrapper[4821]: I1125 10:51:50.026936 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 25 10:51:50 crc kubenswrapper[4821]: I1125 10:51:50.028016 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="e2b8639b-42d3-4de6-8d08-d5e2aac42391" containerName="kube-state-metrics" containerID="cri-o://9b16284d6f14b5581a449f355c6acf99510233d74e0787abff5bdfca3e4d45dd" gracePeriod=30 Nov 25 10:51:50 crc kubenswrapper[4821]: I1125 10:51:50.042670 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v86b8\" (UniqueName: \"kubernetes.io/projected/3d48a46a-3ae2-433a-a70b-cdda6d764e53-kube-api-access-v86b8\") on node \"crc\" DevicePath \"\"" Nov 25 10:51:50 crc kubenswrapper[4821]: I1125 10:51:50.042707 4821 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3d48a46a-3ae2-433a-a70b-cdda6d764e53-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 10:51:50 crc kubenswrapper[4821]: I1125 10:51:50.042721 4821 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d48a46a-3ae2-433a-a70b-cdda6d764e53-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:51:50 crc kubenswrapper[4821]: I1125 10:51:50.042733 4821 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3d48a46a-3ae2-433a-a70b-cdda6d764e53-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 10:51:50 crc kubenswrapper[4821]: I1125 10:51:50.042744 4821 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3d48a46a-3ae2-433a-a70b-cdda6d764e53-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 10:51:50 crc kubenswrapper[4821]: I1125 10:51:50.042756 4821 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3d48a46a-3ae2-433a-a70b-cdda6d764e53-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 25 10:51:50 crc kubenswrapper[4821]: I1125 10:51:50.333501 4821 generic.go:334] "Generic (PLEG): container finished" podID="3d48a46a-3ae2-433a-a70b-cdda6d764e53" containerID="66efde77ebbf8d3d6233842f84c3ca3fba12aabb2793cf543e112fd2340ca7a2" exitCode=0 Nov 25 10:51:50 crc kubenswrapper[4821]: I1125 10:51:50.333560 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-svh8j" event={"ID":"3d48a46a-3ae2-433a-a70b-cdda6d764e53","Type":"ContainerDied","Data":"66efde77ebbf8d3d6233842f84c3ca3fba12aabb2793cf543e112fd2340ca7a2"} Nov 25 10:51:50 crc kubenswrapper[4821]: I1125 10:51:50.333593 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-svh8j" Nov 25 10:51:50 crc kubenswrapper[4821]: I1125 10:51:50.333639 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-svh8j" event={"ID":"3d48a46a-3ae2-433a-a70b-cdda6d764e53","Type":"ContainerDied","Data":"7db3ab42634c39bef4a1bdab6529d7c05eeee67cf31996620083613a8667997a"} Nov 25 10:51:50 crc kubenswrapper[4821]: I1125 10:51:50.333663 4821 scope.go:117] "RemoveContainer" containerID="66efde77ebbf8d3d6233842f84c3ca3fba12aabb2793cf543e112fd2340ca7a2" Nov 25 10:51:50 crc kubenswrapper[4821]: I1125 10:51:50.360660 4821 generic.go:334] "Generic (PLEG): container finished" podID="e2b8639b-42d3-4de6-8d08-d5e2aac42391" containerID="9b16284d6f14b5581a449f355c6acf99510233d74e0787abff5bdfca3e4d45dd" exitCode=2 Nov 25 10:51:50 crc kubenswrapper[4821]: I1125 10:51:50.360765 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"e2b8639b-42d3-4de6-8d08-d5e2aac42391","Type":"ContainerDied","Data":"9b16284d6f14b5581a449f355c6acf99510233d74e0787abff5bdfca3e4d45dd"} Nov 25 10:51:50 crc kubenswrapper[4821]: I1125 10:51:50.379359 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-svh8j"] Nov 25 10:51:50 crc kubenswrapper[4821]: I1125 10:51:50.385660 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-qwwdq" event={"ID":"06c32eab-f3e4-4ebb-bb0e-0a887c0da0de","Type":"ContainerDied","Data":"dda66e472de86b2c228de028bbf2d6e82bd8bc2c20000428e352be8a20b1f0c3"} Nov 25 10:51:50 crc kubenswrapper[4821]: I1125 10:51:50.385705 4821 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dda66e472de86b2c228de028bbf2d6e82bd8bc2c20000428e352be8a20b1f0c3" Nov 25 10:51:50 crc kubenswrapper[4821]: I1125 10:51:50.385799 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-qwwdq" Nov 25 10:51:50 crc kubenswrapper[4821]: I1125 10:51:50.391889 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-svh8j"] Nov 25 10:51:50 crc kubenswrapper[4821]: I1125 10:51:50.493116 4821 scope.go:117] "RemoveContainer" containerID="9ff6a635f643fcc567f969ef5dcb1db96e5cdbfa8ccc746ae9f443de342fa116" Nov 25 10:51:50 crc kubenswrapper[4821]: I1125 10:51:50.507956 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 25 10:51:50 crc kubenswrapper[4821]: I1125 10:51:50.508209 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="87987062-8fb1-4174-829c-cb35c4629905" containerName="nova-api-log" containerID="cri-o://150ef83d8ae7669f561e8aff63be4c4ade17d7f79c71d5a4bbfd49ed6bd39a1b" gracePeriod=30 Nov 25 10:51:50 crc kubenswrapper[4821]: I1125 10:51:50.508308 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="87987062-8fb1-4174-829c-cb35c4629905" containerName="nova-api-api" containerID="cri-o://1e9369c555a202efe04554970bfa90dde3e0bce21d2143f068818cf4ab84f8ff" gracePeriod=30 Nov 25 10:51:50 crc kubenswrapper[4821]: I1125 10:51:50.518919 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 10:51:50 crc kubenswrapper[4821]: I1125 10:51:50.519088 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="2bd16d9e-1992-4537-8376-c8d0da2f4a31" containerName="nova-scheduler-scheduler" containerID="cri-o://69bd9ea9d864923a312714582044157d74603cd071907d70637996049c8c2904" gracePeriod=30 Nov 25 10:51:50 crc kubenswrapper[4821]: I1125 10:51:50.612509 4821 scope.go:117] "RemoveContainer" containerID="66efde77ebbf8d3d6233842f84c3ca3fba12aabb2793cf543e112fd2340ca7a2" Nov 25 10:51:50 crc kubenswrapper[4821]: E1125 10:51:50.613078 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"66efde77ebbf8d3d6233842f84c3ca3fba12aabb2793cf543e112fd2340ca7a2\": container with ID starting with 66efde77ebbf8d3d6233842f84c3ca3fba12aabb2793cf543e112fd2340ca7a2 not found: ID does not exist" containerID="66efde77ebbf8d3d6233842f84c3ca3fba12aabb2793cf543e112fd2340ca7a2" Nov 25 10:51:50 crc kubenswrapper[4821]: I1125 10:51:50.613192 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66efde77ebbf8d3d6233842f84c3ca3fba12aabb2793cf543e112fd2340ca7a2"} err="failed to get container status \"66efde77ebbf8d3d6233842f84c3ca3fba12aabb2793cf543e112fd2340ca7a2\": rpc error: code = NotFound desc = could not find container \"66efde77ebbf8d3d6233842f84c3ca3fba12aabb2793cf543e112fd2340ca7a2\": container with ID starting with 66efde77ebbf8d3d6233842f84c3ca3fba12aabb2793cf543e112fd2340ca7a2 not found: ID does not exist" Nov 25 10:51:50 crc kubenswrapper[4821]: I1125 10:51:50.613227 4821 scope.go:117] "RemoveContainer" containerID="9ff6a635f643fcc567f969ef5dcb1db96e5cdbfa8ccc746ae9f443de342fa116" Nov 25 10:51:50 crc kubenswrapper[4821]: E1125 10:51:50.613680 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ff6a635f643fcc567f969ef5dcb1db96e5cdbfa8ccc746ae9f443de342fa116\": container with ID starting with 9ff6a635f643fcc567f969ef5dcb1db96e5cdbfa8ccc746ae9f443de342fa116 not found: ID does not exist" containerID="9ff6a635f643fcc567f969ef5dcb1db96e5cdbfa8ccc746ae9f443de342fa116" Nov 25 10:51:50 crc kubenswrapper[4821]: I1125 10:51:50.613716 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ff6a635f643fcc567f969ef5dcb1db96e5cdbfa8ccc746ae9f443de342fa116"} err="failed to get container status \"9ff6a635f643fcc567f969ef5dcb1db96e5cdbfa8ccc746ae9f443de342fa116\": rpc error: code = NotFound desc = could not find container \"9ff6a635f643fcc567f969ef5dcb1db96e5cdbfa8ccc746ae9f443de342fa116\": container with ID starting with 9ff6a635f643fcc567f969ef5dcb1db96e5cdbfa8ccc746ae9f443de342fa116 not found: ID does not exist" Nov 25 10:51:50 crc kubenswrapper[4821]: I1125 10:51:50.759490 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 25 10:51:50 crc kubenswrapper[4821]: I1125 10:51:50.874348 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rwnx2\" (UniqueName: \"kubernetes.io/projected/e2b8639b-42d3-4de6-8d08-d5e2aac42391-kube-api-access-rwnx2\") pod \"e2b8639b-42d3-4de6-8d08-d5e2aac42391\" (UID: \"e2b8639b-42d3-4de6-8d08-d5e2aac42391\") " Nov 25 10:51:50 crc kubenswrapper[4821]: I1125 10:51:50.883619 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2b8639b-42d3-4de6-8d08-d5e2aac42391-kube-api-access-rwnx2" (OuterVolumeSpecName: "kube-api-access-rwnx2") pod "e2b8639b-42d3-4de6-8d08-d5e2aac42391" (UID: "e2b8639b-42d3-4de6-8d08-d5e2aac42391"). InnerVolumeSpecName "kube-api-access-rwnx2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:51:50 crc kubenswrapper[4821]: I1125 10:51:50.977993 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rwnx2\" (UniqueName: \"kubernetes.io/projected/e2b8639b-42d3-4de6-8d08-d5e2aac42391-kube-api-access-rwnx2\") on node \"crc\" DevicePath \"\"" Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.072186 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.181101 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87987062-8fb1-4174-829c-cb35c4629905-config-data\") pod \"87987062-8fb1-4174-829c-cb35c4629905\" (UID: \"87987062-8fb1-4174-829c-cb35c4629905\") " Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.181342 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87987062-8fb1-4174-829c-cb35c4629905-combined-ca-bundle\") pod \"87987062-8fb1-4174-829c-cb35c4629905\" (UID: \"87987062-8fb1-4174-829c-cb35c4629905\") " Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.181403 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n2djw\" (UniqueName: \"kubernetes.io/projected/87987062-8fb1-4174-829c-cb35c4629905-kube-api-access-n2djw\") pod \"87987062-8fb1-4174-829c-cb35c4629905\" (UID: \"87987062-8fb1-4174-829c-cb35c4629905\") " Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.181470 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87987062-8fb1-4174-829c-cb35c4629905-logs\") pod \"87987062-8fb1-4174-829c-cb35c4629905\" (UID: \"87987062-8fb1-4174-829c-cb35c4629905\") " Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.182309 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/87987062-8fb1-4174-829c-cb35c4629905-logs" (OuterVolumeSpecName: "logs") pod "87987062-8fb1-4174-829c-cb35c4629905" (UID: "87987062-8fb1-4174-829c-cb35c4629905"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.187580 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87987062-8fb1-4174-829c-cb35c4629905-kube-api-access-n2djw" (OuterVolumeSpecName: "kube-api-access-n2djw") pod "87987062-8fb1-4174-829c-cb35c4629905" (UID: "87987062-8fb1-4174-829c-cb35c4629905"). InnerVolumeSpecName "kube-api-access-n2djw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.213064 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87987062-8fb1-4174-829c-cb35c4629905-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "87987062-8fb1-4174-829c-cb35c4629905" (UID: "87987062-8fb1-4174-829c-cb35c4629905"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.220331 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87987062-8fb1-4174-829c-cb35c4629905-config-data" (OuterVolumeSpecName: "config-data") pod "87987062-8fb1-4174-829c-cb35c4629905" (UID: "87987062-8fb1-4174-829c-cb35c4629905"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.284054 4821 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87987062-8fb1-4174-829c-cb35c4629905-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.284093 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n2djw\" (UniqueName: \"kubernetes.io/projected/87987062-8fb1-4174-829c-cb35c4629905-kube-api-access-n2djw\") on node \"crc\" DevicePath \"\"" Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.284108 4821 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87987062-8fb1-4174-829c-cb35c4629905-logs\") on node \"crc\" DevicePath \"\"" Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.284118 4821 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87987062-8fb1-4174-829c-cb35c4629905-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.396192 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"e2b8639b-42d3-4de6-8d08-d5e2aac42391","Type":"ContainerDied","Data":"cd26802fbbd6b9df5a3b0dd72b2f0570f1d220bdc2538ea6a9db8e716dec4df0"} Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.396245 4821 scope.go:117] "RemoveContainer" containerID="9b16284d6f14b5581a449f355c6acf99510233d74e0787abff5bdfca3e4d45dd" Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.396370 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.405783 4821 generic.go:334] "Generic (PLEG): container finished" podID="87987062-8fb1-4174-829c-cb35c4629905" containerID="1e9369c555a202efe04554970bfa90dde3e0bce21d2143f068818cf4ab84f8ff" exitCode=0 Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.405833 4821 generic.go:334] "Generic (PLEG): container finished" podID="87987062-8fb1-4174-829c-cb35c4629905" containerID="150ef83d8ae7669f561e8aff63be4c4ade17d7f79c71d5a4bbfd49ed6bd39a1b" exitCode=143 Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.405916 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"87987062-8fb1-4174-829c-cb35c4629905","Type":"ContainerDied","Data":"1e9369c555a202efe04554970bfa90dde3e0bce21d2143f068818cf4ab84f8ff"} Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.405951 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"87987062-8fb1-4174-829c-cb35c4629905","Type":"ContainerDied","Data":"150ef83d8ae7669f561e8aff63be4c4ade17d7f79c71d5a4bbfd49ed6bd39a1b"} Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.405962 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"87987062-8fb1-4174-829c-cb35c4629905","Type":"ContainerDied","Data":"efb6210df717a08597b9986d49cbfcc2066330ec7e37ba6f2f136d0fd85eff68"} Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.406082 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.462094 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.488090 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.493724 4821 scope.go:117] "RemoveContainer" containerID="1e9369c555a202efe04554970bfa90dde3e0bce21d2143f068818cf4ab84f8ff" Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.511284 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.528527 4821 scope.go:117] "RemoveContainer" containerID="150ef83d8ae7669f561e8aff63be4c4ade17d7f79c71d5a4bbfd49ed6bd39a1b" Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.532642 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.543005 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 25 10:51:51 crc kubenswrapper[4821]: E1125 10:51:51.543546 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06c32eab-f3e4-4ebb-bb0e-0a887c0da0de" containerName="nova-manage" Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.543572 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="06c32eab-f3e4-4ebb-bb0e-0a887c0da0de" containerName="nova-manage" Nov 25 10:51:51 crc kubenswrapper[4821]: E1125 10:51:51.543601 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87987062-8fb1-4174-829c-cb35c4629905" containerName="nova-api-api" Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.543610 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="87987062-8fb1-4174-829c-cb35c4629905" containerName="nova-api-api" Nov 25 10:51:51 crc kubenswrapper[4821]: E1125 10:51:51.543634 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2b8639b-42d3-4de6-8d08-d5e2aac42391" containerName="kube-state-metrics" Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.543644 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2b8639b-42d3-4de6-8d08-d5e2aac42391" containerName="kube-state-metrics" Nov 25 10:51:51 crc kubenswrapper[4821]: E1125 10:51:51.543655 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d48a46a-3ae2-433a-a70b-cdda6d764e53" containerName="init" Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.543662 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d48a46a-3ae2-433a-a70b-cdda6d764e53" containerName="init" Nov 25 10:51:51 crc kubenswrapper[4821]: E1125 10:51:51.543679 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87987062-8fb1-4174-829c-cb35c4629905" containerName="nova-api-log" Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.543688 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="87987062-8fb1-4174-829c-cb35c4629905" containerName="nova-api-log" Nov 25 10:51:51 crc kubenswrapper[4821]: E1125 10:51:51.543704 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d48a46a-3ae2-433a-a70b-cdda6d764e53" containerName="dnsmasq-dns" Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.543712 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d48a46a-3ae2-433a-a70b-cdda6d764e53" containerName="dnsmasq-dns" Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.543922 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2b8639b-42d3-4de6-8d08-d5e2aac42391" containerName="kube-state-metrics" Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.543943 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="06c32eab-f3e4-4ebb-bb0e-0a887c0da0de" containerName="nova-manage" Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.543958 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d48a46a-3ae2-433a-a70b-cdda6d764e53" containerName="dnsmasq-dns" Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.543969 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="87987062-8fb1-4174-829c-cb35c4629905" containerName="nova-api-api" Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.543989 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="87987062-8fb1-4174-829c-cb35c4629905" containerName="nova-api-log" Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.545336 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.551704 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.557607 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.559374 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.562706 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.562853 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.570519 4821 scope.go:117] "RemoveContainer" containerID="1e9369c555a202efe04554970bfa90dde3e0bce21d2143f068818cf4ab84f8ff" Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.571073 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 25 10:51:51 crc kubenswrapper[4821]: E1125 10:51:51.571082 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e9369c555a202efe04554970bfa90dde3e0bce21d2143f068818cf4ab84f8ff\": container with ID starting with 1e9369c555a202efe04554970bfa90dde3e0bce21d2143f068818cf4ab84f8ff not found: ID does not exist" containerID="1e9369c555a202efe04554970bfa90dde3e0bce21d2143f068818cf4ab84f8ff" Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.571141 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e9369c555a202efe04554970bfa90dde3e0bce21d2143f068818cf4ab84f8ff"} err="failed to get container status \"1e9369c555a202efe04554970bfa90dde3e0bce21d2143f068818cf4ab84f8ff\": rpc error: code = NotFound desc = could not find container \"1e9369c555a202efe04554970bfa90dde3e0bce21d2143f068818cf4ab84f8ff\": container with ID starting with 1e9369c555a202efe04554970bfa90dde3e0bce21d2143f068818cf4ab84f8ff not found: ID does not exist" Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.571189 4821 scope.go:117] "RemoveContainer" containerID="150ef83d8ae7669f561e8aff63be4c4ade17d7f79c71d5a4bbfd49ed6bd39a1b" Nov 25 10:51:51 crc kubenswrapper[4821]: E1125 10:51:51.571647 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"150ef83d8ae7669f561e8aff63be4c4ade17d7f79c71d5a4bbfd49ed6bd39a1b\": container with ID starting with 150ef83d8ae7669f561e8aff63be4c4ade17d7f79c71d5a4bbfd49ed6bd39a1b not found: ID does not exist" containerID="150ef83d8ae7669f561e8aff63be4c4ade17d7f79c71d5a4bbfd49ed6bd39a1b" Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.571670 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"150ef83d8ae7669f561e8aff63be4c4ade17d7f79c71d5a4bbfd49ed6bd39a1b"} err="failed to get container status \"150ef83d8ae7669f561e8aff63be4c4ade17d7f79c71d5a4bbfd49ed6bd39a1b\": rpc error: code = NotFound desc = could not find container \"150ef83d8ae7669f561e8aff63be4c4ade17d7f79c71d5a4bbfd49ed6bd39a1b\": container with ID starting with 150ef83d8ae7669f561e8aff63be4c4ade17d7f79c71d5a4bbfd49ed6bd39a1b not found: ID does not exist" Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.571692 4821 scope.go:117] "RemoveContainer" containerID="1e9369c555a202efe04554970bfa90dde3e0bce21d2143f068818cf4ab84f8ff" Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.571883 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e9369c555a202efe04554970bfa90dde3e0bce21d2143f068818cf4ab84f8ff"} err="failed to get container status \"1e9369c555a202efe04554970bfa90dde3e0bce21d2143f068818cf4ab84f8ff\": rpc error: code = NotFound desc = could not find container \"1e9369c555a202efe04554970bfa90dde3e0bce21d2143f068818cf4ab84f8ff\": container with ID starting with 1e9369c555a202efe04554970bfa90dde3e0bce21d2143f068818cf4ab84f8ff not found: ID does not exist" Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.571896 4821 scope.go:117] "RemoveContainer" containerID="150ef83d8ae7669f561e8aff63be4c4ade17d7f79c71d5a4bbfd49ed6bd39a1b" Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.572058 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"150ef83d8ae7669f561e8aff63be4c4ade17d7f79c71d5a4bbfd49ed6bd39a1b"} err="failed to get container status \"150ef83d8ae7669f561e8aff63be4c4ade17d7f79c71d5a4bbfd49ed6bd39a1b\": rpc error: code = NotFound desc = could not find container \"150ef83d8ae7669f561e8aff63be4c4ade17d7f79c71d5a4bbfd49ed6bd39a1b\": container with ID starting with 150ef83d8ae7669f561e8aff63be4c4ade17d7f79c71d5a4bbfd49ed6bd39a1b not found: ID does not exist" Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.584900 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.691528 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/7c286dad-cd09-4085-a4e0-4b274f30d38d-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"7c286dad-cd09-4085-a4e0-4b274f30d38d\") " pod="openstack/kube-state-metrics-0" Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.691580 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6vtc\" (UniqueName: \"kubernetes.io/projected/31227d89-6706-407e-aa0a-e90b99611462-kube-api-access-q6vtc\") pod \"nova-api-0\" (UID: \"31227d89-6706-407e-aa0a-e90b99611462\") " pod="openstack/nova-api-0" Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.691650 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c286dad-cd09-4085-a4e0-4b274f30d38d-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"7c286dad-cd09-4085-a4e0-4b274f30d38d\") " pod="openstack/kube-state-metrics-0" Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.691767 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31227d89-6706-407e-aa0a-e90b99611462-config-data\") pod \"nova-api-0\" (UID: \"31227d89-6706-407e-aa0a-e90b99611462\") " pod="openstack/nova-api-0" Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.691806 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ssspw\" (UniqueName: \"kubernetes.io/projected/7c286dad-cd09-4085-a4e0-4b274f30d38d-kube-api-access-ssspw\") pod \"kube-state-metrics-0\" (UID: \"7c286dad-cd09-4085-a4e0-4b274f30d38d\") " pod="openstack/kube-state-metrics-0" Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.691840 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c286dad-cd09-4085-a4e0-4b274f30d38d-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"7c286dad-cd09-4085-a4e0-4b274f30d38d\") " pod="openstack/kube-state-metrics-0" Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.691913 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31227d89-6706-407e-aa0a-e90b99611462-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"31227d89-6706-407e-aa0a-e90b99611462\") " pod="openstack/nova-api-0" Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.692011 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/31227d89-6706-407e-aa0a-e90b99611462-logs\") pod \"nova-api-0\" (UID: \"31227d89-6706-407e-aa0a-e90b99611462\") " pod="openstack/nova-api-0" Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.795431 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31227d89-6706-407e-aa0a-e90b99611462-config-data\") pod \"nova-api-0\" (UID: \"31227d89-6706-407e-aa0a-e90b99611462\") " pod="openstack/nova-api-0" Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.795475 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ssspw\" (UniqueName: \"kubernetes.io/projected/7c286dad-cd09-4085-a4e0-4b274f30d38d-kube-api-access-ssspw\") pod \"kube-state-metrics-0\" (UID: \"7c286dad-cd09-4085-a4e0-4b274f30d38d\") " pod="openstack/kube-state-metrics-0" Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.795513 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c286dad-cd09-4085-a4e0-4b274f30d38d-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"7c286dad-cd09-4085-a4e0-4b274f30d38d\") " pod="openstack/kube-state-metrics-0" Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.795568 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31227d89-6706-407e-aa0a-e90b99611462-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"31227d89-6706-407e-aa0a-e90b99611462\") " pod="openstack/nova-api-0" Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.795641 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/31227d89-6706-407e-aa0a-e90b99611462-logs\") pod \"nova-api-0\" (UID: \"31227d89-6706-407e-aa0a-e90b99611462\") " pod="openstack/nova-api-0" Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.795673 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/7c286dad-cd09-4085-a4e0-4b274f30d38d-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"7c286dad-cd09-4085-a4e0-4b274f30d38d\") " pod="openstack/kube-state-metrics-0" Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.795707 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6vtc\" (UniqueName: \"kubernetes.io/projected/31227d89-6706-407e-aa0a-e90b99611462-kube-api-access-q6vtc\") pod \"nova-api-0\" (UID: \"31227d89-6706-407e-aa0a-e90b99611462\") " pod="openstack/nova-api-0" Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.795757 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c286dad-cd09-4085-a4e0-4b274f30d38d-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"7c286dad-cd09-4085-a4e0-4b274f30d38d\") " pod="openstack/kube-state-metrics-0" Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.796447 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/31227d89-6706-407e-aa0a-e90b99611462-logs\") pod \"nova-api-0\" (UID: \"31227d89-6706-407e-aa0a-e90b99611462\") " pod="openstack/nova-api-0" Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.800261 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31227d89-6706-407e-aa0a-e90b99611462-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"31227d89-6706-407e-aa0a-e90b99611462\") " pod="openstack/nova-api-0" Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.800669 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/7c286dad-cd09-4085-a4e0-4b274f30d38d-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"7c286dad-cd09-4085-a4e0-4b274f30d38d\") " pod="openstack/kube-state-metrics-0" Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.803460 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31227d89-6706-407e-aa0a-e90b99611462-config-data\") pod \"nova-api-0\" (UID: \"31227d89-6706-407e-aa0a-e90b99611462\") " pod="openstack/nova-api-0" Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.805411 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c286dad-cd09-4085-a4e0-4b274f30d38d-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"7c286dad-cd09-4085-a4e0-4b274f30d38d\") " pod="openstack/kube-state-metrics-0" Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.812322 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c286dad-cd09-4085-a4e0-4b274f30d38d-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"7c286dad-cd09-4085-a4e0-4b274f30d38d\") " pod="openstack/kube-state-metrics-0" Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.815236 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ssspw\" (UniqueName: \"kubernetes.io/projected/7c286dad-cd09-4085-a4e0-4b274f30d38d-kube-api-access-ssspw\") pod \"kube-state-metrics-0\" (UID: \"7c286dad-cd09-4085-a4e0-4b274f30d38d\") " pod="openstack/kube-state-metrics-0" Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.822982 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6vtc\" (UniqueName: \"kubernetes.io/projected/31227d89-6706-407e-aa0a-e90b99611462-kube-api-access-q6vtc\") pod \"nova-api-0\" (UID: \"31227d89-6706-407e-aa0a-e90b99611462\") " pod="openstack/nova-api-0" Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.880605 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 10:51:51 crc kubenswrapper[4821]: I1125 10:51:51.915518 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 25 10:51:52 crc kubenswrapper[4821]: I1125 10:51:52.131770 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d48a46a-3ae2-433a-a70b-cdda6d764e53" path="/var/lib/kubelet/pods/3d48a46a-3ae2-433a-a70b-cdda6d764e53/volumes" Nov 25 10:51:52 crc kubenswrapper[4821]: I1125 10:51:52.133496 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87987062-8fb1-4174-829c-cb35c4629905" path="/var/lib/kubelet/pods/87987062-8fb1-4174-829c-cb35c4629905/volumes" Nov 25 10:51:52 crc kubenswrapper[4821]: I1125 10:51:52.136808 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e2b8639b-42d3-4de6-8d08-d5e2aac42391" path="/var/lib/kubelet/pods/e2b8639b-42d3-4de6-8d08-d5e2aac42391/volumes" Nov 25 10:51:52 crc kubenswrapper[4821]: I1125 10:51:52.159991 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:51:52 crc kubenswrapper[4821]: I1125 10:51:52.160339 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="93ebfd94-d02a-4a94-bd08-2fa977ece0c2" containerName="ceilometer-central-agent" containerID="cri-o://58854775c62d04df46d082b0b3e917a617f69ac9824ab4befff224cb93919dfa" gracePeriod=30 Nov 25 10:51:52 crc kubenswrapper[4821]: I1125 10:51:52.160464 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="93ebfd94-d02a-4a94-bd08-2fa977ece0c2" containerName="proxy-httpd" containerID="cri-o://7f092fdc0f9a31f25a3a369e3b7e07011c4b9033d12d7a977abf4cf3161f86a1" gracePeriod=30 Nov 25 10:51:52 crc kubenswrapper[4821]: I1125 10:51:52.160517 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="93ebfd94-d02a-4a94-bd08-2fa977ece0c2" containerName="sg-core" containerID="cri-o://5651bfa9d025b88a03615fdbcd8d8ab5c0354dcbfa711ef2309fa1144334340c" gracePeriod=30 Nov 25 10:51:52 crc kubenswrapper[4821]: I1125 10:51:52.160548 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="93ebfd94-d02a-4a94-bd08-2fa977ece0c2" containerName="ceilometer-notification-agent" containerID="cri-o://4524ca39e676a4c2d2a0a5cd6233d1a9da964e9de331e24b23cf829c337da27a" gracePeriod=30 Nov 25 10:51:52 crc kubenswrapper[4821]: I1125 10:51:52.177842 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 25 10:51:52 crc kubenswrapper[4821]: I1125 10:51:52.306654 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f8khc\" (UniqueName: \"kubernetes.io/projected/2bd16d9e-1992-4537-8376-c8d0da2f4a31-kube-api-access-f8khc\") pod \"2bd16d9e-1992-4537-8376-c8d0da2f4a31\" (UID: \"2bd16d9e-1992-4537-8376-c8d0da2f4a31\") " Nov 25 10:51:52 crc kubenswrapper[4821]: I1125 10:51:52.306764 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bd16d9e-1992-4537-8376-c8d0da2f4a31-config-data\") pod \"2bd16d9e-1992-4537-8376-c8d0da2f4a31\" (UID: \"2bd16d9e-1992-4537-8376-c8d0da2f4a31\") " Nov 25 10:51:52 crc kubenswrapper[4821]: I1125 10:51:52.306868 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bd16d9e-1992-4537-8376-c8d0da2f4a31-combined-ca-bundle\") pod \"2bd16d9e-1992-4537-8376-c8d0da2f4a31\" (UID: \"2bd16d9e-1992-4537-8376-c8d0da2f4a31\") " Nov 25 10:51:52 crc kubenswrapper[4821]: I1125 10:51:52.320496 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2bd16d9e-1992-4537-8376-c8d0da2f4a31-kube-api-access-f8khc" (OuterVolumeSpecName: "kube-api-access-f8khc") pod "2bd16d9e-1992-4537-8376-c8d0da2f4a31" (UID: "2bd16d9e-1992-4537-8376-c8d0da2f4a31"). InnerVolumeSpecName "kube-api-access-f8khc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:51:52 crc kubenswrapper[4821]: I1125 10:51:52.332606 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bd16d9e-1992-4537-8376-c8d0da2f4a31-config-data" (OuterVolumeSpecName: "config-data") pod "2bd16d9e-1992-4537-8376-c8d0da2f4a31" (UID: "2bd16d9e-1992-4537-8376-c8d0da2f4a31"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:51:52 crc kubenswrapper[4821]: I1125 10:51:52.335084 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bd16d9e-1992-4537-8376-c8d0da2f4a31-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2bd16d9e-1992-4537-8376-c8d0da2f4a31" (UID: "2bd16d9e-1992-4537-8376-c8d0da2f4a31"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:51:52 crc kubenswrapper[4821]: I1125 10:51:52.404662 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 25 10:51:52 crc kubenswrapper[4821]: I1125 10:51:52.409809 4821 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bd16d9e-1992-4537-8376-c8d0da2f4a31-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:51:52 crc kubenswrapper[4821]: I1125 10:51:52.409856 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f8khc\" (UniqueName: \"kubernetes.io/projected/2bd16d9e-1992-4537-8376-c8d0da2f4a31-kube-api-access-f8khc\") on node \"crc\" DevicePath \"\"" Nov 25 10:51:52 crc kubenswrapper[4821]: I1125 10:51:52.409871 4821 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bd16d9e-1992-4537-8376-c8d0da2f4a31-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:51:52 crc kubenswrapper[4821]: I1125 10:51:52.424202 4821 generic.go:334] "Generic (PLEG): container finished" podID="2bd16d9e-1992-4537-8376-c8d0da2f4a31" containerID="69bd9ea9d864923a312714582044157d74603cd071907d70637996049c8c2904" exitCode=0 Nov 25 10:51:52 crc kubenswrapper[4821]: I1125 10:51:52.424259 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2bd16d9e-1992-4537-8376-c8d0da2f4a31","Type":"ContainerDied","Data":"69bd9ea9d864923a312714582044157d74603cd071907d70637996049c8c2904"} Nov 25 10:51:52 crc kubenswrapper[4821]: I1125 10:51:52.424283 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2bd16d9e-1992-4537-8376-c8d0da2f4a31","Type":"ContainerDied","Data":"aa15114ab525d62b1ea2b5ff34acc1545c341fa46702abcf7db24a155e6e0438"} Nov 25 10:51:52 crc kubenswrapper[4821]: I1125 10:51:52.424299 4821 scope.go:117] "RemoveContainer" containerID="69bd9ea9d864923a312714582044157d74603cd071907d70637996049c8c2904" Nov 25 10:51:52 crc kubenswrapper[4821]: I1125 10:51:52.424381 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 25 10:51:52 crc kubenswrapper[4821]: I1125 10:51:52.430878 4821 generic.go:334] "Generic (PLEG): container finished" podID="93ebfd94-d02a-4a94-bd08-2fa977ece0c2" containerID="7f092fdc0f9a31f25a3a369e3b7e07011c4b9033d12d7a977abf4cf3161f86a1" exitCode=0 Nov 25 10:51:52 crc kubenswrapper[4821]: I1125 10:51:52.430909 4821 generic.go:334] "Generic (PLEG): container finished" podID="93ebfd94-d02a-4a94-bd08-2fa977ece0c2" containerID="5651bfa9d025b88a03615fdbcd8d8ab5c0354dcbfa711ef2309fa1144334340c" exitCode=2 Nov 25 10:51:52 crc kubenswrapper[4821]: I1125 10:51:52.430953 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"93ebfd94-d02a-4a94-bd08-2fa977ece0c2","Type":"ContainerDied","Data":"7f092fdc0f9a31f25a3a369e3b7e07011c4b9033d12d7a977abf4cf3161f86a1"} Nov 25 10:51:52 crc kubenswrapper[4821]: I1125 10:51:52.430993 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"93ebfd94-d02a-4a94-bd08-2fa977ece0c2","Type":"ContainerDied","Data":"5651bfa9d025b88a03615fdbcd8d8ab5c0354dcbfa711ef2309fa1144334340c"} Nov 25 10:51:52 crc kubenswrapper[4821]: I1125 10:51:52.478660 4821 scope.go:117] "RemoveContainer" containerID="69bd9ea9d864923a312714582044157d74603cd071907d70637996049c8c2904" Nov 25 10:51:52 crc kubenswrapper[4821]: E1125 10:51:52.479245 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69bd9ea9d864923a312714582044157d74603cd071907d70637996049c8c2904\": container with ID starting with 69bd9ea9d864923a312714582044157d74603cd071907d70637996049c8c2904 not found: ID does not exist" containerID="69bd9ea9d864923a312714582044157d74603cd071907d70637996049c8c2904" Nov 25 10:51:52 crc kubenswrapper[4821]: I1125 10:51:52.479283 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69bd9ea9d864923a312714582044157d74603cd071907d70637996049c8c2904"} err="failed to get container status \"69bd9ea9d864923a312714582044157d74603cd071907d70637996049c8c2904\": rpc error: code = NotFound desc = could not find container \"69bd9ea9d864923a312714582044157d74603cd071907d70637996049c8c2904\": container with ID starting with 69bd9ea9d864923a312714582044157d74603cd071907d70637996049c8c2904 not found: ID does not exist" Nov 25 10:51:52 crc kubenswrapper[4821]: I1125 10:51:52.482091 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 10:51:52 crc kubenswrapper[4821]: I1125 10:51:52.493016 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 10:51:52 crc kubenswrapper[4821]: I1125 10:51:52.504298 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 10:51:52 crc kubenswrapper[4821]: E1125 10:51:52.506032 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bd16d9e-1992-4537-8376-c8d0da2f4a31" containerName="nova-scheduler-scheduler" Nov 25 10:51:52 crc kubenswrapper[4821]: I1125 10:51:52.506060 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bd16d9e-1992-4537-8376-c8d0da2f4a31" containerName="nova-scheduler-scheduler" Nov 25 10:51:52 crc kubenswrapper[4821]: I1125 10:51:52.506620 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bd16d9e-1992-4537-8376-c8d0da2f4a31" containerName="nova-scheduler-scheduler" Nov 25 10:51:52 crc kubenswrapper[4821]: I1125 10:51:52.507626 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 25 10:51:52 crc kubenswrapper[4821]: I1125 10:51:52.511977 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 25 10:51:52 crc kubenswrapper[4821]: I1125 10:51:52.517419 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 10:51:52 crc kubenswrapper[4821]: I1125 10:51:52.552463 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 25 10:51:52 crc kubenswrapper[4821]: W1125 10:51:52.553445 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7c286dad_cd09_4085_a4e0_4b274f30d38d.slice/crio-a4d8163b3abf3b1b163f760cd29d2a81b8fbec0b4edc449ed4b82bbc571e393c WatchSource:0}: Error finding container a4d8163b3abf3b1b163f760cd29d2a81b8fbec0b4edc449ed4b82bbc571e393c: Status 404 returned error can't find the container with id a4d8163b3abf3b1b163f760cd29d2a81b8fbec0b4edc449ed4b82bbc571e393c Nov 25 10:51:52 crc kubenswrapper[4821]: I1125 10:51:52.613557 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f78n9\" (UniqueName: \"kubernetes.io/projected/c41109a7-e905-4087-9d04-ccd2cccb42b5-kube-api-access-f78n9\") pod \"nova-scheduler-0\" (UID: \"c41109a7-e905-4087-9d04-ccd2cccb42b5\") " pod="openstack/nova-scheduler-0" Nov 25 10:51:52 crc kubenswrapper[4821]: I1125 10:51:52.613685 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c41109a7-e905-4087-9d04-ccd2cccb42b5-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c41109a7-e905-4087-9d04-ccd2cccb42b5\") " pod="openstack/nova-scheduler-0" Nov 25 10:51:52 crc kubenswrapper[4821]: I1125 10:51:52.613762 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c41109a7-e905-4087-9d04-ccd2cccb42b5-config-data\") pod \"nova-scheduler-0\" (UID: \"c41109a7-e905-4087-9d04-ccd2cccb42b5\") " pod="openstack/nova-scheduler-0" Nov 25 10:51:52 crc kubenswrapper[4821]: I1125 10:51:52.715199 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c41109a7-e905-4087-9d04-ccd2cccb42b5-config-data\") pod \"nova-scheduler-0\" (UID: \"c41109a7-e905-4087-9d04-ccd2cccb42b5\") " pod="openstack/nova-scheduler-0" Nov 25 10:51:52 crc kubenswrapper[4821]: I1125 10:51:52.715277 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f78n9\" (UniqueName: \"kubernetes.io/projected/c41109a7-e905-4087-9d04-ccd2cccb42b5-kube-api-access-f78n9\") pod \"nova-scheduler-0\" (UID: \"c41109a7-e905-4087-9d04-ccd2cccb42b5\") " pod="openstack/nova-scheduler-0" Nov 25 10:51:52 crc kubenswrapper[4821]: I1125 10:51:52.715403 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c41109a7-e905-4087-9d04-ccd2cccb42b5-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c41109a7-e905-4087-9d04-ccd2cccb42b5\") " pod="openstack/nova-scheduler-0" Nov 25 10:51:52 crc kubenswrapper[4821]: I1125 10:51:52.720965 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c41109a7-e905-4087-9d04-ccd2cccb42b5-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c41109a7-e905-4087-9d04-ccd2cccb42b5\") " pod="openstack/nova-scheduler-0" Nov 25 10:51:52 crc kubenswrapper[4821]: I1125 10:51:52.721951 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c41109a7-e905-4087-9d04-ccd2cccb42b5-config-data\") pod \"nova-scheduler-0\" (UID: \"c41109a7-e905-4087-9d04-ccd2cccb42b5\") " pod="openstack/nova-scheduler-0" Nov 25 10:51:52 crc kubenswrapper[4821]: I1125 10:51:52.734296 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f78n9\" (UniqueName: \"kubernetes.io/projected/c41109a7-e905-4087-9d04-ccd2cccb42b5-kube-api-access-f78n9\") pod \"nova-scheduler-0\" (UID: \"c41109a7-e905-4087-9d04-ccd2cccb42b5\") " pod="openstack/nova-scheduler-0" Nov 25 10:51:52 crc kubenswrapper[4821]: I1125 10:51:52.936660 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 25 10:51:53 crc kubenswrapper[4821]: W1125 10:51:53.416350 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc41109a7_e905_4087_9d04_ccd2cccb42b5.slice/crio-07af87afbb24e0995e8d0ac6066d64d946c2185a21a95bb878e0c864bd67a08d WatchSource:0}: Error finding container 07af87afbb24e0995e8d0ac6066d64d946c2185a21a95bb878e0c864bd67a08d: Status 404 returned error can't find the container with id 07af87afbb24e0995e8d0ac6066d64d946c2185a21a95bb878e0c864bd67a08d Nov 25 10:51:53 crc kubenswrapper[4821]: I1125 10:51:53.418668 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 10:51:53 crc kubenswrapper[4821]: I1125 10:51:53.441692 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c41109a7-e905-4087-9d04-ccd2cccb42b5","Type":"ContainerStarted","Data":"07af87afbb24e0995e8d0ac6066d64d946c2185a21a95bb878e0c864bd67a08d"} Nov 25 10:51:53 crc kubenswrapper[4821]: I1125 10:51:53.445251 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"31227d89-6706-407e-aa0a-e90b99611462","Type":"ContainerStarted","Data":"6142554888b4dd483754e7df3beec07af51e4423dedfc53e264667a070450805"} Nov 25 10:51:53 crc kubenswrapper[4821]: I1125 10:51:53.445289 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"31227d89-6706-407e-aa0a-e90b99611462","Type":"ContainerStarted","Data":"0f09f615de9732f25086805966d2a9e7cc0742d848f2c1da0f098ffdef5755a0"} Nov 25 10:51:53 crc kubenswrapper[4821]: I1125 10:51:53.445300 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"31227d89-6706-407e-aa0a-e90b99611462","Type":"ContainerStarted","Data":"526823676ee1f29b2682b8e13e0cb391cd014899aec4b55ef0d1c9a89f243a29"} Nov 25 10:51:53 crc kubenswrapper[4821]: I1125 10:51:53.449909 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"7c286dad-cd09-4085-a4e0-4b274f30d38d","Type":"ContainerStarted","Data":"a4d8163b3abf3b1b163f760cd29d2a81b8fbec0b4edc449ed4b82bbc571e393c"} Nov 25 10:51:53 crc kubenswrapper[4821]: I1125 10:51:53.455801 4821 generic.go:334] "Generic (PLEG): container finished" podID="93ebfd94-d02a-4a94-bd08-2fa977ece0c2" containerID="58854775c62d04df46d082b0b3e917a617f69ac9824ab4befff224cb93919dfa" exitCode=0 Nov 25 10:51:53 crc kubenswrapper[4821]: I1125 10:51:53.455839 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"93ebfd94-d02a-4a94-bd08-2fa977ece0c2","Type":"ContainerDied","Data":"58854775c62d04df46d082b0b3e917a617f69ac9824ab4befff224cb93919dfa"} Nov 25 10:51:53 crc kubenswrapper[4821]: I1125 10:51:53.470369 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.470352242 podStartE2EDuration="2.470352242s" podCreationTimestamp="2025-11-25 10:51:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:51:53.463278456 +0000 UTC m=+1183.999598303" watchObservedRunningTime="2025-11-25 10:51:53.470352242 +0000 UTC m=+1184.006672089" Nov 25 10:51:54 crc kubenswrapper[4821]: I1125 10:51:54.126027 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2bd16d9e-1992-4537-8376-c8d0da2f4a31" path="/var/lib/kubelet/pods/2bd16d9e-1992-4537-8376-c8d0da2f4a31/volumes" Nov 25 10:51:54 crc kubenswrapper[4821]: I1125 10:51:54.466798 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c41109a7-e905-4087-9d04-ccd2cccb42b5","Type":"ContainerStarted","Data":"f29d85ca7d099c678b1f559610a961965055b87a55a28a98b8c43a29b9e80703"} Nov 25 10:51:54 crc kubenswrapper[4821]: I1125 10:51:54.469618 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"7c286dad-cd09-4085-a4e0-4b274f30d38d","Type":"ContainerStarted","Data":"d00f4ed1f83aedc19737a80dd80b297e057c1f4ce45e54920e5fed36c203f41c"} Nov 25 10:51:54 crc kubenswrapper[4821]: I1125 10:51:54.469745 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 25 10:51:54 crc kubenswrapper[4821]: I1125 10:51:54.472458 4821 generic.go:334] "Generic (PLEG): container finished" podID="93ebfd94-d02a-4a94-bd08-2fa977ece0c2" containerID="4524ca39e676a4c2d2a0a5cd6233d1a9da964e9de331e24b23cf829c337da27a" exitCode=0 Nov 25 10:51:54 crc kubenswrapper[4821]: I1125 10:51:54.472483 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"93ebfd94-d02a-4a94-bd08-2fa977ece0c2","Type":"ContainerDied","Data":"4524ca39e676a4c2d2a0a5cd6233d1a9da964e9de331e24b23cf829c337da27a"} Nov 25 10:51:54 crc kubenswrapper[4821]: I1125 10:51:54.487423 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.487399974 podStartE2EDuration="2.487399974s" podCreationTimestamp="2025-11-25 10:51:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:51:54.480094502 +0000 UTC m=+1185.016414349" watchObservedRunningTime="2025-11-25 10:51:54.487399974 +0000 UTC m=+1185.023719821" Nov 25 10:51:54 crc kubenswrapper[4821]: I1125 10:51:54.510371 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.5451259569999998 podStartE2EDuration="3.510348708s" podCreationTimestamp="2025-11-25 10:51:51 +0000 UTC" firstStartedPulling="2025-11-25 10:51:52.555045003 +0000 UTC m=+1183.091364850" lastFinishedPulling="2025-11-25 10:51:53.520267754 +0000 UTC m=+1184.056587601" observedRunningTime="2025-11-25 10:51:54.500732105 +0000 UTC m=+1185.037051942" watchObservedRunningTime="2025-11-25 10:51:54.510348708 +0000 UTC m=+1185.046668555" Nov 25 10:51:54 crc kubenswrapper[4821]: I1125 10:51:54.729640 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:51:54 crc kubenswrapper[4821]: I1125 10:51:54.863589 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93ebfd94-d02a-4a94-bd08-2fa977ece0c2-combined-ca-bundle\") pod \"93ebfd94-d02a-4a94-bd08-2fa977ece0c2\" (UID: \"93ebfd94-d02a-4a94-bd08-2fa977ece0c2\") " Nov 25 10:51:54 crc kubenswrapper[4821]: I1125 10:51:54.863644 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/93ebfd94-d02a-4a94-bd08-2fa977ece0c2-log-httpd\") pod \"93ebfd94-d02a-4a94-bd08-2fa977ece0c2\" (UID: \"93ebfd94-d02a-4a94-bd08-2fa977ece0c2\") " Nov 25 10:51:54 crc kubenswrapper[4821]: I1125 10:51:54.863717 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g5fnr\" (UniqueName: \"kubernetes.io/projected/93ebfd94-d02a-4a94-bd08-2fa977ece0c2-kube-api-access-g5fnr\") pod \"93ebfd94-d02a-4a94-bd08-2fa977ece0c2\" (UID: \"93ebfd94-d02a-4a94-bd08-2fa977ece0c2\") " Nov 25 10:51:54 crc kubenswrapper[4821]: I1125 10:51:54.863735 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/93ebfd94-d02a-4a94-bd08-2fa977ece0c2-sg-core-conf-yaml\") pod \"93ebfd94-d02a-4a94-bd08-2fa977ece0c2\" (UID: \"93ebfd94-d02a-4a94-bd08-2fa977ece0c2\") " Nov 25 10:51:54 crc kubenswrapper[4821]: I1125 10:51:54.863823 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93ebfd94-d02a-4a94-bd08-2fa977ece0c2-config-data\") pod \"93ebfd94-d02a-4a94-bd08-2fa977ece0c2\" (UID: \"93ebfd94-d02a-4a94-bd08-2fa977ece0c2\") " Nov 25 10:51:54 crc kubenswrapper[4821]: I1125 10:51:54.863854 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93ebfd94-d02a-4a94-bd08-2fa977ece0c2-scripts\") pod \"93ebfd94-d02a-4a94-bd08-2fa977ece0c2\" (UID: \"93ebfd94-d02a-4a94-bd08-2fa977ece0c2\") " Nov 25 10:51:54 crc kubenswrapper[4821]: I1125 10:51:54.863917 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/93ebfd94-d02a-4a94-bd08-2fa977ece0c2-run-httpd\") pod \"93ebfd94-d02a-4a94-bd08-2fa977ece0c2\" (UID: \"93ebfd94-d02a-4a94-bd08-2fa977ece0c2\") " Nov 25 10:51:54 crc kubenswrapper[4821]: I1125 10:51:54.864723 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/93ebfd94-d02a-4a94-bd08-2fa977ece0c2-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "93ebfd94-d02a-4a94-bd08-2fa977ece0c2" (UID: "93ebfd94-d02a-4a94-bd08-2fa977ece0c2"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:51:54 crc kubenswrapper[4821]: I1125 10:51:54.865630 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/93ebfd94-d02a-4a94-bd08-2fa977ece0c2-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "93ebfd94-d02a-4a94-bd08-2fa977ece0c2" (UID: "93ebfd94-d02a-4a94-bd08-2fa977ece0c2"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:51:54 crc kubenswrapper[4821]: I1125 10:51:54.869129 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93ebfd94-d02a-4a94-bd08-2fa977ece0c2-scripts" (OuterVolumeSpecName: "scripts") pod "93ebfd94-d02a-4a94-bd08-2fa977ece0c2" (UID: "93ebfd94-d02a-4a94-bd08-2fa977ece0c2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:51:54 crc kubenswrapper[4821]: I1125 10:51:54.869410 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93ebfd94-d02a-4a94-bd08-2fa977ece0c2-kube-api-access-g5fnr" (OuterVolumeSpecName: "kube-api-access-g5fnr") pod "93ebfd94-d02a-4a94-bd08-2fa977ece0c2" (UID: "93ebfd94-d02a-4a94-bd08-2fa977ece0c2"). InnerVolumeSpecName "kube-api-access-g5fnr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:51:54 crc kubenswrapper[4821]: I1125 10:51:54.912697 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93ebfd94-d02a-4a94-bd08-2fa977ece0c2-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "93ebfd94-d02a-4a94-bd08-2fa977ece0c2" (UID: "93ebfd94-d02a-4a94-bd08-2fa977ece0c2"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:51:54 crc kubenswrapper[4821]: I1125 10:51:54.950470 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93ebfd94-d02a-4a94-bd08-2fa977ece0c2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "93ebfd94-d02a-4a94-bd08-2fa977ece0c2" (UID: "93ebfd94-d02a-4a94-bd08-2fa977ece0c2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:51:54 crc kubenswrapper[4821]: I1125 10:51:54.966482 4821 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93ebfd94-d02a-4a94-bd08-2fa977ece0c2-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:51:54 crc kubenswrapper[4821]: I1125 10:51:54.966725 4821 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/93ebfd94-d02a-4a94-bd08-2fa977ece0c2-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 10:51:54 crc kubenswrapper[4821]: I1125 10:51:54.966796 4821 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93ebfd94-d02a-4a94-bd08-2fa977ece0c2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:51:54 crc kubenswrapper[4821]: I1125 10:51:54.966836 4821 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/93ebfd94-d02a-4a94-bd08-2fa977ece0c2-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 10:51:54 crc kubenswrapper[4821]: I1125 10:51:54.966848 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g5fnr\" (UniqueName: \"kubernetes.io/projected/93ebfd94-d02a-4a94-bd08-2fa977ece0c2-kube-api-access-g5fnr\") on node \"crc\" DevicePath \"\"" Nov 25 10:51:54 crc kubenswrapper[4821]: I1125 10:51:54.966861 4821 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/93ebfd94-d02a-4a94-bd08-2fa977ece0c2-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 25 10:51:54 crc kubenswrapper[4821]: I1125 10:51:54.992387 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93ebfd94-d02a-4a94-bd08-2fa977ece0c2-config-data" (OuterVolumeSpecName: "config-data") pod "93ebfd94-d02a-4a94-bd08-2fa977ece0c2" (UID: "93ebfd94-d02a-4a94-bd08-2fa977ece0c2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:51:55 crc kubenswrapper[4821]: I1125 10:51:55.068984 4821 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93ebfd94-d02a-4a94-bd08-2fa977ece0c2-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:51:55 crc kubenswrapper[4821]: I1125 10:51:55.484084 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:51:55 crc kubenswrapper[4821]: I1125 10:51:55.490344 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"93ebfd94-d02a-4a94-bd08-2fa977ece0c2","Type":"ContainerDied","Data":"02ce8cb24bca19c9c9402a011d797633829934b981dcce8b74b6a27133ad4607"} Nov 25 10:51:55 crc kubenswrapper[4821]: I1125 10:51:55.490413 4821 scope.go:117] "RemoveContainer" containerID="7f092fdc0f9a31f25a3a369e3b7e07011c4b9033d12d7a977abf4cf3161f86a1" Nov 25 10:51:55 crc kubenswrapper[4821]: I1125 10:51:55.524309 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:51:55 crc kubenswrapper[4821]: I1125 10:51:55.535388 4821 scope.go:117] "RemoveContainer" containerID="5651bfa9d025b88a03615fdbcd8d8ab5c0354dcbfa711ef2309fa1144334340c" Nov 25 10:51:55 crc kubenswrapper[4821]: I1125 10:51:55.549678 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:51:55 crc kubenswrapper[4821]: I1125 10:51:55.557631 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:51:55 crc kubenswrapper[4821]: E1125 10:51:55.558045 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93ebfd94-d02a-4a94-bd08-2fa977ece0c2" containerName="ceilometer-central-agent" Nov 25 10:51:55 crc kubenswrapper[4821]: I1125 10:51:55.558061 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="93ebfd94-d02a-4a94-bd08-2fa977ece0c2" containerName="ceilometer-central-agent" Nov 25 10:51:55 crc kubenswrapper[4821]: E1125 10:51:55.558083 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93ebfd94-d02a-4a94-bd08-2fa977ece0c2" containerName="sg-core" Nov 25 10:51:55 crc kubenswrapper[4821]: I1125 10:51:55.558089 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="93ebfd94-d02a-4a94-bd08-2fa977ece0c2" containerName="sg-core" Nov 25 10:51:55 crc kubenswrapper[4821]: E1125 10:51:55.558103 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93ebfd94-d02a-4a94-bd08-2fa977ece0c2" containerName="proxy-httpd" Nov 25 10:51:55 crc kubenswrapper[4821]: I1125 10:51:55.558108 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="93ebfd94-d02a-4a94-bd08-2fa977ece0c2" containerName="proxy-httpd" Nov 25 10:51:55 crc kubenswrapper[4821]: E1125 10:51:55.558120 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93ebfd94-d02a-4a94-bd08-2fa977ece0c2" containerName="ceilometer-notification-agent" Nov 25 10:51:55 crc kubenswrapper[4821]: I1125 10:51:55.558126 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="93ebfd94-d02a-4a94-bd08-2fa977ece0c2" containerName="ceilometer-notification-agent" Nov 25 10:51:55 crc kubenswrapper[4821]: I1125 10:51:55.558344 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="93ebfd94-d02a-4a94-bd08-2fa977ece0c2" containerName="ceilometer-central-agent" Nov 25 10:51:55 crc kubenswrapper[4821]: I1125 10:51:55.558361 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="93ebfd94-d02a-4a94-bd08-2fa977ece0c2" containerName="proxy-httpd" Nov 25 10:51:55 crc kubenswrapper[4821]: I1125 10:51:55.558369 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="93ebfd94-d02a-4a94-bd08-2fa977ece0c2" containerName="ceilometer-notification-agent" Nov 25 10:51:55 crc kubenswrapper[4821]: I1125 10:51:55.558386 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="93ebfd94-d02a-4a94-bd08-2fa977ece0c2" containerName="sg-core" Nov 25 10:51:55 crc kubenswrapper[4821]: I1125 10:51:55.560249 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:51:55 crc kubenswrapper[4821]: I1125 10:51:55.562806 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 25 10:51:55 crc kubenswrapper[4821]: I1125 10:51:55.563063 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 25 10:51:55 crc kubenswrapper[4821]: I1125 10:51:55.568429 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 25 10:51:55 crc kubenswrapper[4821]: I1125 10:51:55.572181 4821 scope.go:117] "RemoveContainer" containerID="4524ca39e676a4c2d2a0a5cd6233d1a9da964e9de331e24b23cf829c337da27a" Nov 25 10:51:55 crc kubenswrapper[4821]: I1125 10:51:55.573591 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:51:55 crc kubenswrapper[4821]: I1125 10:51:55.605638 4821 scope.go:117] "RemoveContainer" containerID="58854775c62d04df46d082b0b3e917a617f69ac9824ab4befff224cb93919dfa" Nov 25 10:51:55 crc kubenswrapper[4821]: I1125 10:51:55.682068 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a-log-httpd\") pod \"ceilometer-0\" (UID: \"8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a\") " pod="openstack/ceilometer-0" Nov 25 10:51:55 crc kubenswrapper[4821]: I1125 10:51:55.682245 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a-scripts\") pod \"ceilometer-0\" (UID: \"8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a\") " pod="openstack/ceilometer-0" Nov 25 10:51:55 crc kubenswrapper[4821]: I1125 10:51:55.682309 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a-run-httpd\") pod \"ceilometer-0\" (UID: \"8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a\") " pod="openstack/ceilometer-0" Nov 25 10:51:55 crc kubenswrapper[4821]: I1125 10:51:55.682559 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a\") " pod="openstack/ceilometer-0" Nov 25 10:51:55 crc kubenswrapper[4821]: I1125 10:51:55.682611 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a\") " pod="openstack/ceilometer-0" Nov 25 10:51:55 crc kubenswrapper[4821]: I1125 10:51:55.682744 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a-config-data\") pod \"ceilometer-0\" (UID: \"8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a\") " pod="openstack/ceilometer-0" Nov 25 10:51:55 crc kubenswrapper[4821]: I1125 10:51:55.682869 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-scssg\" (UniqueName: \"kubernetes.io/projected/8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a-kube-api-access-scssg\") pod \"ceilometer-0\" (UID: \"8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a\") " pod="openstack/ceilometer-0" Nov 25 10:51:55 crc kubenswrapper[4821]: I1125 10:51:55.682972 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a\") " pod="openstack/ceilometer-0" Nov 25 10:51:55 crc kubenswrapper[4821]: I1125 10:51:55.784560 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-scssg\" (UniqueName: \"kubernetes.io/projected/8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a-kube-api-access-scssg\") pod \"ceilometer-0\" (UID: \"8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a\") " pod="openstack/ceilometer-0" Nov 25 10:51:55 crc kubenswrapper[4821]: I1125 10:51:55.784656 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a\") " pod="openstack/ceilometer-0" Nov 25 10:51:55 crc kubenswrapper[4821]: I1125 10:51:55.784746 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a-log-httpd\") pod \"ceilometer-0\" (UID: \"8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a\") " pod="openstack/ceilometer-0" Nov 25 10:51:55 crc kubenswrapper[4821]: I1125 10:51:55.784803 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a-scripts\") pod \"ceilometer-0\" (UID: \"8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a\") " pod="openstack/ceilometer-0" Nov 25 10:51:55 crc kubenswrapper[4821]: I1125 10:51:55.784831 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a-run-httpd\") pod \"ceilometer-0\" (UID: \"8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a\") " pod="openstack/ceilometer-0" Nov 25 10:51:55 crc kubenswrapper[4821]: I1125 10:51:55.784914 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a\") " pod="openstack/ceilometer-0" Nov 25 10:51:55 crc kubenswrapper[4821]: I1125 10:51:55.784993 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a\") " pod="openstack/ceilometer-0" Nov 25 10:51:55 crc kubenswrapper[4821]: I1125 10:51:55.785075 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a-config-data\") pod \"ceilometer-0\" (UID: \"8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a\") " pod="openstack/ceilometer-0" Nov 25 10:51:55 crc kubenswrapper[4821]: I1125 10:51:55.785517 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a-log-httpd\") pod \"ceilometer-0\" (UID: \"8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a\") " pod="openstack/ceilometer-0" Nov 25 10:51:55 crc kubenswrapper[4821]: I1125 10:51:55.785957 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a-run-httpd\") pod \"ceilometer-0\" (UID: \"8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a\") " pod="openstack/ceilometer-0" Nov 25 10:51:55 crc kubenswrapper[4821]: I1125 10:51:55.789686 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a\") " pod="openstack/ceilometer-0" Nov 25 10:51:55 crc kubenswrapper[4821]: I1125 10:51:55.790352 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a\") " pod="openstack/ceilometer-0" Nov 25 10:51:55 crc kubenswrapper[4821]: I1125 10:51:55.790421 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a-scripts\") pod \"ceilometer-0\" (UID: \"8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a\") " pod="openstack/ceilometer-0" Nov 25 10:51:55 crc kubenswrapper[4821]: I1125 10:51:55.790439 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a\") " pod="openstack/ceilometer-0" Nov 25 10:51:55 crc kubenswrapper[4821]: I1125 10:51:55.790945 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a-config-data\") pod \"ceilometer-0\" (UID: \"8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a\") " pod="openstack/ceilometer-0" Nov 25 10:51:55 crc kubenswrapper[4821]: I1125 10:51:55.812887 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-scssg\" (UniqueName: \"kubernetes.io/projected/8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a-kube-api-access-scssg\") pod \"ceilometer-0\" (UID: \"8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a\") " pod="openstack/ceilometer-0" Nov 25 10:51:55 crc kubenswrapper[4821]: I1125 10:51:55.882803 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:51:56 crc kubenswrapper[4821]: I1125 10:51:56.126592 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93ebfd94-d02a-4a94-bd08-2fa977ece0c2" path="/var/lib/kubelet/pods/93ebfd94-d02a-4a94-bd08-2fa977ece0c2/volumes" Nov 25 10:51:56 crc kubenswrapper[4821]: I1125 10:51:56.383460 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:51:56 crc kubenswrapper[4821]: W1125 10:51:56.386907 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8d0e63e3_7a64_4127_9df2_5c7f00fa4d2a.slice/crio-70e100b8fb0bfaf3ab5658cb324558d472ffab7e5bb4793012ad97e235a36311 WatchSource:0}: Error finding container 70e100b8fb0bfaf3ab5658cb324558d472ffab7e5bb4793012ad97e235a36311: Status 404 returned error can't find the container with id 70e100b8fb0bfaf3ab5658cb324558d472ffab7e5bb4793012ad97e235a36311 Nov 25 10:51:56 crc kubenswrapper[4821]: I1125 10:51:56.494415 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a","Type":"ContainerStarted","Data":"70e100b8fb0bfaf3ab5658cb324558d472ffab7e5bb4793012ad97e235a36311"} Nov 25 10:51:57 crc kubenswrapper[4821]: I1125 10:51:57.499735 4821 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 10:51:57 crc kubenswrapper[4821]: I1125 10:51:57.507622 4821 generic.go:334] "Generic (PLEG): container finished" podID="e22e60b5-5698-493b-9a3f-8cfdf289c6d9" containerID="0c2fa815c3e17265e814630a7030173763f041b2d9166a413539261379e799be" exitCode=0 Nov 25 10:51:57 crc kubenswrapper[4821]: I1125 10:51:57.507695 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-vwzcd" event={"ID":"e22e60b5-5698-493b-9a3f-8cfdf289c6d9","Type":"ContainerDied","Data":"0c2fa815c3e17265e814630a7030173763f041b2d9166a413539261379e799be"} Nov 25 10:51:57 crc kubenswrapper[4821]: I1125 10:51:57.509722 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a","Type":"ContainerStarted","Data":"4ab8fdfa62f13fbd1abb83257ca5cdeba846d33627a9e30d10610757e61b64d7"} Nov 25 10:51:57 crc kubenswrapper[4821]: I1125 10:51:57.937376 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 25 10:51:58 crc kubenswrapper[4821]: I1125 10:51:58.527509 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a","Type":"ContainerStarted","Data":"7e6f9aabb7e768eb0a92de8459ab86f56405e962404785b3975090f9a2bb2cb6"} Nov 25 10:51:58 crc kubenswrapper[4821]: I1125 10:51:58.970830 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-vwzcd" Nov 25 10:51:59 crc kubenswrapper[4821]: I1125 10:51:59.043909 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e22e60b5-5698-493b-9a3f-8cfdf289c6d9-config-data\") pod \"e22e60b5-5698-493b-9a3f-8cfdf289c6d9\" (UID: \"e22e60b5-5698-493b-9a3f-8cfdf289c6d9\") " Nov 25 10:51:59 crc kubenswrapper[4821]: I1125 10:51:59.044156 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f9h9w\" (UniqueName: \"kubernetes.io/projected/e22e60b5-5698-493b-9a3f-8cfdf289c6d9-kube-api-access-f9h9w\") pod \"e22e60b5-5698-493b-9a3f-8cfdf289c6d9\" (UID: \"e22e60b5-5698-493b-9a3f-8cfdf289c6d9\") " Nov 25 10:51:59 crc kubenswrapper[4821]: I1125 10:51:59.044260 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e22e60b5-5698-493b-9a3f-8cfdf289c6d9-scripts\") pod \"e22e60b5-5698-493b-9a3f-8cfdf289c6d9\" (UID: \"e22e60b5-5698-493b-9a3f-8cfdf289c6d9\") " Nov 25 10:51:59 crc kubenswrapper[4821]: I1125 10:51:59.044284 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e22e60b5-5698-493b-9a3f-8cfdf289c6d9-combined-ca-bundle\") pod \"e22e60b5-5698-493b-9a3f-8cfdf289c6d9\" (UID: \"e22e60b5-5698-493b-9a3f-8cfdf289c6d9\") " Nov 25 10:51:59 crc kubenswrapper[4821]: I1125 10:51:59.049576 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e22e60b5-5698-493b-9a3f-8cfdf289c6d9-kube-api-access-f9h9w" (OuterVolumeSpecName: "kube-api-access-f9h9w") pod "e22e60b5-5698-493b-9a3f-8cfdf289c6d9" (UID: "e22e60b5-5698-493b-9a3f-8cfdf289c6d9"). InnerVolumeSpecName "kube-api-access-f9h9w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:51:59 crc kubenswrapper[4821]: I1125 10:51:59.049981 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e22e60b5-5698-493b-9a3f-8cfdf289c6d9-scripts" (OuterVolumeSpecName: "scripts") pod "e22e60b5-5698-493b-9a3f-8cfdf289c6d9" (UID: "e22e60b5-5698-493b-9a3f-8cfdf289c6d9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:51:59 crc kubenswrapper[4821]: I1125 10:51:59.078650 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e22e60b5-5698-493b-9a3f-8cfdf289c6d9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e22e60b5-5698-493b-9a3f-8cfdf289c6d9" (UID: "e22e60b5-5698-493b-9a3f-8cfdf289c6d9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:51:59 crc kubenswrapper[4821]: I1125 10:51:59.082453 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e22e60b5-5698-493b-9a3f-8cfdf289c6d9-config-data" (OuterVolumeSpecName: "config-data") pod "e22e60b5-5698-493b-9a3f-8cfdf289c6d9" (UID: "e22e60b5-5698-493b-9a3f-8cfdf289c6d9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:51:59 crc kubenswrapper[4821]: I1125 10:51:59.146268 4821 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e22e60b5-5698-493b-9a3f-8cfdf289c6d9-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:51:59 crc kubenswrapper[4821]: I1125 10:51:59.146303 4821 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e22e60b5-5698-493b-9a3f-8cfdf289c6d9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:51:59 crc kubenswrapper[4821]: I1125 10:51:59.146317 4821 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e22e60b5-5698-493b-9a3f-8cfdf289c6d9-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:51:59 crc kubenswrapper[4821]: I1125 10:51:59.146325 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f9h9w\" (UniqueName: \"kubernetes.io/projected/e22e60b5-5698-493b-9a3f-8cfdf289c6d9-kube-api-access-f9h9w\") on node \"crc\" DevicePath \"\"" Nov 25 10:51:59 crc kubenswrapper[4821]: I1125 10:51:59.540730 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a","Type":"ContainerStarted","Data":"082c0727decabeea221fa43abfb8945aa71d5bc81de092541d63970c64ce4ac9"} Nov 25 10:51:59 crc kubenswrapper[4821]: I1125 10:51:59.544093 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-vwzcd" event={"ID":"e22e60b5-5698-493b-9a3f-8cfdf289c6d9","Type":"ContainerDied","Data":"511a20d2922d8228a6ff5ae595c75af89e038e1d037dead5400c48d1c09b4edb"} Nov 25 10:51:59 crc kubenswrapper[4821]: I1125 10:51:59.544120 4821 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="511a20d2922d8228a6ff5ae595c75af89e038e1d037dead5400c48d1c09b4edb" Nov 25 10:51:59 crc kubenswrapper[4821]: I1125 10:51:59.544182 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-vwzcd" Nov 25 10:51:59 crc kubenswrapper[4821]: I1125 10:51:59.625949 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 25 10:51:59 crc kubenswrapper[4821]: E1125 10:51:59.626433 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e22e60b5-5698-493b-9a3f-8cfdf289c6d9" containerName="nova-cell1-conductor-db-sync" Nov 25 10:51:59 crc kubenswrapper[4821]: I1125 10:51:59.626451 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="e22e60b5-5698-493b-9a3f-8cfdf289c6d9" containerName="nova-cell1-conductor-db-sync" Nov 25 10:51:59 crc kubenswrapper[4821]: I1125 10:51:59.626670 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="e22e60b5-5698-493b-9a3f-8cfdf289c6d9" containerName="nova-cell1-conductor-db-sync" Nov 25 10:51:59 crc kubenswrapper[4821]: I1125 10:51:59.627310 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 25 10:51:59 crc kubenswrapper[4821]: I1125 10:51:59.630889 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 25 10:51:59 crc kubenswrapper[4821]: I1125 10:51:59.662768 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 25 10:51:59 crc kubenswrapper[4821]: I1125 10:51:59.766312 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dsptm\" (UniqueName: \"kubernetes.io/projected/fb540939-5683-495a-b8df-d7a000fedaa2-kube-api-access-dsptm\") pod \"nova-cell1-conductor-0\" (UID: \"fb540939-5683-495a-b8df-d7a000fedaa2\") " pod="openstack/nova-cell1-conductor-0" Nov 25 10:51:59 crc kubenswrapper[4821]: I1125 10:51:59.766556 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb540939-5683-495a-b8df-d7a000fedaa2-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"fb540939-5683-495a-b8df-d7a000fedaa2\") " pod="openstack/nova-cell1-conductor-0" Nov 25 10:51:59 crc kubenswrapper[4821]: I1125 10:51:59.766623 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb540939-5683-495a-b8df-d7a000fedaa2-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"fb540939-5683-495a-b8df-d7a000fedaa2\") " pod="openstack/nova-cell1-conductor-0" Nov 25 10:51:59 crc kubenswrapper[4821]: I1125 10:51:59.868083 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dsptm\" (UniqueName: \"kubernetes.io/projected/fb540939-5683-495a-b8df-d7a000fedaa2-kube-api-access-dsptm\") pod \"nova-cell1-conductor-0\" (UID: \"fb540939-5683-495a-b8df-d7a000fedaa2\") " pod="openstack/nova-cell1-conductor-0" Nov 25 10:51:59 crc kubenswrapper[4821]: I1125 10:51:59.868195 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb540939-5683-495a-b8df-d7a000fedaa2-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"fb540939-5683-495a-b8df-d7a000fedaa2\") " pod="openstack/nova-cell1-conductor-0" Nov 25 10:51:59 crc kubenswrapper[4821]: I1125 10:51:59.868219 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb540939-5683-495a-b8df-d7a000fedaa2-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"fb540939-5683-495a-b8df-d7a000fedaa2\") " pod="openstack/nova-cell1-conductor-0" Nov 25 10:51:59 crc kubenswrapper[4821]: I1125 10:51:59.873233 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb540939-5683-495a-b8df-d7a000fedaa2-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"fb540939-5683-495a-b8df-d7a000fedaa2\") " pod="openstack/nova-cell1-conductor-0" Nov 25 10:51:59 crc kubenswrapper[4821]: I1125 10:51:59.875953 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb540939-5683-495a-b8df-d7a000fedaa2-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"fb540939-5683-495a-b8df-d7a000fedaa2\") " pod="openstack/nova-cell1-conductor-0" Nov 25 10:51:59 crc kubenswrapper[4821]: I1125 10:51:59.894594 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dsptm\" (UniqueName: \"kubernetes.io/projected/fb540939-5683-495a-b8df-d7a000fedaa2-kube-api-access-dsptm\") pod \"nova-cell1-conductor-0\" (UID: \"fb540939-5683-495a-b8df-d7a000fedaa2\") " pod="openstack/nova-cell1-conductor-0" Nov 25 10:51:59 crc kubenswrapper[4821]: I1125 10:51:59.955696 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 25 10:52:00 crc kubenswrapper[4821]: I1125 10:52:00.451371 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 25 10:52:00 crc kubenswrapper[4821]: I1125 10:52:00.554094 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a","Type":"ContainerStarted","Data":"7d1fd9f582251094b332ecec2b092c8237aaa47bb73dc61f7e9258c72e486ecd"} Nov 25 10:52:00 crc kubenswrapper[4821]: I1125 10:52:00.554251 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 25 10:52:00 crc kubenswrapper[4821]: I1125 10:52:00.555581 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"fb540939-5683-495a-b8df-d7a000fedaa2","Type":"ContainerStarted","Data":"8b18f4e81501678d4f2da7050fee44c96c2b1b94cb5856d744c94bd183541cb5"} Nov 25 10:52:00 crc kubenswrapper[4821]: I1125 10:52:00.582417 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.7245025250000001 podStartE2EDuration="5.582402609s" podCreationTimestamp="2025-11-25 10:51:55 +0000 UTC" firstStartedPulling="2025-11-25 10:51:56.389633034 +0000 UTC m=+1186.925952881" lastFinishedPulling="2025-11-25 10:52:00.247533118 +0000 UTC m=+1190.783852965" observedRunningTime="2025-11-25 10:52:00.581009713 +0000 UTC m=+1191.117329570" watchObservedRunningTime="2025-11-25 10:52:00.582402609 +0000 UTC m=+1191.118722456" Nov 25 10:52:01 crc kubenswrapper[4821]: I1125 10:52:01.565260 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"fb540939-5683-495a-b8df-d7a000fedaa2","Type":"ContainerStarted","Data":"8521d6fbc4f0880bbc98e4d94d971389a88ee99f59e6d41fe4edde4a8c8a52ee"} Nov 25 10:52:01 crc kubenswrapper[4821]: I1125 10:52:01.589430 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.5894106690000003 podStartE2EDuration="2.589410669s" podCreationTimestamp="2025-11-25 10:51:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:52:01.586152603 +0000 UTC m=+1192.122472450" watchObservedRunningTime="2025-11-25 10:52:01.589410669 +0000 UTC m=+1192.125730516" Nov 25 10:52:01 crc kubenswrapper[4821]: I1125 10:52:01.882132 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 25 10:52:01 crc kubenswrapper[4821]: I1125 10:52:01.882202 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 25 10:52:01 crc kubenswrapper[4821]: I1125 10:52:01.933661 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 25 10:52:02 crc kubenswrapper[4821]: I1125 10:52:02.574744 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Nov 25 10:52:02 crc kubenswrapper[4821]: I1125 10:52:02.937122 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 25 10:52:02 crc kubenswrapper[4821]: I1125 10:52:02.964876 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 25 10:52:02 crc kubenswrapper[4821]: I1125 10:52:02.966585 4821 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="31227d89-6706-407e-aa0a-e90b99611462" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.192:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 25 10:52:02 crc kubenswrapper[4821]: I1125 10:52:02.966749 4821 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="31227d89-6706-407e-aa0a-e90b99611462" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.192:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 25 10:52:03 crc kubenswrapper[4821]: I1125 10:52:03.627979 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 25 10:52:04 crc kubenswrapper[4821]: I1125 10:52:04.714445 4821 patch_prober.go:28] interesting pod/machine-config-daemon-2krbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:52:04 crc kubenswrapper[4821]: I1125 10:52:04.714794 4821 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:52:04 crc kubenswrapper[4821]: I1125 10:52:04.714842 4821 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" Nov 25 10:52:04 crc kubenswrapper[4821]: I1125 10:52:04.715543 4821 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8e91d42cd72af932de083af71bf0c37ad5242c32b692e67d8b88f724fd1c26a8"} pod="openshift-machine-config-operator/machine-config-daemon-2krbf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 10:52:04 crc kubenswrapper[4821]: I1125 10:52:04.715612 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerName="machine-config-daemon" containerID="cri-o://8e91d42cd72af932de083af71bf0c37ad5242c32b692e67d8b88f724fd1c26a8" gracePeriod=600 Nov 25 10:52:05 crc kubenswrapper[4821]: I1125 10:52:05.600961 4821 generic.go:334] "Generic (PLEG): container finished" podID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerID="8e91d42cd72af932de083af71bf0c37ad5242c32b692e67d8b88f724fd1c26a8" exitCode=0 Nov 25 10:52:05 crc kubenswrapper[4821]: I1125 10:52:05.601000 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" event={"ID":"5f948b87-ac86-4de6-ad64-c2ef947f84d4","Type":"ContainerDied","Data":"8e91d42cd72af932de083af71bf0c37ad5242c32b692e67d8b88f724fd1c26a8"} Nov 25 10:52:05 crc kubenswrapper[4821]: I1125 10:52:05.601332 4821 scope.go:117] "RemoveContainer" containerID="c1742b8b51dd841b8f5beefd1352b72fe8669e2095b0356503a696291b279fa5" Nov 25 10:52:06 crc kubenswrapper[4821]: I1125 10:52:06.614322 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" event={"ID":"5f948b87-ac86-4de6-ad64-c2ef947f84d4","Type":"ContainerStarted","Data":"1cba28a33f40c656b6dedea9b9df6fa03370b5e6b1aff86ef7fcc571ea800429"} Nov 25 10:52:09 crc kubenswrapper[4821]: I1125 10:52:09.996586 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Nov 25 10:52:11 crc kubenswrapper[4821]: I1125 10:52:11.885411 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 25 10:52:11 crc kubenswrapper[4821]: I1125 10:52:11.886491 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 25 10:52:11 crc kubenswrapper[4821]: I1125 10:52:11.888373 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 25 10:52:11 crc kubenswrapper[4821]: I1125 10:52:11.890542 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 25 10:52:12 crc kubenswrapper[4821]: I1125 10:52:12.696807 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 25 10:52:12 crc kubenswrapper[4821]: I1125 10:52:12.700114 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 25 10:52:12 crc kubenswrapper[4821]: I1125 10:52:12.899226 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-xd4pf"] Nov 25 10:52:12 crc kubenswrapper[4821]: I1125 10:52:12.904512 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-xd4pf" Nov 25 10:52:12 crc kubenswrapper[4821]: I1125 10:52:12.936761 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-xd4pf"] Nov 25 10:52:12 crc kubenswrapper[4821]: I1125 10:52:12.946556 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db92740e-3cd9-4b81-b153-3f866fd3db62-config\") pod \"dnsmasq-dns-cd5cbd7b9-xd4pf\" (UID: \"db92740e-3cd9-4b81-b153-3f866fd3db62\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-xd4pf" Nov 25 10:52:12 crc kubenswrapper[4821]: I1125 10:52:12.946600 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/db92740e-3cd9-4b81-b153-3f866fd3db62-ovsdbserver-sb\") pod \"dnsmasq-dns-cd5cbd7b9-xd4pf\" (UID: \"db92740e-3cd9-4b81-b153-3f866fd3db62\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-xd4pf" Nov 25 10:52:12 crc kubenswrapper[4821]: I1125 10:52:12.946677 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42sws\" (UniqueName: \"kubernetes.io/projected/db92740e-3cd9-4b81-b153-3f866fd3db62-kube-api-access-42sws\") pod \"dnsmasq-dns-cd5cbd7b9-xd4pf\" (UID: \"db92740e-3cd9-4b81-b153-3f866fd3db62\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-xd4pf" Nov 25 10:52:12 crc kubenswrapper[4821]: I1125 10:52:12.946697 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/db92740e-3cd9-4b81-b153-3f866fd3db62-dns-swift-storage-0\") pod \"dnsmasq-dns-cd5cbd7b9-xd4pf\" (UID: \"db92740e-3cd9-4b81-b153-3f866fd3db62\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-xd4pf" Nov 25 10:52:12 crc kubenswrapper[4821]: I1125 10:52:12.946759 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/db92740e-3cd9-4b81-b153-3f866fd3db62-ovsdbserver-nb\") pod \"dnsmasq-dns-cd5cbd7b9-xd4pf\" (UID: \"db92740e-3cd9-4b81-b153-3f866fd3db62\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-xd4pf" Nov 25 10:52:12 crc kubenswrapper[4821]: I1125 10:52:12.946782 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/db92740e-3cd9-4b81-b153-3f866fd3db62-dns-svc\") pod \"dnsmasq-dns-cd5cbd7b9-xd4pf\" (UID: \"db92740e-3cd9-4b81-b153-3f866fd3db62\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-xd4pf" Nov 25 10:52:13 crc kubenswrapper[4821]: I1125 10:52:13.566639 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/db92740e-3cd9-4b81-b153-3f866fd3db62-ovsdbserver-nb\") pod \"dnsmasq-dns-cd5cbd7b9-xd4pf\" (UID: \"db92740e-3cd9-4b81-b153-3f866fd3db62\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-xd4pf" Nov 25 10:52:13 crc kubenswrapper[4821]: I1125 10:52:13.566879 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/db92740e-3cd9-4b81-b153-3f866fd3db62-dns-svc\") pod \"dnsmasq-dns-cd5cbd7b9-xd4pf\" (UID: \"db92740e-3cd9-4b81-b153-3f866fd3db62\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-xd4pf" Nov 25 10:52:13 crc kubenswrapper[4821]: I1125 10:52:13.566924 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db92740e-3cd9-4b81-b153-3f866fd3db62-config\") pod \"dnsmasq-dns-cd5cbd7b9-xd4pf\" (UID: \"db92740e-3cd9-4b81-b153-3f866fd3db62\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-xd4pf" Nov 25 10:52:13 crc kubenswrapper[4821]: I1125 10:52:13.566945 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/db92740e-3cd9-4b81-b153-3f866fd3db62-ovsdbserver-sb\") pod \"dnsmasq-dns-cd5cbd7b9-xd4pf\" (UID: \"db92740e-3cd9-4b81-b153-3f866fd3db62\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-xd4pf" Nov 25 10:52:13 crc kubenswrapper[4821]: I1125 10:52:13.567018 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42sws\" (UniqueName: \"kubernetes.io/projected/db92740e-3cd9-4b81-b153-3f866fd3db62-kube-api-access-42sws\") pod \"dnsmasq-dns-cd5cbd7b9-xd4pf\" (UID: \"db92740e-3cd9-4b81-b153-3f866fd3db62\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-xd4pf" Nov 25 10:52:13 crc kubenswrapper[4821]: I1125 10:52:13.567039 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/db92740e-3cd9-4b81-b153-3f866fd3db62-dns-swift-storage-0\") pod \"dnsmasq-dns-cd5cbd7b9-xd4pf\" (UID: \"db92740e-3cd9-4b81-b153-3f866fd3db62\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-xd4pf" Nov 25 10:52:13 crc kubenswrapper[4821]: I1125 10:52:13.567941 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/db92740e-3cd9-4b81-b153-3f866fd3db62-dns-swift-storage-0\") pod \"dnsmasq-dns-cd5cbd7b9-xd4pf\" (UID: \"db92740e-3cd9-4b81-b153-3f866fd3db62\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-xd4pf" Nov 25 10:52:13 crc kubenswrapper[4821]: I1125 10:52:13.568891 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/db92740e-3cd9-4b81-b153-3f866fd3db62-ovsdbserver-nb\") pod \"dnsmasq-dns-cd5cbd7b9-xd4pf\" (UID: \"db92740e-3cd9-4b81-b153-3f866fd3db62\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-xd4pf" Nov 25 10:52:13 crc kubenswrapper[4821]: I1125 10:52:13.569486 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/db92740e-3cd9-4b81-b153-3f866fd3db62-dns-svc\") pod \"dnsmasq-dns-cd5cbd7b9-xd4pf\" (UID: \"db92740e-3cd9-4b81-b153-3f866fd3db62\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-xd4pf" Nov 25 10:52:13 crc kubenswrapper[4821]: I1125 10:52:13.570013 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db92740e-3cd9-4b81-b153-3f866fd3db62-config\") pod \"dnsmasq-dns-cd5cbd7b9-xd4pf\" (UID: \"db92740e-3cd9-4b81-b153-3f866fd3db62\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-xd4pf" Nov 25 10:52:13 crc kubenswrapper[4821]: I1125 10:52:13.571257 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/db92740e-3cd9-4b81-b153-3f866fd3db62-ovsdbserver-sb\") pod \"dnsmasq-dns-cd5cbd7b9-xd4pf\" (UID: \"db92740e-3cd9-4b81-b153-3f866fd3db62\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-xd4pf" Nov 25 10:52:13 crc kubenswrapper[4821]: I1125 10:52:13.609353 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42sws\" (UniqueName: \"kubernetes.io/projected/db92740e-3cd9-4b81-b153-3f866fd3db62-kube-api-access-42sws\") pod \"dnsmasq-dns-cd5cbd7b9-xd4pf\" (UID: \"db92740e-3cd9-4b81-b153-3f866fd3db62\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-xd4pf" Nov 25 10:52:13 crc kubenswrapper[4821]: I1125 10:52:13.834101 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-xd4pf" Nov 25 10:52:14 crc kubenswrapper[4821]: I1125 10:52:14.195342 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:52:14 crc kubenswrapper[4821]: I1125 10:52:14.372065 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-xd4pf"] Nov 25 10:52:14 crc kubenswrapper[4821]: W1125 10:52:14.373814 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddb92740e_3cd9_4b81_b153_3f866fd3db62.slice/crio-57c57fb91904c23186e70ab11e5d85b3e5886f55850beade32a56686f77fbd0c WatchSource:0}: Error finding container 57c57fb91904c23186e70ab11e5d85b3e5886f55850beade32a56686f77fbd0c: Status 404 returned error can't find the container with id 57c57fb91904c23186e70ab11e5d85b3e5886f55850beade32a56686f77fbd0c Nov 25 10:52:14 crc kubenswrapper[4821]: I1125 10:52:14.381063 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tvnnv\" (UniqueName: \"kubernetes.io/projected/98e78fd5-b460-436b-990a-fdf91c8d95c9-kube-api-access-tvnnv\") pod \"98e78fd5-b460-436b-990a-fdf91c8d95c9\" (UID: \"98e78fd5-b460-436b-990a-fdf91c8d95c9\") " Nov 25 10:52:14 crc kubenswrapper[4821]: I1125 10:52:14.381121 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98e78fd5-b460-436b-990a-fdf91c8d95c9-combined-ca-bundle\") pod \"98e78fd5-b460-436b-990a-fdf91c8d95c9\" (UID: \"98e78fd5-b460-436b-990a-fdf91c8d95c9\") " Nov 25 10:52:14 crc kubenswrapper[4821]: I1125 10:52:14.381178 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98e78fd5-b460-436b-990a-fdf91c8d95c9-config-data\") pod \"98e78fd5-b460-436b-990a-fdf91c8d95c9\" (UID: \"98e78fd5-b460-436b-990a-fdf91c8d95c9\") " Nov 25 10:52:14 crc kubenswrapper[4821]: I1125 10:52:14.409517 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98e78fd5-b460-436b-990a-fdf91c8d95c9-kube-api-access-tvnnv" (OuterVolumeSpecName: "kube-api-access-tvnnv") pod "98e78fd5-b460-436b-990a-fdf91c8d95c9" (UID: "98e78fd5-b460-436b-990a-fdf91c8d95c9"). InnerVolumeSpecName "kube-api-access-tvnnv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:52:14 crc kubenswrapper[4821]: I1125 10:52:14.425531 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98e78fd5-b460-436b-990a-fdf91c8d95c9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "98e78fd5-b460-436b-990a-fdf91c8d95c9" (UID: "98e78fd5-b460-436b-990a-fdf91c8d95c9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:52:14 crc kubenswrapper[4821]: I1125 10:52:14.433243 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98e78fd5-b460-436b-990a-fdf91c8d95c9-config-data" (OuterVolumeSpecName: "config-data") pod "98e78fd5-b460-436b-990a-fdf91c8d95c9" (UID: "98e78fd5-b460-436b-990a-fdf91c8d95c9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:52:14 crc kubenswrapper[4821]: I1125 10:52:14.483189 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tvnnv\" (UniqueName: \"kubernetes.io/projected/98e78fd5-b460-436b-990a-fdf91c8d95c9-kube-api-access-tvnnv\") on node \"crc\" DevicePath \"\"" Nov 25 10:52:14 crc kubenswrapper[4821]: I1125 10:52:14.483230 4821 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98e78fd5-b460-436b-990a-fdf91c8d95c9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:52:14 crc kubenswrapper[4821]: I1125 10:52:14.483239 4821 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98e78fd5-b460-436b-990a-fdf91c8d95c9-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:52:14 crc kubenswrapper[4821]: I1125 10:52:14.713529 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-xd4pf" event={"ID":"db92740e-3cd9-4b81-b153-3f866fd3db62","Type":"ContainerStarted","Data":"57c57fb91904c23186e70ab11e5d85b3e5886f55850beade32a56686f77fbd0c"} Nov 25 10:52:14 crc kubenswrapper[4821]: I1125 10:52:14.728077 4821 generic.go:334] "Generic (PLEG): container finished" podID="b044e354-e28e-43c0-ba43-bb51e8a21761" containerID="9dc278880a147a8959c8dae238461baf2c4e42d54c0ed542c1d08aa498941b9a" exitCode=137 Nov 25 10:52:14 crc kubenswrapper[4821]: I1125 10:52:14.728267 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b044e354-e28e-43c0-ba43-bb51e8a21761","Type":"ContainerDied","Data":"9dc278880a147a8959c8dae238461baf2c4e42d54c0ed542c1d08aa498941b9a"} Nov 25 10:52:14 crc kubenswrapper[4821]: I1125 10:52:14.728300 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b044e354-e28e-43c0-ba43-bb51e8a21761","Type":"ContainerDied","Data":"526a6ad2288f544d51b53c64077bbf3622a1a7297731400ef71216736743a1cd"} Nov 25 10:52:14 crc kubenswrapper[4821]: I1125 10:52:14.728314 4821 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="526a6ad2288f544d51b53c64077bbf3622a1a7297731400ef71216736743a1cd" Nov 25 10:52:14 crc kubenswrapper[4821]: I1125 10:52:14.734009 4821 generic.go:334] "Generic (PLEG): container finished" podID="98e78fd5-b460-436b-990a-fdf91c8d95c9" containerID="a1f49d02914e3f07788cfd9a71fcc19a3a39bc6a75bc5c0637a3e999d698bac2" exitCode=137 Nov 25 10:52:14 crc kubenswrapper[4821]: I1125 10:52:14.734112 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:52:14 crc kubenswrapper[4821]: I1125 10:52:14.734187 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"98e78fd5-b460-436b-990a-fdf91c8d95c9","Type":"ContainerDied","Data":"a1f49d02914e3f07788cfd9a71fcc19a3a39bc6a75bc5c0637a3e999d698bac2"} Nov 25 10:52:14 crc kubenswrapper[4821]: I1125 10:52:14.734233 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"98e78fd5-b460-436b-990a-fdf91c8d95c9","Type":"ContainerDied","Data":"f41e01f314cebee4e04c54d4906c22432d231b7630752977b6aa0bf358721afa"} Nov 25 10:52:14 crc kubenswrapper[4821]: I1125 10:52:14.734257 4821 scope.go:117] "RemoveContainer" containerID="a1f49d02914e3f07788cfd9a71fcc19a3a39bc6a75bc5c0637a3e999d698bac2" Nov 25 10:52:14 crc kubenswrapper[4821]: I1125 10:52:14.814550 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 10:52:14 crc kubenswrapper[4821]: I1125 10:52:14.830033 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 25 10:52:14 crc kubenswrapper[4821]: I1125 10:52:14.840839 4821 scope.go:117] "RemoveContainer" containerID="a1f49d02914e3f07788cfd9a71fcc19a3a39bc6a75bc5c0637a3e999d698bac2" Nov 25 10:52:14 crc kubenswrapper[4821]: I1125 10:52:14.841036 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 25 10:52:14 crc kubenswrapper[4821]: I1125 10:52:14.846323 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 25 10:52:14 crc kubenswrapper[4821]: E1125 10:52:14.847098 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b044e354-e28e-43c0-ba43-bb51e8a21761" containerName="nova-metadata-metadata" Nov 25 10:52:14 crc kubenswrapper[4821]: I1125 10:52:14.847115 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="b044e354-e28e-43c0-ba43-bb51e8a21761" containerName="nova-metadata-metadata" Nov 25 10:52:14 crc kubenswrapper[4821]: E1125 10:52:14.847138 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98e78fd5-b460-436b-990a-fdf91c8d95c9" containerName="nova-cell1-novncproxy-novncproxy" Nov 25 10:52:14 crc kubenswrapper[4821]: I1125 10:52:14.847147 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="98e78fd5-b460-436b-990a-fdf91c8d95c9" containerName="nova-cell1-novncproxy-novncproxy" Nov 25 10:52:14 crc kubenswrapper[4821]: E1125 10:52:14.847184 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b044e354-e28e-43c0-ba43-bb51e8a21761" containerName="nova-metadata-log" Nov 25 10:52:14 crc kubenswrapper[4821]: I1125 10:52:14.847196 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="b044e354-e28e-43c0-ba43-bb51e8a21761" containerName="nova-metadata-log" Nov 25 10:52:14 crc kubenswrapper[4821]: I1125 10:52:14.847382 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="98e78fd5-b460-436b-990a-fdf91c8d95c9" containerName="nova-cell1-novncproxy-novncproxy" Nov 25 10:52:14 crc kubenswrapper[4821]: I1125 10:52:14.847405 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="b044e354-e28e-43c0-ba43-bb51e8a21761" containerName="nova-metadata-log" Nov 25 10:52:14 crc kubenswrapper[4821]: I1125 10:52:14.847416 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="b044e354-e28e-43c0-ba43-bb51e8a21761" containerName="nova-metadata-metadata" Nov 25 10:52:14 crc kubenswrapper[4821]: I1125 10:52:14.848007 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:52:14 crc kubenswrapper[4821]: E1125 10:52:14.848710 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1f49d02914e3f07788cfd9a71fcc19a3a39bc6a75bc5c0637a3e999d698bac2\": container with ID starting with a1f49d02914e3f07788cfd9a71fcc19a3a39bc6a75bc5c0637a3e999d698bac2 not found: ID does not exist" containerID="a1f49d02914e3f07788cfd9a71fcc19a3a39bc6a75bc5c0637a3e999d698bac2" Nov 25 10:52:14 crc kubenswrapper[4821]: I1125 10:52:14.848774 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1f49d02914e3f07788cfd9a71fcc19a3a39bc6a75bc5c0637a3e999d698bac2"} err="failed to get container status \"a1f49d02914e3f07788cfd9a71fcc19a3a39bc6a75bc5c0637a3e999d698bac2\": rpc error: code = NotFound desc = could not find container \"a1f49d02914e3f07788cfd9a71fcc19a3a39bc6a75bc5c0637a3e999d698bac2\": container with ID starting with a1f49d02914e3f07788cfd9a71fcc19a3a39bc6a75bc5c0637a3e999d698bac2 not found: ID does not exist" Nov 25 10:52:14 crc kubenswrapper[4821]: I1125 10:52:14.850518 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Nov 25 10:52:14 crc kubenswrapper[4821]: I1125 10:52:14.850787 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Nov 25 10:52:14 crc kubenswrapper[4821]: I1125 10:52:14.850986 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 25 10:52:14 crc kubenswrapper[4821]: I1125 10:52:14.891737 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 25 10:52:14 crc kubenswrapper[4821]: I1125 10:52:14.995550 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-79qdb\" (UniqueName: \"kubernetes.io/projected/b044e354-e28e-43c0-ba43-bb51e8a21761-kube-api-access-79qdb\") pod \"b044e354-e28e-43c0-ba43-bb51e8a21761\" (UID: \"b044e354-e28e-43c0-ba43-bb51e8a21761\") " Nov 25 10:52:14 crc kubenswrapper[4821]: I1125 10:52:14.995607 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b044e354-e28e-43c0-ba43-bb51e8a21761-combined-ca-bundle\") pod \"b044e354-e28e-43c0-ba43-bb51e8a21761\" (UID: \"b044e354-e28e-43c0-ba43-bb51e8a21761\") " Nov 25 10:52:14 crc kubenswrapper[4821]: I1125 10:52:14.996208 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b044e354-e28e-43c0-ba43-bb51e8a21761-logs\") pod \"b044e354-e28e-43c0-ba43-bb51e8a21761\" (UID: \"b044e354-e28e-43c0-ba43-bb51e8a21761\") " Nov 25 10:52:14 crc kubenswrapper[4821]: I1125 10:52:14.996267 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b044e354-e28e-43c0-ba43-bb51e8a21761-config-data\") pod \"b044e354-e28e-43c0-ba43-bb51e8a21761\" (UID: \"b044e354-e28e-43c0-ba43-bb51e8a21761\") " Nov 25 10:52:14 crc kubenswrapper[4821]: I1125 10:52:14.996548 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/5b8402cb-db96-4ea4-8f95-a2ea84f202f5-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"5b8402cb-db96-4ea4-8f95-a2ea84f202f5\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:52:14 crc kubenswrapper[4821]: I1125 10:52:14.996553 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b044e354-e28e-43c0-ba43-bb51e8a21761-logs" (OuterVolumeSpecName: "logs") pod "b044e354-e28e-43c0-ba43-bb51e8a21761" (UID: "b044e354-e28e-43c0-ba43-bb51e8a21761"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:52:14 crc kubenswrapper[4821]: I1125 10:52:14.996579 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b8402cb-db96-4ea4-8f95-a2ea84f202f5-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"5b8402cb-db96-4ea4-8f95-a2ea84f202f5\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:52:14 crc kubenswrapper[4821]: I1125 10:52:14.996816 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/5b8402cb-db96-4ea4-8f95-a2ea84f202f5-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"5b8402cb-db96-4ea4-8f95-a2ea84f202f5\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:52:14 crc kubenswrapper[4821]: I1125 10:52:14.996992 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b8402cb-db96-4ea4-8f95-a2ea84f202f5-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"5b8402cb-db96-4ea4-8f95-a2ea84f202f5\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:52:14 crc kubenswrapper[4821]: I1125 10:52:14.997037 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7dvg\" (UniqueName: \"kubernetes.io/projected/5b8402cb-db96-4ea4-8f95-a2ea84f202f5-kube-api-access-n7dvg\") pod \"nova-cell1-novncproxy-0\" (UID: \"5b8402cb-db96-4ea4-8f95-a2ea84f202f5\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:52:14 crc kubenswrapper[4821]: I1125 10:52:14.997125 4821 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b044e354-e28e-43c0-ba43-bb51e8a21761-logs\") on node \"crc\" DevicePath \"\"" Nov 25 10:52:15 crc kubenswrapper[4821]: I1125 10:52:15.011586 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b044e354-e28e-43c0-ba43-bb51e8a21761-kube-api-access-79qdb" (OuterVolumeSpecName: "kube-api-access-79qdb") pod "b044e354-e28e-43c0-ba43-bb51e8a21761" (UID: "b044e354-e28e-43c0-ba43-bb51e8a21761"). InnerVolumeSpecName "kube-api-access-79qdb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:52:15 crc kubenswrapper[4821]: I1125 10:52:15.020477 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b044e354-e28e-43c0-ba43-bb51e8a21761-config-data" (OuterVolumeSpecName: "config-data") pod "b044e354-e28e-43c0-ba43-bb51e8a21761" (UID: "b044e354-e28e-43c0-ba43-bb51e8a21761"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:52:15 crc kubenswrapper[4821]: I1125 10:52:15.020922 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b044e354-e28e-43c0-ba43-bb51e8a21761-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b044e354-e28e-43c0-ba43-bb51e8a21761" (UID: "b044e354-e28e-43c0-ba43-bb51e8a21761"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:52:15 crc kubenswrapper[4821]: I1125 10:52:15.098502 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b8402cb-db96-4ea4-8f95-a2ea84f202f5-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"5b8402cb-db96-4ea4-8f95-a2ea84f202f5\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:52:15 crc kubenswrapper[4821]: I1125 10:52:15.098583 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7dvg\" (UniqueName: \"kubernetes.io/projected/5b8402cb-db96-4ea4-8f95-a2ea84f202f5-kube-api-access-n7dvg\") pod \"nova-cell1-novncproxy-0\" (UID: \"5b8402cb-db96-4ea4-8f95-a2ea84f202f5\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:52:15 crc kubenswrapper[4821]: I1125 10:52:15.098631 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/5b8402cb-db96-4ea4-8f95-a2ea84f202f5-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"5b8402cb-db96-4ea4-8f95-a2ea84f202f5\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:52:15 crc kubenswrapper[4821]: I1125 10:52:15.098664 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b8402cb-db96-4ea4-8f95-a2ea84f202f5-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"5b8402cb-db96-4ea4-8f95-a2ea84f202f5\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:52:15 crc kubenswrapper[4821]: I1125 10:52:15.098768 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/5b8402cb-db96-4ea4-8f95-a2ea84f202f5-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"5b8402cb-db96-4ea4-8f95-a2ea84f202f5\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:52:15 crc kubenswrapper[4821]: I1125 10:52:15.098872 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-79qdb\" (UniqueName: \"kubernetes.io/projected/b044e354-e28e-43c0-ba43-bb51e8a21761-kube-api-access-79qdb\") on node \"crc\" DevicePath \"\"" Nov 25 10:52:15 crc kubenswrapper[4821]: I1125 10:52:15.098886 4821 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b044e354-e28e-43c0-ba43-bb51e8a21761-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:52:15 crc kubenswrapper[4821]: I1125 10:52:15.098902 4821 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b044e354-e28e-43c0-ba43-bb51e8a21761-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:52:15 crc kubenswrapper[4821]: I1125 10:52:15.102513 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/5b8402cb-db96-4ea4-8f95-a2ea84f202f5-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"5b8402cb-db96-4ea4-8f95-a2ea84f202f5\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:52:15 crc kubenswrapper[4821]: I1125 10:52:15.102745 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b8402cb-db96-4ea4-8f95-a2ea84f202f5-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"5b8402cb-db96-4ea4-8f95-a2ea84f202f5\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:52:15 crc kubenswrapper[4821]: I1125 10:52:15.103959 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b8402cb-db96-4ea4-8f95-a2ea84f202f5-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"5b8402cb-db96-4ea4-8f95-a2ea84f202f5\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:52:15 crc kubenswrapper[4821]: I1125 10:52:15.104585 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/5b8402cb-db96-4ea4-8f95-a2ea84f202f5-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"5b8402cb-db96-4ea4-8f95-a2ea84f202f5\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:52:15 crc kubenswrapper[4821]: I1125 10:52:15.118393 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7dvg\" (UniqueName: \"kubernetes.io/projected/5b8402cb-db96-4ea4-8f95-a2ea84f202f5-kube-api-access-n7dvg\") pod \"nova-cell1-novncproxy-0\" (UID: \"5b8402cb-db96-4ea4-8f95-a2ea84f202f5\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:52:15 crc kubenswrapper[4821]: I1125 10:52:15.174693 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:52:15 crc kubenswrapper[4821]: I1125 10:52:15.599845 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 25 10:52:15 crc kubenswrapper[4821]: I1125 10:52:15.744046 4821 generic.go:334] "Generic (PLEG): container finished" podID="db92740e-3cd9-4b81-b153-3f866fd3db62" containerID="f7a503d2e982596c45acc041ea6221e887ebb186f6f3e48e63644ac930367f7d" exitCode=0 Nov 25 10:52:15 crc kubenswrapper[4821]: I1125 10:52:15.744134 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-xd4pf" event={"ID":"db92740e-3cd9-4b81-b153-3f866fd3db62","Type":"ContainerDied","Data":"f7a503d2e982596c45acc041ea6221e887ebb186f6f3e48e63644ac930367f7d"} Nov 25 10:52:15 crc kubenswrapper[4821]: I1125 10:52:15.748616 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 10:52:15 crc kubenswrapper[4821]: I1125 10:52:15.749120 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="31227d89-6706-407e-aa0a-e90b99611462" containerName="nova-api-log" containerID="cri-o://0f09f615de9732f25086805966d2a9e7cc0742d848f2c1da0f098ffdef5755a0" gracePeriod=30 Nov 25 10:52:15 crc kubenswrapper[4821]: I1125 10:52:15.749233 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="31227d89-6706-407e-aa0a-e90b99611462" containerName="nova-api-api" containerID="cri-o://6142554888b4dd483754e7df3beec07af51e4423dedfc53e264667a070450805" gracePeriod=30 Nov 25 10:52:15 crc kubenswrapper[4821]: W1125 10:52:15.806615 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5b8402cb_db96_4ea4_8f95_a2ea84f202f5.slice/crio-909a4ad595970cc28d2065ddcf2dcd2274b16867a3d7f0cd91e6a39898401aed WatchSource:0}: Error finding container 909a4ad595970cc28d2065ddcf2dcd2274b16867a3d7f0cd91e6a39898401aed: Status 404 returned error can't find the container with id 909a4ad595970cc28d2065ddcf2dcd2274b16867a3d7f0cd91e6a39898401aed Nov 25 10:52:15 crc kubenswrapper[4821]: I1125 10:52:15.807406 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 25 10:52:16 crc kubenswrapper[4821]: I1125 10:52:16.010001 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 10:52:16 crc kubenswrapper[4821]: I1125 10:52:16.024471 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 10:52:16 crc kubenswrapper[4821]: I1125 10:52:16.037865 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 25 10:52:16 crc kubenswrapper[4821]: I1125 10:52:16.039850 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 10:52:16 crc kubenswrapper[4821]: I1125 10:52:16.042054 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 25 10:52:16 crc kubenswrapper[4821]: I1125 10:52:16.042282 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 25 10:52:16 crc kubenswrapper[4821]: I1125 10:52:16.069533 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 10:52:16 crc kubenswrapper[4821]: I1125 10:52:16.121492 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbldn\" (UniqueName: \"kubernetes.io/projected/802c0cd6-5dd0-41d9-ae10-12cd23aa0467-kube-api-access-lbldn\") pod \"nova-metadata-0\" (UID: \"802c0cd6-5dd0-41d9-ae10-12cd23aa0467\") " pod="openstack/nova-metadata-0" Nov 25 10:52:16 crc kubenswrapper[4821]: I1125 10:52:16.121813 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/802c0cd6-5dd0-41d9-ae10-12cd23aa0467-config-data\") pod \"nova-metadata-0\" (UID: \"802c0cd6-5dd0-41d9-ae10-12cd23aa0467\") " pod="openstack/nova-metadata-0" Nov 25 10:52:16 crc kubenswrapper[4821]: I1125 10:52:16.121948 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/802c0cd6-5dd0-41d9-ae10-12cd23aa0467-logs\") pod \"nova-metadata-0\" (UID: \"802c0cd6-5dd0-41d9-ae10-12cd23aa0467\") " pod="openstack/nova-metadata-0" Nov 25 10:52:16 crc kubenswrapper[4821]: I1125 10:52:16.122055 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/802c0cd6-5dd0-41d9-ae10-12cd23aa0467-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"802c0cd6-5dd0-41d9-ae10-12cd23aa0467\") " pod="openstack/nova-metadata-0" Nov 25 10:52:16 crc kubenswrapper[4821]: I1125 10:52:16.122216 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/802c0cd6-5dd0-41d9-ae10-12cd23aa0467-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"802c0cd6-5dd0-41d9-ae10-12cd23aa0467\") " pod="openstack/nova-metadata-0" Nov 25 10:52:16 crc kubenswrapper[4821]: I1125 10:52:16.126870 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98e78fd5-b460-436b-990a-fdf91c8d95c9" path="/var/lib/kubelet/pods/98e78fd5-b460-436b-990a-fdf91c8d95c9/volumes" Nov 25 10:52:16 crc kubenswrapper[4821]: I1125 10:52:16.127454 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b044e354-e28e-43c0-ba43-bb51e8a21761" path="/var/lib/kubelet/pods/b044e354-e28e-43c0-ba43-bb51e8a21761/volumes" Nov 25 10:52:16 crc kubenswrapper[4821]: I1125 10:52:16.128012 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:52:16 crc kubenswrapper[4821]: I1125 10:52:16.128271 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a" containerName="ceilometer-central-agent" containerID="cri-o://4ab8fdfa62f13fbd1abb83257ca5cdeba846d33627a9e30d10610757e61b64d7" gracePeriod=30 Nov 25 10:52:16 crc kubenswrapper[4821]: I1125 10:52:16.128499 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a" containerName="proxy-httpd" containerID="cri-o://7d1fd9f582251094b332ecec2b092c8237aaa47bb73dc61f7e9258c72e486ecd" gracePeriod=30 Nov 25 10:52:16 crc kubenswrapper[4821]: I1125 10:52:16.128529 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a" containerName="ceilometer-notification-agent" containerID="cri-o://7e6f9aabb7e768eb0a92de8459ab86f56405e962404785b3975090f9a2bb2cb6" gracePeriod=30 Nov 25 10:52:16 crc kubenswrapper[4821]: I1125 10:52:16.128897 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a" containerName="sg-core" containerID="cri-o://082c0727decabeea221fa43abfb8945aa71d5bc81de092541d63970c64ce4ac9" gracePeriod=30 Nov 25 10:52:16 crc kubenswrapper[4821]: I1125 10:52:16.139043 4821 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.195:3000/\": EOF" Nov 25 10:52:16 crc kubenswrapper[4821]: I1125 10:52:16.223895 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbldn\" (UniqueName: \"kubernetes.io/projected/802c0cd6-5dd0-41d9-ae10-12cd23aa0467-kube-api-access-lbldn\") pod \"nova-metadata-0\" (UID: \"802c0cd6-5dd0-41d9-ae10-12cd23aa0467\") " pod="openstack/nova-metadata-0" Nov 25 10:52:16 crc kubenswrapper[4821]: I1125 10:52:16.223989 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/802c0cd6-5dd0-41d9-ae10-12cd23aa0467-config-data\") pod \"nova-metadata-0\" (UID: \"802c0cd6-5dd0-41d9-ae10-12cd23aa0467\") " pod="openstack/nova-metadata-0" Nov 25 10:52:16 crc kubenswrapper[4821]: I1125 10:52:16.224032 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/802c0cd6-5dd0-41d9-ae10-12cd23aa0467-logs\") pod \"nova-metadata-0\" (UID: \"802c0cd6-5dd0-41d9-ae10-12cd23aa0467\") " pod="openstack/nova-metadata-0" Nov 25 10:52:16 crc kubenswrapper[4821]: I1125 10:52:16.224072 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/802c0cd6-5dd0-41d9-ae10-12cd23aa0467-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"802c0cd6-5dd0-41d9-ae10-12cd23aa0467\") " pod="openstack/nova-metadata-0" Nov 25 10:52:16 crc kubenswrapper[4821]: I1125 10:52:16.224133 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/802c0cd6-5dd0-41d9-ae10-12cd23aa0467-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"802c0cd6-5dd0-41d9-ae10-12cd23aa0467\") " pod="openstack/nova-metadata-0" Nov 25 10:52:16 crc kubenswrapper[4821]: I1125 10:52:16.224873 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/802c0cd6-5dd0-41d9-ae10-12cd23aa0467-logs\") pod \"nova-metadata-0\" (UID: \"802c0cd6-5dd0-41d9-ae10-12cd23aa0467\") " pod="openstack/nova-metadata-0" Nov 25 10:52:16 crc kubenswrapper[4821]: I1125 10:52:16.232895 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/802c0cd6-5dd0-41d9-ae10-12cd23aa0467-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"802c0cd6-5dd0-41d9-ae10-12cd23aa0467\") " pod="openstack/nova-metadata-0" Nov 25 10:52:16 crc kubenswrapper[4821]: I1125 10:52:16.232992 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/802c0cd6-5dd0-41d9-ae10-12cd23aa0467-config-data\") pod \"nova-metadata-0\" (UID: \"802c0cd6-5dd0-41d9-ae10-12cd23aa0467\") " pod="openstack/nova-metadata-0" Nov 25 10:52:16 crc kubenswrapper[4821]: I1125 10:52:16.233631 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/802c0cd6-5dd0-41d9-ae10-12cd23aa0467-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"802c0cd6-5dd0-41d9-ae10-12cd23aa0467\") " pod="openstack/nova-metadata-0" Nov 25 10:52:16 crc kubenswrapper[4821]: I1125 10:52:16.242622 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbldn\" (UniqueName: \"kubernetes.io/projected/802c0cd6-5dd0-41d9-ae10-12cd23aa0467-kube-api-access-lbldn\") pod \"nova-metadata-0\" (UID: \"802c0cd6-5dd0-41d9-ae10-12cd23aa0467\") " pod="openstack/nova-metadata-0" Nov 25 10:52:16 crc kubenswrapper[4821]: I1125 10:52:16.363897 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 10:52:16 crc kubenswrapper[4821]: I1125 10:52:16.759450 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"5b8402cb-db96-4ea4-8f95-a2ea84f202f5","Type":"ContainerStarted","Data":"d8ecd13fe4312c073b8118060ddbb49f4804ccfa299645e3f937e6dd17a86abd"} Nov 25 10:52:16 crc kubenswrapper[4821]: I1125 10:52:16.759508 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"5b8402cb-db96-4ea4-8f95-a2ea84f202f5","Type":"ContainerStarted","Data":"909a4ad595970cc28d2065ddcf2dcd2274b16867a3d7f0cd91e6a39898401aed"} Nov 25 10:52:16 crc kubenswrapper[4821]: I1125 10:52:16.763965 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-xd4pf" event={"ID":"db92740e-3cd9-4b81-b153-3f866fd3db62","Type":"ContainerStarted","Data":"f1f4e312d2979b1204af23dde344c8ca9c9547dab34f7e7fdd9fb0184f5f582f"} Nov 25 10:52:16 crc kubenswrapper[4821]: I1125 10:52:16.764079 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-cd5cbd7b9-xd4pf" Nov 25 10:52:16 crc kubenswrapper[4821]: I1125 10:52:16.767809 4821 generic.go:334] "Generic (PLEG): container finished" podID="8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a" containerID="7d1fd9f582251094b332ecec2b092c8237aaa47bb73dc61f7e9258c72e486ecd" exitCode=0 Nov 25 10:52:16 crc kubenswrapper[4821]: I1125 10:52:16.767842 4821 generic.go:334] "Generic (PLEG): container finished" podID="8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a" containerID="082c0727decabeea221fa43abfb8945aa71d5bc81de092541d63970c64ce4ac9" exitCode=2 Nov 25 10:52:16 crc kubenswrapper[4821]: I1125 10:52:16.767849 4821 generic.go:334] "Generic (PLEG): container finished" podID="8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a" containerID="4ab8fdfa62f13fbd1abb83257ca5cdeba846d33627a9e30d10610757e61b64d7" exitCode=0 Nov 25 10:52:16 crc kubenswrapper[4821]: I1125 10:52:16.767853 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a","Type":"ContainerDied","Data":"7d1fd9f582251094b332ecec2b092c8237aaa47bb73dc61f7e9258c72e486ecd"} Nov 25 10:52:16 crc kubenswrapper[4821]: I1125 10:52:16.767912 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a","Type":"ContainerDied","Data":"082c0727decabeea221fa43abfb8945aa71d5bc81de092541d63970c64ce4ac9"} Nov 25 10:52:16 crc kubenswrapper[4821]: I1125 10:52:16.767928 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a","Type":"ContainerDied","Data":"4ab8fdfa62f13fbd1abb83257ca5cdeba846d33627a9e30d10610757e61b64d7"} Nov 25 10:52:16 crc kubenswrapper[4821]: I1125 10:52:16.770303 4821 generic.go:334] "Generic (PLEG): container finished" podID="31227d89-6706-407e-aa0a-e90b99611462" containerID="0f09f615de9732f25086805966d2a9e7cc0742d848f2c1da0f098ffdef5755a0" exitCode=143 Nov 25 10:52:16 crc kubenswrapper[4821]: I1125 10:52:16.770349 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"31227d89-6706-407e-aa0a-e90b99611462","Type":"ContainerDied","Data":"0f09f615de9732f25086805966d2a9e7cc0742d848f2c1da0f098ffdef5755a0"} Nov 25 10:52:16 crc kubenswrapper[4821]: I1125 10:52:16.793446 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.793421104 podStartE2EDuration="2.793421104s" podCreationTimestamp="2025-11-25 10:52:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:52:16.785392543 +0000 UTC m=+1207.321712400" watchObservedRunningTime="2025-11-25 10:52:16.793421104 +0000 UTC m=+1207.329740951" Nov 25 10:52:16 crc kubenswrapper[4821]: I1125 10:52:16.818024 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-cd5cbd7b9-xd4pf" podStartSLOduration=4.817998489 podStartE2EDuration="4.817998489s" podCreationTimestamp="2025-11-25 10:52:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:52:16.804139515 +0000 UTC m=+1207.340459362" watchObservedRunningTime="2025-11-25 10:52:16.817998489 +0000 UTC m=+1207.354318346" Nov 25 10:52:16 crc kubenswrapper[4821]: I1125 10:52:16.858624 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 10:52:17 crc kubenswrapper[4821]: I1125 10:52:17.781283 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"802c0cd6-5dd0-41d9-ae10-12cd23aa0467","Type":"ContainerStarted","Data":"9e2d63189a92dad7e806e26293da984c4e4eba7281abfb9f21111046c6cbe494"} Nov 25 10:52:17 crc kubenswrapper[4821]: I1125 10:52:17.781893 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"802c0cd6-5dd0-41d9-ae10-12cd23aa0467","Type":"ContainerStarted","Data":"078a3f7579d31eb42966404567599c09306bdc5804c0c47e780e7f31e77f7928"} Nov 25 10:52:17 crc kubenswrapper[4821]: I1125 10:52:17.781913 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"802c0cd6-5dd0-41d9-ae10-12cd23aa0467","Type":"ContainerStarted","Data":"ef3b6426ee31c32e17d75bc1dedc62c6896e26ab182a1e940d6bb8b42f77eb48"} Nov 25 10:52:17 crc kubenswrapper[4821]: I1125 10:52:17.806427 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=1.80640467 podStartE2EDuration="1.80640467s" podCreationTimestamp="2025-11-25 10:52:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:52:17.79767245 +0000 UTC m=+1208.333992297" watchObservedRunningTime="2025-11-25 10:52:17.80640467 +0000 UTC m=+1208.342724527" Nov 25 10:52:19 crc kubenswrapper[4821]: I1125 10:52:19.356083 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 10:52:19 crc kubenswrapper[4821]: I1125 10:52:19.489917 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31227d89-6706-407e-aa0a-e90b99611462-config-data\") pod \"31227d89-6706-407e-aa0a-e90b99611462\" (UID: \"31227d89-6706-407e-aa0a-e90b99611462\") " Nov 25 10:52:19 crc kubenswrapper[4821]: I1125 10:52:19.489991 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/31227d89-6706-407e-aa0a-e90b99611462-logs\") pod \"31227d89-6706-407e-aa0a-e90b99611462\" (UID: \"31227d89-6706-407e-aa0a-e90b99611462\") " Nov 25 10:52:19 crc kubenswrapper[4821]: I1125 10:52:19.490175 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q6vtc\" (UniqueName: \"kubernetes.io/projected/31227d89-6706-407e-aa0a-e90b99611462-kube-api-access-q6vtc\") pod \"31227d89-6706-407e-aa0a-e90b99611462\" (UID: \"31227d89-6706-407e-aa0a-e90b99611462\") " Nov 25 10:52:19 crc kubenswrapper[4821]: I1125 10:52:19.490273 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31227d89-6706-407e-aa0a-e90b99611462-combined-ca-bundle\") pod \"31227d89-6706-407e-aa0a-e90b99611462\" (UID: \"31227d89-6706-407e-aa0a-e90b99611462\") " Nov 25 10:52:19 crc kubenswrapper[4821]: I1125 10:52:19.491243 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/31227d89-6706-407e-aa0a-e90b99611462-logs" (OuterVolumeSpecName: "logs") pod "31227d89-6706-407e-aa0a-e90b99611462" (UID: "31227d89-6706-407e-aa0a-e90b99611462"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:52:19 crc kubenswrapper[4821]: I1125 10:52:19.502950 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31227d89-6706-407e-aa0a-e90b99611462-kube-api-access-q6vtc" (OuterVolumeSpecName: "kube-api-access-q6vtc") pod "31227d89-6706-407e-aa0a-e90b99611462" (UID: "31227d89-6706-407e-aa0a-e90b99611462"). InnerVolumeSpecName "kube-api-access-q6vtc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:52:19 crc kubenswrapper[4821]: I1125 10:52:19.524145 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31227d89-6706-407e-aa0a-e90b99611462-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "31227d89-6706-407e-aa0a-e90b99611462" (UID: "31227d89-6706-407e-aa0a-e90b99611462"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:52:19 crc kubenswrapper[4821]: I1125 10:52:19.524256 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31227d89-6706-407e-aa0a-e90b99611462-config-data" (OuterVolumeSpecName: "config-data") pod "31227d89-6706-407e-aa0a-e90b99611462" (UID: "31227d89-6706-407e-aa0a-e90b99611462"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:52:19 crc kubenswrapper[4821]: I1125 10:52:19.592756 4821 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31227d89-6706-407e-aa0a-e90b99611462-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:52:19 crc kubenswrapper[4821]: I1125 10:52:19.592791 4821 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31227d89-6706-407e-aa0a-e90b99611462-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:52:19 crc kubenswrapper[4821]: I1125 10:52:19.592799 4821 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/31227d89-6706-407e-aa0a-e90b99611462-logs\") on node \"crc\" DevicePath \"\"" Nov 25 10:52:19 crc kubenswrapper[4821]: I1125 10:52:19.592808 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q6vtc\" (UniqueName: \"kubernetes.io/projected/31227d89-6706-407e-aa0a-e90b99611462-kube-api-access-q6vtc\") on node \"crc\" DevicePath \"\"" Nov 25 10:52:19 crc kubenswrapper[4821]: I1125 10:52:19.801134 4821 generic.go:334] "Generic (PLEG): container finished" podID="8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a" containerID="7e6f9aabb7e768eb0a92de8459ab86f56405e962404785b3975090f9a2bb2cb6" exitCode=0 Nov 25 10:52:19 crc kubenswrapper[4821]: I1125 10:52:19.801198 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a","Type":"ContainerDied","Data":"7e6f9aabb7e768eb0a92de8459ab86f56405e962404785b3975090f9a2bb2cb6"} Nov 25 10:52:19 crc kubenswrapper[4821]: I1125 10:52:19.802783 4821 generic.go:334] "Generic (PLEG): container finished" podID="31227d89-6706-407e-aa0a-e90b99611462" containerID="6142554888b4dd483754e7df3beec07af51e4423dedfc53e264667a070450805" exitCode=0 Nov 25 10:52:19 crc kubenswrapper[4821]: I1125 10:52:19.802817 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"31227d89-6706-407e-aa0a-e90b99611462","Type":"ContainerDied","Data":"6142554888b4dd483754e7df3beec07af51e4423dedfc53e264667a070450805"} Nov 25 10:52:19 crc kubenswrapper[4821]: I1125 10:52:19.802849 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"31227d89-6706-407e-aa0a-e90b99611462","Type":"ContainerDied","Data":"526823676ee1f29b2682b8e13e0cb391cd014899aec4b55ef0d1c9a89f243a29"} Nov 25 10:52:19 crc kubenswrapper[4821]: I1125 10:52:19.802872 4821 scope.go:117] "RemoveContainer" containerID="6142554888b4dd483754e7df3beec07af51e4423dedfc53e264667a070450805" Nov 25 10:52:19 crc kubenswrapper[4821]: I1125 10:52:19.802874 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 10:52:19 crc kubenswrapper[4821]: I1125 10:52:19.837569 4821 scope.go:117] "RemoveContainer" containerID="0f09f615de9732f25086805966d2a9e7cc0742d848f2c1da0f098ffdef5755a0" Nov 25 10:52:19 crc kubenswrapper[4821]: I1125 10:52:19.847299 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 25 10:52:19 crc kubenswrapper[4821]: I1125 10:52:19.879341 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 25 10:52:19 crc kubenswrapper[4821]: I1125 10:52:19.895535 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 25 10:52:19 crc kubenswrapper[4821]: E1125 10:52:19.896082 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31227d89-6706-407e-aa0a-e90b99611462" containerName="nova-api-log" Nov 25 10:52:19 crc kubenswrapper[4821]: I1125 10:52:19.896107 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="31227d89-6706-407e-aa0a-e90b99611462" containerName="nova-api-log" Nov 25 10:52:19 crc kubenswrapper[4821]: E1125 10:52:19.896132 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31227d89-6706-407e-aa0a-e90b99611462" containerName="nova-api-api" Nov 25 10:52:19 crc kubenswrapper[4821]: I1125 10:52:19.896139 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="31227d89-6706-407e-aa0a-e90b99611462" containerName="nova-api-api" Nov 25 10:52:19 crc kubenswrapper[4821]: I1125 10:52:19.896462 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="31227d89-6706-407e-aa0a-e90b99611462" containerName="nova-api-api" Nov 25 10:52:19 crc kubenswrapper[4821]: I1125 10:52:19.896511 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="31227d89-6706-407e-aa0a-e90b99611462" containerName="nova-api-log" Nov 25 10:52:19 crc kubenswrapper[4821]: I1125 10:52:19.897708 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 10:52:19 crc kubenswrapper[4821]: I1125 10:52:19.902716 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 25 10:52:19 crc kubenswrapper[4821]: I1125 10:52:19.902980 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 25 10:52:19 crc kubenswrapper[4821]: I1125 10:52:19.904931 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 25 10:52:19 crc kubenswrapper[4821]: I1125 10:52:19.905550 4821 scope.go:117] "RemoveContainer" containerID="6142554888b4dd483754e7df3beec07af51e4423dedfc53e264667a070450805" Nov 25 10:52:19 crc kubenswrapper[4821]: E1125 10:52:19.907768 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6142554888b4dd483754e7df3beec07af51e4423dedfc53e264667a070450805\": container with ID starting with 6142554888b4dd483754e7df3beec07af51e4423dedfc53e264667a070450805 not found: ID does not exist" containerID="6142554888b4dd483754e7df3beec07af51e4423dedfc53e264667a070450805" Nov 25 10:52:19 crc kubenswrapper[4821]: I1125 10:52:19.907793 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6142554888b4dd483754e7df3beec07af51e4423dedfc53e264667a070450805"} err="failed to get container status \"6142554888b4dd483754e7df3beec07af51e4423dedfc53e264667a070450805\": rpc error: code = NotFound desc = could not find container \"6142554888b4dd483754e7df3beec07af51e4423dedfc53e264667a070450805\": container with ID starting with 6142554888b4dd483754e7df3beec07af51e4423dedfc53e264667a070450805 not found: ID does not exist" Nov 25 10:52:19 crc kubenswrapper[4821]: I1125 10:52:19.907815 4821 scope.go:117] "RemoveContainer" containerID="0f09f615de9732f25086805966d2a9e7cc0742d848f2c1da0f098ffdef5755a0" Nov 25 10:52:19 crc kubenswrapper[4821]: E1125 10:52:19.916242 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f09f615de9732f25086805966d2a9e7cc0742d848f2c1da0f098ffdef5755a0\": container with ID starting with 0f09f615de9732f25086805966d2a9e7cc0742d848f2c1da0f098ffdef5755a0 not found: ID does not exist" containerID="0f09f615de9732f25086805966d2a9e7cc0742d848f2c1da0f098ffdef5755a0" Nov 25 10:52:19 crc kubenswrapper[4821]: I1125 10:52:19.916283 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f09f615de9732f25086805966d2a9e7cc0742d848f2c1da0f098ffdef5755a0"} err="failed to get container status \"0f09f615de9732f25086805966d2a9e7cc0742d848f2c1da0f098ffdef5755a0\": rpc error: code = NotFound desc = could not find container \"0f09f615de9732f25086805966d2a9e7cc0742d848f2c1da0f098ffdef5755a0\": container with ID starting with 0f09f615de9732f25086805966d2a9e7cc0742d848f2c1da0f098ffdef5755a0 not found: ID does not exist" Nov 25 10:52:19 crc kubenswrapper[4821]: I1125 10:52:19.923094 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 25 10:52:20 crc kubenswrapper[4821]: I1125 10:52:20.003910 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e305fedf-6c46-4c0e-9f05-ce67a8b2668c-public-tls-certs\") pod \"nova-api-0\" (UID: \"e305fedf-6c46-4c0e-9f05-ce67a8b2668c\") " pod="openstack/nova-api-0" Nov 25 10:52:20 crc kubenswrapper[4821]: I1125 10:52:20.004343 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xgbbz\" (UniqueName: \"kubernetes.io/projected/e305fedf-6c46-4c0e-9f05-ce67a8b2668c-kube-api-access-xgbbz\") pod \"nova-api-0\" (UID: \"e305fedf-6c46-4c0e-9f05-ce67a8b2668c\") " pod="openstack/nova-api-0" Nov 25 10:52:20 crc kubenswrapper[4821]: I1125 10:52:20.004396 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e305fedf-6c46-4c0e-9f05-ce67a8b2668c-logs\") pod \"nova-api-0\" (UID: \"e305fedf-6c46-4c0e-9f05-ce67a8b2668c\") " pod="openstack/nova-api-0" Nov 25 10:52:20 crc kubenswrapper[4821]: I1125 10:52:20.004487 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e305fedf-6c46-4c0e-9f05-ce67a8b2668c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e305fedf-6c46-4c0e-9f05-ce67a8b2668c\") " pod="openstack/nova-api-0" Nov 25 10:52:20 crc kubenswrapper[4821]: I1125 10:52:20.004508 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e305fedf-6c46-4c0e-9f05-ce67a8b2668c-internal-tls-certs\") pod \"nova-api-0\" (UID: \"e305fedf-6c46-4c0e-9f05-ce67a8b2668c\") " pod="openstack/nova-api-0" Nov 25 10:52:20 crc kubenswrapper[4821]: I1125 10:52:20.004655 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e305fedf-6c46-4c0e-9f05-ce67a8b2668c-config-data\") pod \"nova-api-0\" (UID: \"e305fedf-6c46-4c0e-9f05-ce67a8b2668c\") " pod="openstack/nova-api-0" Nov 25 10:52:20 crc kubenswrapper[4821]: I1125 10:52:20.105809 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e305fedf-6c46-4c0e-9f05-ce67a8b2668c-config-data\") pod \"nova-api-0\" (UID: \"e305fedf-6c46-4c0e-9f05-ce67a8b2668c\") " pod="openstack/nova-api-0" Nov 25 10:52:20 crc kubenswrapper[4821]: I1125 10:52:20.105860 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e305fedf-6c46-4c0e-9f05-ce67a8b2668c-public-tls-certs\") pod \"nova-api-0\" (UID: \"e305fedf-6c46-4c0e-9f05-ce67a8b2668c\") " pod="openstack/nova-api-0" Nov 25 10:52:20 crc kubenswrapper[4821]: I1125 10:52:20.105901 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xgbbz\" (UniqueName: \"kubernetes.io/projected/e305fedf-6c46-4c0e-9f05-ce67a8b2668c-kube-api-access-xgbbz\") pod \"nova-api-0\" (UID: \"e305fedf-6c46-4c0e-9f05-ce67a8b2668c\") " pod="openstack/nova-api-0" Nov 25 10:52:20 crc kubenswrapper[4821]: I1125 10:52:20.105934 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e305fedf-6c46-4c0e-9f05-ce67a8b2668c-logs\") pod \"nova-api-0\" (UID: \"e305fedf-6c46-4c0e-9f05-ce67a8b2668c\") " pod="openstack/nova-api-0" Nov 25 10:52:20 crc kubenswrapper[4821]: I1125 10:52:20.105993 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e305fedf-6c46-4c0e-9f05-ce67a8b2668c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e305fedf-6c46-4c0e-9f05-ce67a8b2668c\") " pod="openstack/nova-api-0" Nov 25 10:52:20 crc kubenswrapper[4821]: I1125 10:52:20.106006 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e305fedf-6c46-4c0e-9f05-ce67a8b2668c-internal-tls-certs\") pod \"nova-api-0\" (UID: \"e305fedf-6c46-4c0e-9f05-ce67a8b2668c\") " pod="openstack/nova-api-0" Nov 25 10:52:20 crc kubenswrapper[4821]: I1125 10:52:20.106845 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e305fedf-6c46-4c0e-9f05-ce67a8b2668c-logs\") pod \"nova-api-0\" (UID: \"e305fedf-6c46-4c0e-9f05-ce67a8b2668c\") " pod="openstack/nova-api-0" Nov 25 10:52:20 crc kubenswrapper[4821]: I1125 10:52:20.111804 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e305fedf-6c46-4c0e-9f05-ce67a8b2668c-public-tls-certs\") pod \"nova-api-0\" (UID: \"e305fedf-6c46-4c0e-9f05-ce67a8b2668c\") " pod="openstack/nova-api-0" Nov 25 10:52:20 crc kubenswrapper[4821]: I1125 10:52:20.111890 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e305fedf-6c46-4c0e-9f05-ce67a8b2668c-internal-tls-certs\") pod \"nova-api-0\" (UID: \"e305fedf-6c46-4c0e-9f05-ce67a8b2668c\") " pod="openstack/nova-api-0" Nov 25 10:52:20 crc kubenswrapper[4821]: I1125 10:52:20.112266 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e305fedf-6c46-4c0e-9f05-ce67a8b2668c-config-data\") pod \"nova-api-0\" (UID: \"e305fedf-6c46-4c0e-9f05-ce67a8b2668c\") " pod="openstack/nova-api-0" Nov 25 10:52:20 crc kubenswrapper[4821]: I1125 10:52:20.112469 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e305fedf-6c46-4c0e-9f05-ce67a8b2668c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e305fedf-6c46-4c0e-9f05-ce67a8b2668c\") " pod="openstack/nova-api-0" Nov 25 10:52:20 crc kubenswrapper[4821]: I1125 10:52:20.123808 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xgbbz\" (UniqueName: \"kubernetes.io/projected/e305fedf-6c46-4c0e-9f05-ce67a8b2668c-kube-api-access-xgbbz\") pod \"nova-api-0\" (UID: \"e305fedf-6c46-4c0e-9f05-ce67a8b2668c\") " pod="openstack/nova-api-0" Nov 25 10:52:20 crc kubenswrapper[4821]: I1125 10:52:20.125646 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31227d89-6706-407e-aa0a-e90b99611462" path="/var/lib/kubelet/pods/31227d89-6706-407e-aa0a-e90b99611462/volumes" Nov 25 10:52:20 crc kubenswrapper[4821]: I1125 10:52:20.175497 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:52:20 crc kubenswrapper[4821]: I1125 10:52:20.213399 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:52:20 crc kubenswrapper[4821]: I1125 10:52:20.222012 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 10:52:20 crc kubenswrapper[4821]: I1125 10:52:20.310048 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a-run-httpd\") pod \"8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a\" (UID: \"8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a\") " Nov 25 10:52:20 crc kubenswrapper[4821]: I1125 10:52:20.310099 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a-sg-core-conf-yaml\") pod \"8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a\" (UID: \"8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a\") " Nov 25 10:52:20 crc kubenswrapper[4821]: I1125 10:52:20.310137 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a-combined-ca-bundle\") pod \"8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a\" (UID: \"8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a\") " Nov 25 10:52:20 crc kubenswrapper[4821]: I1125 10:52:20.310245 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a-config-data\") pod \"8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a\" (UID: \"8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a\") " Nov 25 10:52:20 crc kubenswrapper[4821]: I1125 10:52:20.310307 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a-ceilometer-tls-certs\") pod \"8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a\" (UID: \"8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a\") " Nov 25 10:52:20 crc kubenswrapper[4821]: I1125 10:52:20.310590 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a" (UID: "8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:52:20 crc kubenswrapper[4821]: I1125 10:52:20.310693 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a-scripts\") pod \"8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a\" (UID: \"8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a\") " Nov 25 10:52:20 crc kubenswrapper[4821]: I1125 10:52:20.310714 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-scssg\" (UniqueName: \"kubernetes.io/projected/8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a-kube-api-access-scssg\") pod \"8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a\" (UID: \"8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a\") " Nov 25 10:52:20 crc kubenswrapper[4821]: I1125 10:52:20.310759 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a-log-httpd\") pod \"8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a\" (UID: \"8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a\") " Nov 25 10:52:20 crc kubenswrapper[4821]: I1125 10:52:20.312397 4821 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 10:52:20 crc kubenswrapper[4821]: I1125 10:52:20.313497 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a" (UID: "8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:52:20 crc kubenswrapper[4821]: I1125 10:52:20.327480 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a-scripts" (OuterVolumeSpecName: "scripts") pod "8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a" (UID: "8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:52:20 crc kubenswrapper[4821]: I1125 10:52:20.327529 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a-kube-api-access-scssg" (OuterVolumeSpecName: "kube-api-access-scssg") pod "8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a" (UID: "8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a"). InnerVolumeSpecName "kube-api-access-scssg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:52:20 crc kubenswrapper[4821]: I1125 10:52:20.347842 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a" (UID: "8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:52:20 crc kubenswrapper[4821]: I1125 10:52:20.418638 4821 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:52:20 crc kubenswrapper[4821]: I1125 10:52:20.418683 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-scssg\" (UniqueName: \"kubernetes.io/projected/8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a-kube-api-access-scssg\") on node \"crc\" DevicePath \"\"" Nov 25 10:52:20 crc kubenswrapper[4821]: I1125 10:52:20.418697 4821 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 10:52:20 crc kubenswrapper[4821]: I1125 10:52:20.418708 4821 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 25 10:52:20 crc kubenswrapper[4821]: I1125 10:52:20.421648 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a" (UID: "8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:52:20 crc kubenswrapper[4821]: I1125 10:52:20.450011 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a" (UID: "8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:52:20 crc kubenswrapper[4821]: I1125 10:52:20.507365 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a-config-data" (OuterVolumeSpecName: "config-data") pod "8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a" (UID: "8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:52:20 crc kubenswrapper[4821]: I1125 10:52:20.520931 4821 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:52:20 crc kubenswrapper[4821]: I1125 10:52:20.520969 4821 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:52:20 crc kubenswrapper[4821]: I1125 10:52:20.520977 4821 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 10:52:20 crc kubenswrapper[4821]: I1125 10:52:20.735301 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 25 10:52:20 crc kubenswrapper[4821]: W1125 10:52:20.738376 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode305fedf_6c46_4c0e_9f05_ce67a8b2668c.slice/crio-648a4a3e2d0645f1953bda347d075246f296aaef440150dcdf92f8db43a974db WatchSource:0}: Error finding container 648a4a3e2d0645f1953bda347d075246f296aaef440150dcdf92f8db43a974db: Status 404 returned error can't find the container with id 648a4a3e2d0645f1953bda347d075246f296aaef440150dcdf92f8db43a974db Nov 25 10:52:20 crc kubenswrapper[4821]: I1125 10:52:20.816321 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e305fedf-6c46-4c0e-9f05-ce67a8b2668c","Type":"ContainerStarted","Data":"648a4a3e2d0645f1953bda347d075246f296aaef440150dcdf92f8db43a974db"} Nov 25 10:52:20 crc kubenswrapper[4821]: I1125 10:52:20.819059 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a","Type":"ContainerDied","Data":"70e100b8fb0bfaf3ab5658cb324558d472ffab7e5bb4793012ad97e235a36311"} Nov 25 10:52:20 crc kubenswrapper[4821]: I1125 10:52:20.819118 4821 scope.go:117] "RemoveContainer" containerID="7d1fd9f582251094b332ecec2b092c8237aaa47bb73dc61f7e9258c72e486ecd" Nov 25 10:52:20 crc kubenswrapper[4821]: I1125 10:52:20.819272 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:52:20 crc kubenswrapper[4821]: I1125 10:52:20.896424 4821 scope.go:117] "RemoveContainer" containerID="082c0727decabeea221fa43abfb8945aa71d5bc81de092541d63970c64ce4ac9" Nov 25 10:52:20 crc kubenswrapper[4821]: I1125 10:52:20.907358 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:52:20 crc kubenswrapper[4821]: I1125 10:52:20.923547 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:52:20 crc kubenswrapper[4821]: I1125 10:52:20.937421 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:52:20 crc kubenswrapper[4821]: E1125 10:52:20.947071 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a" containerName="proxy-httpd" Nov 25 10:52:20 crc kubenswrapper[4821]: I1125 10:52:20.947186 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a" containerName="proxy-httpd" Nov 25 10:52:20 crc kubenswrapper[4821]: E1125 10:52:20.947224 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a" containerName="ceilometer-central-agent" Nov 25 10:52:20 crc kubenswrapper[4821]: I1125 10:52:20.947235 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a" containerName="ceilometer-central-agent" Nov 25 10:52:20 crc kubenswrapper[4821]: E1125 10:52:20.947250 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a" containerName="ceilometer-notification-agent" Nov 25 10:52:20 crc kubenswrapper[4821]: I1125 10:52:20.947259 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a" containerName="ceilometer-notification-agent" Nov 25 10:52:20 crc kubenswrapper[4821]: E1125 10:52:20.947276 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a" containerName="sg-core" Nov 25 10:52:20 crc kubenswrapper[4821]: I1125 10:52:20.947283 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a" containerName="sg-core" Nov 25 10:52:20 crc kubenswrapper[4821]: I1125 10:52:20.947709 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a" containerName="sg-core" Nov 25 10:52:20 crc kubenswrapper[4821]: I1125 10:52:20.947754 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a" containerName="ceilometer-central-agent" Nov 25 10:52:20 crc kubenswrapper[4821]: I1125 10:52:20.947773 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a" containerName="ceilometer-notification-agent" Nov 25 10:52:20 crc kubenswrapper[4821]: I1125 10:52:20.947785 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a" containerName="proxy-httpd" Nov 25 10:52:20 crc kubenswrapper[4821]: I1125 10:52:20.959074 4821 scope.go:117] "RemoveContainer" containerID="7e6f9aabb7e768eb0a92de8459ab86f56405e962404785b3975090f9a2bb2cb6" Nov 25 10:52:20 crc kubenswrapper[4821]: I1125 10:52:20.962181 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:52:20 crc kubenswrapper[4821]: I1125 10:52:20.968122 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 25 10:52:20 crc kubenswrapper[4821]: I1125 10:52:20.968353 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 25 10:52:20 crc kubenswrapper[4821]: I1125 10:52:20.968596 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 25 10:52:20 crc kubenswrapper[4821]: I1125 10:52:20.974531 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:52:21 crc kubenswrapper[4821]: I1125 10:52:21.137235 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxpb2\" (UniqueName: \"kubernetes.io/projected/08ee9d9a-c488-4289-9d8f-a87d338992d6-kube-api-access-jxpb2\") pod \"ceilometer-0\" (UID: \"08ee9d9a-c488-4289-9d8f-a87d338992d6\") " pod="openstack/ceilometer-0" Nov 25 10:52:21 crc kubenswrapper[4821]: I1125 10:52:21.137297 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/08ee9d9a-c488-4289-9d8f-a87d338992d6-scripts\") pod \"ceilometer-0\" (UID: \"08ee9d9a-c488-4289-9d8f-a87d338992d6\") " pod="openstack/ceilometer-0" Nov 25 10:52:21 crc kubenswrapper[4821]: I1125 10:52:21.137443 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/08ee9d9a-c488-4289-9d8f-a87d338992d6-log-httpd\") pod \"ceilometer-0\" (UID: \"08ee9d9a-c488-4289-9d8f-a87d338992d6\") " pod="openstack/ceilometer-0" Nov 25 10:52:21 crc kubenswrapper[4821]: I1125 10:52:21.137528 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08ee9d9a-c488-4289-9d8f-a87d338992d6-config-data\") pod \"ceilometer-0\" (UID: \"08ee9d9a-c488-4289-9d8f-a87d338992d6\") " pod="openstack/ceilometer-0" Nov 25 10:52:21 crc kubenswrapper[4821]: I1125 10:52:21.137576 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08ee9d9a-c488-4289-9d8f-a87d338992d6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"08ee9d9a-c488-4289-9d8f-a87d338992d6\") " pod="openstack/ceilometer-0" Nov 25 10:52:21 crc kubenswrapper[4821]: I1125 10:52:21.137655 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/08ee9d9a-c488-4289-9d8f-a87d338992d6-run-httpd\") pod \"ceilometer-0\" (UID: \"08ee9d9a-c488-4289-9d8f-a87d338992d6\") " pod="openstack/ceilometer-0" Nov 25 10:52:21 crc kubenswrapper[4821]: I1125 10:52:21.137680 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/08ee9d9a-c488-4289-9d8f-a87d338992d6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"08ee9d9a-c488-4289-9d8f-a87d338992d6\") " pod="openstack/ceilometer-0" Nov 25 10:52:21 crc kubenswrapper[4821]: I1125 10:52:21.137735 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/08ee9d9a-c488-4289-9d8f-a87d338992d6-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"08ee9d9a-c488-4289-9d8f-a87d338992d6\") " pod="openstack/ceilometer-0" Nov 25 10:52:21 crc kubenswrapper[4821]: I1125 10:52:21.189631 4821 scope.go:117] "RemoveContainer" containerID="4ab8fdfa62f13fbd1abb83257ca5cdeba846d33627a9e30d10610757e61b64d7" Nov 25 10:52:21 crc kubenswrapper[4821]: I1125 10:52:21.239146 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxpb2\" (UniqueName: \"kubernetes.io/projected/08ee9d9a-c488-4289-9d8f-a87d338992d6-kube-api-access-jxpb2\") pod \"ceilometer-0\" (UID: \"08ee9d9a-c488-4289-9d8f-a87d338992d6\") " pod="openstack/ceilometer-0" Nov 25 10:52:21 crc kubenswrapper[4821]: I1125 10:52:21.239298 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/08ee9d9a-c488-4289-9d8f-a87d338992d6-scripts\") pod \"ceilometer-0\" (UID: \"08ee9d9a-c488-4289-9d8f-a87d338992d6\") " pod="openstack/ceilometer-0" Nov 25 10:52:21 crc kubenswrapper[4821]: I1125 10:52:21.239442 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/08ee9d9a-c488-4289-9d8f-a87d338992d6-log-httpd\") pod \"ceilometer-0\" (UID: \"08ee9d9a-c488-4289-9d8f-a87d338992d6\") " pod="openstack/ceilometer-0" Nov 25 10:52:21 crc kubenswrapper[4821]: I1125 10:52:21.239471 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08ee9d9a-c488-4289-9d8f-a87d338992d6-config-data\") pod \"ceilometer-0\" (UID: \"08ee9d9a-c488-4289-9d8f-a87d338992d6\") " pod="openstack/ceilometer-0" Nov 25 10:52:21 crc kubenswrapper[4821]: I1125 10:52:21.239590 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08ee9d9a-c488-4289-9d8f-a87d338992d6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"08ee9d9a-c488-4289-9d8f-a87d338992d6\") " pod="openstack/ceilometer-0" Nov 25 10:52:21 crc kubenswrapper[4821]: I1125 10:52:21.239628 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/08ee9d9a-c488-4289-9d8f-a87d338992d6-run-httpd\") pod \"ceilometer-0\" (UID: \"08ee9d9a-c488-4289-9d8f-a87d338992d6\") " pod="openstack/ceilometer-0" Nov 25 10:52:21 crc kubenswrapper[4821]: I1125 10:52:21.239783 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/08ee9d9a-c488-4289-9d8f-a87d338992d6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"08ee9d9a-c488-4289-9d8f-a87d338992d6\") " pod="openstack/ceilometer-0" Nov 25 10:52:21 crc kubenswrapper[4821]: I1125 10:52:21.239812 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/08ee9d9a-c488-4289-9d8f-a87d338992d6-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"08ee9d9a-c488-4289-9d8f-a87d338992d6\") " pod="openstack/ceilometer-0" Nov 25 10:52:21 crc kubenswrapper[4821]: I1125 10:52:21.240656 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/08ee9d9a-c488-4289-9d8f-a87d338992d6-run-httpd\") pod \"ceilometer-0\" (UID: \"08ee9d9a-c488-4289-9d8f-a87d338992d6\") " pod="openstack/ceilometer-0" Nov 25 10:52:21 crc kubenswrapper[4821]: I1125 10:52:21.241033 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/08ee9d9a-c488-4289-9d8f-a87d338992d6-log-httpd\") pod \"ceilometer-0\" (UID: \"08ee9d9a-c488-4289-9d8f-a87d338992d6\") " pod="openstack/ceilometer-0" Nov 25 10:52:21 crc kubenswrapper[4821]: I1125 10:52:21.245249 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/08ee9d9a-c488-4289-9d8f-a87d338992d6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"08ee9d9a-c488-4289-9d8f-a87d338992d6\") " pod="openstack/ceilometer-0" Nov 25 10:52:21 crc kubenswrapper[4821]: I1125 10:52:21.245558 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/08ee9d9a-c488-4289-9d8f-a87d338992d6-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"08ee9d9a-c488-4289-9d8f-a87d338992d6\") " pod="openstack/ceilometer-0" Nov 25 10:52:21 crc kubenswrapper[4821]: I1125 10:52:21.245964 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08ee9d9a-c488-4289-9d8f-a87d338992d6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"08ee9d9a-c488-4289-9d8f-a87d338992d6\") " pod="openstack/ceilometer-0" Nov 25 10:52:21 crc kubenswrapper[4821]: I1125 10:52:21.245994 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08ee9d9a-c488-4289-9d8f-a87d338992d6-config-data\") pod \"ceilometer-0\" (UID: \"08ee9d9a-c488-4289-9d8f-a87d338992d6\") " pod="openstack/ceilometer-0" Nov 25 10:52:21 crc kubenswrapper[4821]: I1125 10:52:21.246886 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/08ee9d9a-c488-4289-9d8f-a87d338992d6-scripts\") pod \"ceilometer-0\" (UID: \"08ee9d9a-c488-4289-9d8f-a87d338992d6\") " pod="openstack/ceilometer-0" Nov 25 10:52:21 crc kubenswrapper[4821]: I1125 10:52:21.257822 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxpb2\" (UniqueName: \"kubernetes.io/projected/08ee9d9a-c488-4289-9d8f-a87d338992d6-kube-api-access-jxpb2\") pod \"ceilometer-0\" (UID: \"08ee9d9a-c488-4289-9d8f-a87d338992d6\") " pod="openstack/ceilometer-0" Nov 25 10:52:21 crc kubenswrapper[4821]: I1125 10:52:21.365755 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 25 10:52:21 crc kubenswrapper[4821]: I1125 10:52:21.366248 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 25 10:52:21 crc kubenswrapper[4821]: I1125 10:52:21.491433 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:52:21 crc kubenswrapper[4821]: I1125 10:52:21.878236 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e305fedf-6c46-4c0e-9f05-ce67a8b2668c","Type":"ContainerStarted","Data":"4884242d869fcb68acbcbd8e4bdbb081b5a3c1dd10fa4dca42f46a2290bc70cc"} Nov 25 10:52:21 crc kubenswrapper[4821]: I1125 10:52:21.878273 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e305fedf-6c46-4c0e-9f05-ce67a8b2668c","Type":"ContainerStarted","Data":"efc640699b2f0c761ecc6cdd93bb7592387a23fd71b7c895e0a87b94aeeb0420"} Nov 25 10:52:21 crc kubenswrapper[4821]: I1125 10:52:21.905468 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.905450003 podStartE2EDuration="2.905450003s" podCreationTimestamp="2025-11-25 10:52:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:52:21.899565947 +0000 UTC m=+1212.435885804" watchObservedRunningTime="2025-11-25 10:52:21.905450003 +0000 UTC m=+1212.441769850" Nov 25 10:52:21 crc kubenswrapper[4821]: I1125 10:52:21.958885 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:52:21 crc kubenswrapper[4821]: W1125 10:52:21.965373 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08ee9d9a_c488_4289_9d8f_a87d338992d6.slice/crio-733a015154109420fda47eb84065779b370bf859e4b9750e02c26a0562c23257 WatchSource:0}: Error finding container 733a015154109420fda47eb84065779b370bf859e4b9750e02c26a0562c23257: Status 404 returned error can't find the container with id 733a015154109420fda47eb84065779b370bf859e4b9750e02c26a0562c23257 Nov 25 10:52:22 crc kubenswrapper[4821]: I1125 10:52:22.130301 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a" path="/var/lib/kubelet/pods/8d0e63e3-7a64-4127-9df2-5c7f00fa4d2a/volumes" Nov 25 10:52:22 crc kubenswrapper[4821]: I1125 10:52:22.886682 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"08ee9d9a-c488-4289-9d8f-a87d338992d6","Type":"ContainerStarted","Data":"733a015154109420fda47eb84065779b370bf859e4b9750e02c26a0562c23257"} Nov 25 10:52:23 crc kubenswrapper[4821]: I1125 10:52:23.835371 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-cd5cbd7b9-xd4pf" Nov 25 10:52:23 crc kubenswrapper[4821]: I1125 10:52:23.898215 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"08ee9d9a-c488-4289-9d8f-a87d338992d6","Type":"ContainerStarted","Data":"2200f93eeec32061204385a7b0a6093dc6fd4a9953ea5f3ae58bc2c1d3a13606"} Nov 25 10:52:23 crc kubenswrapper[4821]: I1125 10:52:23.915967 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-wsvhq"] Nov 25 10:52:23 crc kubenswrapper[4821]: I1125 10:52:23.916221 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-bccf8f775-wsvhq" podUID="77c23b7d-3861-4de3-b70f-d109bd949332" containerName="dnsmasq-dns" containerID="cri-o://751a176c0aa45d407c93fde3d89d4d41fa1f9c1889e805cc20cb48e901ba8d91" gracePeriod=10 Nov 25 10:52:24 crc kubenswrapper[4821]: I1125 10:52:24.158546 4821 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-bccf8f775-wsvhq" podUID="77c23b7d-3861-4de3-b70f-d109bd949332" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.190:5353: connect: connection refused" Nov 25 10:52:24 crc kubenswrapper[4821]: I1125 10:52:24.468067 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-wsvhq" Nov 25 10:52:24 crc kubenswrapper[4821]: I1125 10:52:24.652125 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/77c23b7d-3861-4de3-b70f-d109bd949332-ovsdbserver-sb\") pod \"77c23b7d-3861-4de3-b70f-d109bd949332\" (UID: \"77c23b7d-3861-4de3-b70f-d109bd949332\") " Nov 25 10:52:24 crc kubenswrapper[4821]: I1125 10:52:24.652234 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/77c23b7d-3861-4de3-b70f-d109bd949332-dns-svc\") pod \"77c23b7d-3861-4de3-b70f-d109bd949332\" (UID: \"77c23b7d-3861-4de3-b70f-d109bd949332\") " Nov 25 10:52:24 crc kubenswrapper[4821]: I1125 10:52:24.652277 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/77c23b7d-3861-4de3-b70f-d109bd949332-config\") pod \"77c23b7d-3861-4de3-b70f-d109bd949332\" (UID: \"77c23b7d-3861-4de3-b70f-d109bd949332\") " Nov 25 10:52:24 crc kubenswrapper[4821]: I1125 10:52:24.652331 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-59dq7\" (UniqueName: \"kubernetes.io/projected/77c23b7d-3861-4de3-b70f-d109bd949332-kube-api-access-59dq7\") pod \"77c23b7d-3861-4de3-b70f-d109bd949332\" (UID: \"77c23b7d-3861-4de3-b70f-d109bd949332\") " Nov 25 10:52:24 crc kubenswrapper[4821]: I1125 10:52:24.652374 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/77c23b7d-3861-4de3-b70f-d109bd949332-dns-swift-storage-0\") pod \"77c23b7d-3861-4de3-b70f-d109bd949332\" (UID: \"77c23b7d-3861-4de3-b70f-d109bd949332\") " Nov 25 10:52:24 crc kubenswrapper[4821]: I1125 10:52:24.652430 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/77c23b7d-3861-4de3-b70f-d109bd949332-ovsdbserver-nb\") pod \"77c23b7d-3861-4de3-b70f-d109bd949332\" (UID: \"77c23b7d-3861-4de3-b70f-d109bd949332\") " Nov 25 10:52:24 crc kubenswrapper[4821]: I1125 10:52:24.697746 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77c23b7d-3861-4de3-b70f-d109bd949332-kube-api-access-59dq7" (OuterVolumeSpecName: "kube-api-access-59dq7") pod "77c23b7d-3861-4de3-b70f-d109bd949332" (UID: "77c23b7d-3861-4de3-b70f-d109bd949332"). InnerVolumeSpecName "kube-api-access-59dq7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:52:24 crc kubenswrapper[4821]: I1125 10:52:24.755236 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-59dq7\" (UniqueName: \"kubernetes.io/projected/77c23b7d-3861-4de3-b70f-d109bd949332-kube-api-access-59dq7\") on node \"crc\" DevicePath \"\"" Nov 25 10:52:24 crc kubenswrapper[4821]: I1125 10:52:24.786238 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/77c23b7d-3861-4de3-b70f-d109bd949332-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "77c23b7d-3861-4de3-b70f-d109bd949332" (UID: "77c23b7d-3861-4de3-b70f-d109bd949332"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:52:24 crc kubenswrapper[4821]: I1125 10:52:24.786274 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/77c23b7d-3861-4de3-b70f-d109bd949332-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "77c23b7d-3861-4de3-b70f-d109bd949332" (UID: "77c23b7d-3861-4de3-b70f-d109bd949332"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:52:24 crc kubenswrapper[4821]: I1125 10:52:24.813212 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/77c23b7d-3861-4de3-b70f-d109bd949332-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "77c23b7d-3861-4de3-b70f-d109bd949332" (UID: "77c23b7d-3861-4de3-b70f-d109bd949332"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:52:24 crc kubenswrapper[4821]: I1125 10:52:24.824039 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/77c23b7d-3861-4de3-b70f-d109bd949332-config" (OuterVolumeSpecName: "config") pod "77c23b7d-3861-4de3-b70f-d109bd949332" (UID: "77c23b7d-3861-4de3-b70f-d109bd949332"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:52:24 crc kubenswrapper[4821]: I1125 10:52:24.846761 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/77c23b7d-3861-4de3-b70f-d109bd949332-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "77c23b7d-3861-4de3-b70f-d109bd949332" (UID: "77c23b7d-3861-4de3-b70f-d109bd949332"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:52:24 crc kubenswrapper[4821]: I1125 10:52:24.858147 4821 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/77c23b7d-3861-4de3-b70f-d109bd949332-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 10:52:24 crc kubenswrapper[4821]: I1125 10:52:24.858226 4821 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/77c23b7d-3861-4de3-b70f-d109bd949332-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 10:52:24 crc kubenswrapper[4821]: I1125 10:52:24.858239 4821 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/77c23b7d-3861-4de3-b70f-d109bd949332-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:52:24 crc kubenswrapper[4821]: I1125 10:52:24.858250 4821 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/77c23b7d-3861-4de3-b70f-d109bd949332-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 25 10:52:24 crc kubenswrapper[4821]: I1125 10:52:24.858263 4821 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/77c23b7d-3861-4de3-b70f-d109bd949332-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 10:52:24 crc kubenswrapper[4821]: I1125 10:52:24.908794 4821 generic.go:334] "Generic (PLEG): container finished" podID="77c23b7d-3861-4de3-b70f-d109bd949332" containerID="751a176c0aa45d407c93fde3d89d4d41fa1f9c1889e805cc20cb48e901ba8d91" exitCode=0 Nov 25 10:52:24 crc kubenswrapper[4821]: I1125 10:52:24.908869 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-wsvhq" event={"ID":"77c23b7d-3861-4de3-b70f-d109bd949332","Type":"ContainerDied","Data":"751a176c0aa45d407c93fde3d89d4d41fa1f9c1889e805cc20cb48e901ba8d91"} Nov 25 10:52:24 crc kubenswrapper[4821]: I1125 10:52:24.908908 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-wsvhq" event={"ID":"77c23b7d-3861-4de3-b70f-d109bd949332","Type":"ContainerDied","Data":"f8273fbd9ec7b7cbec8d0714eb38beea4a7567dfd2c5c6998b30a3df8b54e6f2"} Nov 25 10:52:24 crc kubenswrapper[4821]: I1125 10:52:24.908930 4821 scope.go:117] "RemoveContainer" containerID="751a176c0aa45d407c93fde3d89d4d41fa1f9c1889e805cc20cb48e901ba8d91" Nov 25 10:52:24 crc kubenswrapper[4821]: I1125 10:52:24.909136 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-wsvhq" Nov 25 10:52:24 crc kubenswrapper[4821]: I1125 10:52:24.944334 4821 scope.go:117] "RemoveContainer" containerID="487c366b60dacb4ec493cea4a8993ad585a343d876f19bb13ce7f26915f559e1" Nov 25 10:52:24 crc kubenswrapper[4821]: I1125 10:52:24.960513 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-wsvhq"] Nov 25 10:52:24 crc kubenswrapper[4821]: I1125 10:52:24.976349 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-wsvhq"] Nov 25 10:52:24 crc kubenswrapper[4821]: I1125 10:52:24.982543 4821 scope.go:117] "RemoveContainer" containerID="751a176c0aa45d407c93fde3d89d4d41fa1f9c1889e805cc20cb48e901ba8d91" Nov 25 10:52:24 crc kubenswrapper[4821]: E1125 10:52:24.983049 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"751a176c0aa45d407c93fde3d89d4d41fa1f9c1889e805cc20cb48e901ba8d91\": container with ID starting with 751a176c0aa45d407c93fde3d89d4d41fa1f9c1889e805cc20cb48e901ba8d91 not found: ID does not exist" containerID="751a176c0aa45d407c93fde3d89d4d41fa1f9c1889e805cc20cb48e901ba8d91" Nov 25 10:52:24 crc kubenswrapper[4821]: I1125 10:52:24.983093 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"751a176c0aa45d407c93fde3d89d4d41fa1f9c1889e805cc20cb48e901ba8d91"} err="failed to get container status \"751a176c0aa45d407c93fde3d89d4d41fa1f9c1889e805cc20cb48e901ba8d91\": rpc error: code = NotFound desc = could not find container \"751a176c0aa45d407c93fde3d89d4d41fa1f9c1889e805cc20cb48e901ba8d91\": container with ID starting with 751a176c0aa45d407c93fde3d89d4d41fa1f9c1889e805cc20cb48e901ba8d91 not found: ID does not exist" Nov 25 10:52:24 crc kubenswrapper[4821]: I1125 10:52:24.983120 4821 scope.go:117] "RemoveContainer" containerID="487c366b60dacb4ec493cea4a8993ad585a343d876f19bb13ce7f26915f559e1" Nov 25 10:52:24 crc kubenswrapper[4821]: E1125 10:52:24.983873 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"487c366b60dacb4ec493cea4a8993ad585a343d876f19bb13ce7f26915f559e1\": container with ID starting with 487c366b60dacb4ec493cea4a8993ad585a343d876f19bb13ce7f26915f559e1 not found: ID does not exist" containerID="487c366b60dacb4ec493cea4a8993ad585a343d876f19bb13ce7f26915f559e1" Nov 25 10:52:24 crc kubenswrapper[4821]: I1125 10:52:24.983909 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"487c366b60dacb4ec493cea4a8993ad585a343d876f19bb13ce7f26915f559e1"} err="failed to get container status \"487c366b60dacb4ec493cea4a8993ad585a343d876f19bb13ce7f26915f559e1\": rpc error: code = NotFound desc = could not find container \"487c366b60dacb4ec493cea4a8993ad585a343d876f19bb13ce7f26915f559e1\": container with ID starting with 487c366b60dacb4ec493cea4a8993ad585a343d876f19bb13ce7f26915f559e1 not found: ID does not exist" Nov 25 10:52:25 crc kubenswrapper[4821]: I1125 10:52:25.175390 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:52:25 crc kubenswrapper[4821]: I1125 10:52:25.194096 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:52:25 crc kubenswrapper[4821]: I1125 10:52:25.923059 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"08ee9d9a-c488-4289-9d8f-a87d338992d6","Type":"ContainerStarted","Data":"c619538ee1a375af6e06c44e258d883dd7e49ec9b1a5f121bb2c437bc4d98d3f"} Nov 25 10:52:25 crc kubenswrapper[4821]: I1125 10:52:25.924143 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"08ee9d9a-c488-4289-9d8f-a87d338992d6","Type":"ContainerStarted","Data":"39f555f9b1828d5b4d5a6813e10eeddfeff4fb20d8bf7e5451c6c6f81b66b6fb"} Nov 25 10:52:25 crc kubenswrapper[4821]: I1125 10:52:25.942814 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:52:26 crc kubenswrapper[4821]: I1125 10:52:26.126184 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="77c23b7d-3861-4de3-b70f-d109bd949332" path="/var/lib/kubelet/pods/77c23b7d-3861-4de3-b70f-d109bd949332/volumes" Nov 25 10:52:26 crc kubenswrapper[4821]: I1125 10:52:26.232220 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-p42p6"] Nov 25 10:52:26 crc kubenswrapper[4821]: E1125 10:52:26.232727 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77c23b7d-3861-4de3-b70f-d109bd949332" containerName="init" Nov 25 10:52:26 crc kubenswrapper[4821]: I1125 10:52:26.232755 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="77c23b7d-3861-4de3-b70f-d109bd949332" containerName="init" Nov 25 10:52:26 crc kubenswrapper[4821]: E1125 10:52:26.232769 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77c23b7d-3861-4de3-b70f-d109bd949332" containerName="dnsmasq-dns" Nov 25 10:52:26 crc kubenswrapper[4821]: I1125 10:52:26.232780 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="77c23b7d-3861-4de3-b70f-d109bd949332" containerName="dnsmasq-dns" Nov 25 10:52:26 crc kubenswrapper[4821]: I1125 10:52:26.233028 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="77c23b7d-3861-4de3-b70f-d109bd949332" containerName="dnsmasq-dns" Nov 25 10:52:26 crc kubenswrapper[4821]: I1125 10:52:26.234645 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-p42p6" Nov 25 10:52:26 crc kubenswrapper[4821]: I1125 10:52:26.239581 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Nov 25 10:52:26 crc kubenswrapper[4821]: I1125 10:52:26.239833 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Nov 25 10:52:26 crc kubenswrapper[4821]: I1125 10:52:26.248681 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-p42p6"] Nov 25 10:52:26 crc kubenswrapper[4821]: I1125 10:52:26.292396 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d463843-fed0-4215-a85c-aac27aa1de50-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-p42p6\" (UID: \"0d463843-fed0-4215-a85c-aac27aa1de50\") " pod="openstack/nova-cell1-cell-mapping-p42p6" Nov 25 10:52:26 crc kubenswrapper[4821]: I1125 10:52:26.292550 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzrxh\" (UniqueName: \"kubernetes.io/projected/0d463843-fed0-4215-a85c-aac27aa1de50-kube-api-access-fzrxh\") pod \"nova-cell1-cell-mapping-p42p6\" (UID: \"0d463843-fed0-4215-a85c-aac27aa1de50\") " pod="openstack/nova-cell1-cell-mapping-p42p6" Nov 25 10:52:26 crc kubenswrapper[4821]: I1125 10:52:26.292636 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d463843-fed0-4215-a85c-aac27aa1de50-config-data\") pod \"nova-cell1-cell-mapping-p42p6\" (UID: \"0d463843-fed0-4215-a85c-aac27aa1de50\") " pod="openstack/nova-cell1-cell-mapping-p42p6" Nov 25 10:52:26 crc kubenswrapper[4821]: I1125 10:52:26.292714 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d463843-fed0-4215-a85c-aac27aa1de50-scripts\") pod \"nova-cell1-cell-mapping-p42p6\" (UID: \"0d463843-fed0-4215-a85c-aac27aa1de50\") " pod="openstack/nova-cell1-cell-mapping-p42p6" Nov 25 10:52:26 crc kubenswrapper[4821]: I1125 10:52:26.365723 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 25 10:52:26 crc kubenswrapper[4821]: I1125 10:52:26.365789 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 25 10:52:26 crc kubenswrapper[4821]: I1125 10:52:26.398512 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d463843-fed0-4215-a85c-aac27aa1de50-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-p42p6\" (UID: \"0d463843-fed0-4215-a85c-aac27aa1de50\") " pod="openstack/nova-cell1-cell-mapping-p42p6" Nov 25 10:52:26 crc kubenswrapper[4821]: I1125 10:52:26.398623 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzrxh\" (UniqueName: \"kubernetes.io/projected/0d463843-fed0-4215-a85c-aac27aa1de50-kube-api-access-fzrxh\") pod \"nova-cell1-cell-mapping-p42p6\" (UID: \"0d463843-fed0-4215-a85c-aac27aa1de50\") " pod="openstack/nova-cell1-cell-mapping-p42p6" Nov 25 10:52:26 crc kubenswrapper[4821]: I1125 10:52:26.398681 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d463843-fed0-4215-a85c-aac27aa1de50-config-data\") pod \"nova-cell1-cell-mapping-p42p6\" (UID: \"0d463843-fed0-4215-a85c-aac27aa1de50\") " pod="openstack/nova-cell1-cell-mapping-p42p6" Nov 25 10:52:26 crc kubenswrapper[4821]: I1125 10:52:26.398744 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d463843-fed0-4215-a85c-aac27aa1de50-scripts\") pod \"nova-cell1-cell-mapping-p42p6\" (UID: \"0d463843-fed0-4215-a85c-aac27aa1de50\") " pod="openstack/nova-cell1-cell-mapping-p42p6" Nov 25 10:52:26 crc kubenswrapper[4821]: I1125 10:52:26.405866 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d463843-fed0-4215-a85c-aac27aa1de50-config-data\") pod \"nova-cell1-cell-mapping-p42p6\" (UID: \"0d463843-fed0-4215-a85c-aac27aa1de50\") " pod="openstack/nova-cell1-cell-mapping-p42p6" Nov 25 10:52:26 crc kubenswrapper[4821]: I1125 10:52:26.406330 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d463843-fed0-4215-a85c-aac27aa1de50-scripts\") pod \"nova-cell1-cell-mapping-p42p6\" (UID: \"0d463843-fed0-4215-a85c-aac27aa1de50\") " pod="openstack/nova-cell1-cell-mapping-p42p6" Nov 25 10:52:26 crc kubenswrapper[4821]: I1125 10:52:26.416839 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d463843-fed0-4215-a85c-aac27aa1de50-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-p42p6\" (UID: \"0d463843-fed0-4215-a85c-aac27aa1de50\") " pod="openstack/nova-cell1-cell-mapping-p42p6" Nov 25 10:52:26 crc kubenswrapper[4821]: I1125 10:52:26.430880 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzrxh\" (UniqueName: \"kubernetes.io/projected/0d463843-fed0-4215-a85c-aac27aa1de50-kube-api-access-fzrxh\") pod \"nova-cell1-cell-mapping-p42p6\" (UID: \"0d463843-fed0-4215-a85c-aac27aa1de50\") " pod="openstack/nova-cell1-cell-mapping-p42p6" Nov 25 10:52:26 crc kubenswrapper[4821]: I1125 10:52:26.554030 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-p42p6" Nov 25 10:52:27 crc kubenswrapper[4821]: I1125 10:52:27.051786 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-p42p6"] Nov 25 10:52:27 crc kubenswrapper[4821]: I1125 10:52:27.378373 4821 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="802c0cd6-5dd0-41d9-ae10-12cd23aa0467" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.199:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 25 10:52:27 crc kubenswrapper[4821]: I1125 10:52:27.378373 4821 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="802c0cd6-5dd0-41d9-ae10-12cd23aa0467" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.199:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 25 10:52:27 crc kubenswrapper[4821]: I1125 10:52:27.944278 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-p42p6" event={"ID":"0d463843-fed0-4215-a85c-aac27aa1de50","Type":"ContainerStarted","Data":"93034c9c531c3037c12f30ef6bbf8418d7c92ca6f75ffe9429cb9b35c22c56c2"} Nov 25 10:52:27 crc kubenswrapper[4821]: I1125 10:52:27.944320 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-p42p6" event={"ID":"0d463843-fed0-4215-a85c-aac27aa1de50","Type":"ContainerStarted","Data":"6a3efd21d183c7120c95815b30ce571d09280d3508806104ade0e42541c00361"} Nov 25 10:52:27 crc kubenswrapper[4821]: I1125 10:52:27.946442 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"08ee9d9a-c488-4289-9d8f-a87d338992d6","Type":"ContainerStarted","Data":"2b4700a1f4eac34d5d1ff0de837b2c4d7c98eb4d1e5435aa3a77b5ec9c6e8e3c"} Nov 25 10:52:27 crc kubenswrapper[4821]: I1125 10:52:27.947505 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 25 10:52:27 crc kubenswrapper[4821]: I1125 10:52:27.963904 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-p42p6" podStartSLOduration=1.963883898 podStartE2EDuration="1.963883898s" podCreationTimestamp="2025-11-25 10:52:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:52:27.959278766 +0000 UTC m=+1218.495598613" watchObservedRunningTime="2025-11-25 10:52:27.963883898 +0000 UTC m=+1218.500203745" Nov 25 10:52:27 crc kubenswrapper[4821]: I1125 10:52:27.995910 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.748625305 podStartE2EDuration="7.995884828s" podCreationTimestamp="2025-11-25 10:52:20 +0000 UTC" firstStartedPulling="2025-11-25 10:52:21.970729418 +0000 UTC m=+1212.507049265" lastFinishedPulling="2025-11-25 10:52:27.217988941 +0000 UTC m=+1217.754308788" observedRunningTime="2025-11-25 10:52:27.980988537 +0000 UTC m=+1218.517308384" watchObservedRunningTime="2025-11-25 10:52:27.995884828 +0000 UTC m=+1218.532204675" Nov 25 10:52:30 crc kubenswrapper[4821]: I1125 10:52:30.222767 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 25 10:52:30 crc kubenswrapper[4821]: I1125 10:52:30.222819 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 25 10:52:31 crc kubenswrapper[4821]: I1125 10:52:31.239728 4821 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="e305fedf-6c46-4c0e-9f05-ce67a8b2668c" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.200:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 25 10:52:31 crc kubenswrapper[4821]: I1125 10:52:31.239789 4821 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="e305fedf-6c46-4c0e-9f05-ce67a8b2668c" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.200:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 25 10:52:32 crc kubenswrapper[4821]: I1125 10:52:32.997763 4821 generic.go:334] "Generic (PLEG): container finished" podID="0d463843-fed0-4215-a85c-aac27aa1de50" containerID="93034c9c531c3037c12f30ef6bbf8418d7c92ca6f75ffe9429cb9b35c22c56c2" exitCode=0 Nov 25 10:52:32 crc kubenswrapper[4821]: I1125 10:52:32.997852 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-p42p6" event={"ID":"0d463843-fed0-4215-a85c-aac27aa1de50","Type":"ContainerDied","Data":"93034c9c531c3037c12f30ef6bbf8418d7c92ca6f75ffe9429cb9b35c22c56c2"} Nov 25 10:52:34 crc kubenswrapper[4821]: I1125 10:52:34.408564 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-p42p6" Nov 25 10:52:34 crc kubenswrapper[4821]: I1125 10:52:34.490790 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fzrxh\" (UniqueName: \"kubernetes.io/projected/0d463843-fed0-4215-a85c-aac27aa1de50-kube-api-access-fzrxh\") pod \"0d463843-fed0-4215-a85c-aac27aa1de50\" (UID: \"0d463843-fed0-4215-a85c-aac27aa1de50\") " Nov 25 10:52:34 crc kubenswrapper[4821]: I1125 10:52:34.490946 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d463843-fed0-4215-a85c-aac27aa1de50-config-data\") pod \"0d463843-fed0-4215-a85c-aac27aa1de50\" (UID: \"0d463843-fed0-4215-a85c-aac27aa1de50\") " Nov 25 10:52:34 crc kubenswrapper[4821]: I1125 10:52:34.491119 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d463843-fed0-4215-a85c-aac27aa1de50-scripts\") pod \"0d463843-fed0-4215-a85c-aac27aa1de50\" (UID: \"0d463843-fed0-4215-a85c-aac27aa1de50\") " Nov 25 10:52:34 crc kubenswrapper[4821]: I1125 10:52:34.491146 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d463843-fed0-4215-a85c-aac27aa1de50-combined-ca-bundle\") pod \"0d463843-fed0-4215-a85c-aac27aa1de50\" (UID: \"0d463843-fed0-4215-a85c-aac27aa1de50\") " Nov 25 10:52:34 crc kubenswrapper[4821]: I1125 10:52:34.496873 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d463843-fed0-4215-a85c-aac27aa1de50-scripts" (OuterVolumeSpecName: "scripts") pod "0d463843-fed0-4215-a85c-aac27aa1de50" (UID: "0d463843-fed0-4215-a85c-aac27aa1de50"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:52:34 crc kubenswrapper[4821]: I1125 10:52:34.497307 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d463843-fed0-4215-a85c-aac27aa1de50-kube-api-access-fzrxh" (OuterVolumeSpecName: "kube-api-access-fzrxh") pod "0d463843-fed0-4215-a85c-aac27aa1de50" (UID: "0d463843-fed0-4215-a85c-aac27aa1de50"). InnerVolumeSpecName "kube-api-access-fzrxh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:52:34 crc kubenswrapper[4821]: I1125 10:52:34.521670 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d463843-fed0-4215-a85c-aac27aa1de50-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0d463843-fed0-4215-a85c-aac27aa1de50" (UID: "0d463843-fed0-4215-a85c-aac27aa1de50"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:52:34 crc kubenswrapper[4821]: I1125 10:52:34.523901 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d463843-fed0-4215-a85c-aac27aa1de50-config-data" (OuterVolumeSpecName: "config-data") pod "0d463843-fed0-4215-a85c-aac27aa1de50" (UID: "0d463843-fed0-4215-a85c-aac27aa1de50"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:52:34 crc kubenswrapper[4821]: I1125 10:52:34.594627 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fzrxh\" (UniqueName: \"kubernetes.io/projected/0d463843-fed0-4215-a85c-aac27aa1de50-kube-api-access-fzrxh\") on node \"crc\" DevicePath \"\"" Nov 25 10:52:34 crc kubenswrapper[4821]: I1125 10:52:34.594695 4821 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d463843-fed0-4215-a85c-aac27aa1de50-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:52:34 crc kubenswrapper[4821]: I1125 10:52:34.594708 4821 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d463843-fed0-4215-a85c-aac27aa1de50-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:52:34 crc kubenswrapper[4821]: I1125 10:52:34.594718 4821 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d463843-fed0-4215-a85c-aac27aa1de50-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:52:35 crc kubenswrapper[4821]: I1125 10:52:35.017540 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-p42p6" event={"ID":"0d463843-fed0-4215-a85c-aac27aa1de50","Type":"ContainerDied","Data":"6a3efd21d183c7120c95815b30ce571d09280d3508806104ade0e42541c00361"} Nov 25 10:52:35 crc kubenswrapper[4821]: I1125 10:52:35.017616 4821 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6a3efd21d183c7120c95815b30ce571d09280d3508806104ade0e42541c00361" Nov 25 10:52:35 crc kubenswrapper[4821]: I1125 10:52:35.017626 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-p42p6" Nov 25 10:52:35 crc kubenswrapper[4821]: I1125 10:52:35.189239 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 25 10:52:35 crc kubenswrapper[4821]: I1125 10:52:35.189610 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="e305fedf-6c46-4c0e-9f05-ce67a8b2668c" containerName="nova-api-log" containerID="cri-o://efc640699b2f0c761ecc6cdd93bb7592387a23fd71b7c895e0a87b94aeeb0420" gracePeriod=30 Nov 25 10:52:35 crc kubenswrapper[4821]: I1125 10:52:35.189757 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="e305fedf-6c46-4c0e-9f05-ce67a8b2668c" containerName="nova-api-api" containerID="cri-o://4884242d869fcb68acbcbd8e4bdbb081b5a3c1dd10fa4dca42f46a2290bc70cc" gracePeriod=30 Nov 25 10:52:35 crc kubenswrapper[4821]: I1125 10:52:35.199341 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 10:52:35 crc kubenswrapper[4821]: I1125 10:52:35.199588 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="c41109a7-e905-4087-9d04-ccd2cccb42b5" containerName="nova-scheduler-scheduler" containerID="cri-o://f29d85ca7d099c678b1f559610a961965055b87a55a28a98b8c43a29b9e80703" gracePeriod=30 Nov 25 10:52:35 crc kubenswrapper[4821]: I1125 10:52:35.218765 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 10:52:35 crc kubenswrapper[4821]: I1125 10:52:35.219324 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="802c0cd6-5dd0-41d9-ae10-12cd23aa0467" containerName="nova-metadata-log" containerID="cri-o://078a3f7579d31eb42966404567599c09306bdc5804c0c47e780e7f31e77f7928" gracePeriod=30 Nov 25 10:52:35 crc kubenswrapper[4821]: I1125 10:52:35.219436 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="802c0cd6-5dd0-41d9-ae10-12cd23aa0467" containerName="nova-metadata-metadata" containerID="cri-o://9e2d63189a92dad7e806e26293da984c4e4eba7281abfb9f21111046c6cbe494" gracePeriod=30 Nov 25 10:52:36 crc kubenswrapper[4821]: I1125 10:52:36.032779 4821 generic.go:334] "Generic (PLEG): container finished" podID="e305fedf-6c46-4c0e-9f05-ce67a8b2668c" containerID="efc640699b2f0c761ecc6cdd93bb7592387a23fd71b7c895e0a87b94aeeb0420" exitCode=143 Nov 25 10:52:36 crc kubenswrapper[4821]: I1125 10:52:36.032878 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e305fedf-6c46-4c0e-9f05-ce67a8b2668c","Type":"ContainerDied","Data":"efc640699b2f0c761ecc6cdd93bb7592387a23fd71b7c895e0a87b94aeeb0420"} Nov 25 10:52:36 crc kubenswrapper[4821]: I1125 10:52:36.036624 4821 generic.go:334] "Generic (PLEG): container finished" podID="802c0cd6-5dd0-41d9-ae10-12cd23aa0467" containerID="078a3f7579d31eb42966404567599c09306bdc5804c0c47e780e7f31e77f7928" exitCode=143 Nov 25 10:52:36 crc kubenswrapper[4821]: I1125 10:52:36.036669 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"802c0cd6-5dd0-41d9-ae10-12cd23aa0467","Type":"ContainerDied","Data":"078a3f7579d31eb42966404567599c09306bdc5804c0c47e780e7f31e77f7928"} Nov 25 10:52:37 crc kubenswrapper[4821]: I1125 10:52:37.054808 4821 generic.go:334] "Generic (PLEG): container finished" podID="c41109a7-e905-4087-9d04-ccd2cccb42b5" containerID="f29d85ca7d099c678b1f559610a961965055b87a55a28a98b8c43a29b9e80703" exitCode=0 Nov 25 10:52:37 crc kubenswrapper[4821]: I1125 10:52:37.054959 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c41109a7-e905-4087-9d04-ccd2cccb42b5","Type":"ContainerDied","Data":"f29d85ca7d099c678b1f559610a961965055b87a55a28a98b8c43a29b9e80703"} Nov 25 10:52:37 crc kubenswrapper[4821]: I1125 10:52:37.342122 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 25 10:52:37 crc kubenswrapper[4821]: I1125 10:52:37.449773 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f78n9\" (UniqueName: \"kubernetes.io/projected/c41109a7-e905-4087-9d04-ccd2cccb42b5-kube-api-access-f78n9\") pod \"c41109a7-e905-4087-9d04-ccd2cccb42b5\" (UID: \"c41109a7-e905-4087-9d04-ccd2cccb42b5\") " Nov 25 10:52:37 crc kubenswrapper[4821]: I1125 10:52:37.449883 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c41109a7-e905-4087-9d04-ccd2cccb42b5-combined-ca-bundle\") pod \"c41109a7-e905-4087-9d04-ccd2cccb42b5\" (UID: \"c41109a7-e905-4087-9d04-ccd2cccb42b5\") " Nov 25 10:52:37 crc kubenswrapper[4821]: I1125 10:52:37.450056 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c41109a7-e905-4087-9d04-ccd2cccb42b5-config-data\") pod \"c41109a7-e905-4087-9d04-ccd2cccb42b5\" (UID: \"c41109a7-e905-4087-9d04-ccd2cccb42b5\") " Nov 25 10:52:37 crc kubenswrapper[4821]: I1125 10:52:37.456175 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c41109a7-e905-4087-9d04-ccd2cccb42b5-kube-api-access-f78n9" (OuterVolumeSpecName: "kube-api-access-f78n9") pod "c41109a7-e905-4087-9d04-ccd2cccb42b5" (UID: "c41109a7-e905-4087-9d04-ccd2cccb42b5"). InnerVolumeSpecName "kube-api-access-f78n9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:52:37 crc kubenswrapper[4821]: I1125 10:52:37.479383 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c41109a7-e905-4087-9d04-ccd2cccb42b5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c41109a7-e905-4087-9d04-ccd2cccb42b5" (UID: "c41109a7-e905-4087-9d04-ccd2cccb42b5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:52:37 crc kubenswrapper[4821]: I1125 10:52:37.484599 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c41109a7-e905-4087-9d04-ccd2cccb42b5-config-data" (OuterVolumeSpecName: "config-data") pod "c41109a7-e905-4087-9d04-ccd2cccb42b5" (UID: "c41109a7-e905-4087-9d04-ccd2cccb42b5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:52:37 crc kubenswrapper[4821]: I1125 10:52:37.553708 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f78n9\" (UniqueName: \"kubernetes.io/projected/c41109a7-e905-4087-9d04-ccd2cccb42b5-kube-api-access-f78n9\") on node \"crc\" DevicePath \"\"" Nov 25 10:52:37 crc kubenswrapper[4821]: I1125 10:52:37.553763 4821 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c41109a7-e905-4087-9d04-ccd2cccb42b5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:52:37 crc kubenswrapper[4821]: I1125 10:52:37.553789 4821 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c41109a7-e905-4087-9d04-ccd2cccb42b5-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:52:38 crc kubenswrapper[4821]: I1125 10:52:38.067225 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c41109a7-e905-4087-9d04-ccd2cccb42b5","Type":"ContainerDied","Data":"07af87afbb24e0995e8d0ac6066d64d946c2185a21a95bb878e0c864bd67a08d"} Nov 25 10:52:38 crc kubenswrapper[4821]: I1125 10:52:38.067287 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 25 10:52:38 crc kubenswrapper[4821]: I1125 10:52:38.067295 4821 scope.go:117] "RemoveContainer" containerID="f29d85ca7d099c678b1f559610a961965055b87a55a28a98b8c43a29b9e80703" Nov 25 10:52:38 crc kubenswrapper[4821]: I1125 10:52:38.111733 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 10:52:38 crc kubenswrapper[4821]: I1125 10:52:38.131486 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 10:52:38 crc kubenswrapper[4821]: I1125 10:52:38.142262 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 10:52:38 crc kubenswrapper[4821]: E1125 10:52:38.142727 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d463843-fed0-4215-a85c-aac27aa1de50" containerName="nova-manage" Nov 25 10:52:38 crc kubenswrapper[4821]: I1125 10:52:38.142750 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d463843-fed0-4215-a85c-aac27aa1de50" containerName="nova-manage" Nov 25 10:52:38 crc kubenswrapper[4821]: E1125 10:52:38.142821 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c41109a7-e905-4087-9d04-ccd2cccb42b5" containerName="nova-scheduler-scheduler" Nov 25 10:52:38 crc kubenswrapper[4821]: I1125 10:52:38.142831 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="c41109a7-e905-4087-9d04-ccd2cccb42b5" containerName="nova-scheduler-scheduler" Nov 25 10:52:38 crc kubenswrapper[4821]: I1125 10:52:38.143037 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="c41109a7-e905-4087-9d04-ccd2cccb42b5" containerName="nova-scheduler-scheduler" Nov 25 10:52:38 crc kubenswrapper[4821]: I1125 10:52:38.143083 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d463843-fed0-4215-a85c-aac27aa1de50" containerName="nova-manage" Nov 25 10:52:38 crc kubenswrapper[4821]: I1125 10:52:38.143834 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 25 10:52:38 crc kubenswrapper[4821]: I1125 10:52:38.153500 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 25 10:52:38 crc kubenswrapper[4821]: I1125 10:52:38.163436 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 10:52:38 crc kubenswrapper[4821]: I1125 10:52:38.266456 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bdff849-753c-42b5-9e06-2e37dfe42666-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"3bdff849-753c-42b5-9e06-2e37dfe42666\") " pod="openstack/nova-scheduler-0" Nov 25 10:52:38 crc kubenswrapper[4821]: I1125 10:52:38.266682 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cvmr9\" (UniqueName: \"kubernetes.io/projected/3bdff849-753c-42b5-9e06-2e37dfe42666-kube-api-access-cvmr9\") pod \"nova-scheduler-0\" (UID: \"3bdff849-753c-42b5-9e06-2e37dfe42666\") " pod="openstack/nova-scheduler-0" Nov 25 10:52:38 crc kubenswrapper[4821]: I1125 10:52:38.266716 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3bdff849-753c-42b5-9e06-2e37dfe42666-config-data\") pod \"nova-scheduler-0\" (UID: \"3bdff849-753c-42b5-9e06-2e37dfe42666\") " pod="openstack/nova-scheduler-0" Nov 25 10:52:38 crc kubenswrapper[4821]: I1125 10:52:38.368599 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3bdff849-753c-42b5-9e06-2e37dfe42666-config-data\") pod \"nova-scheduler-0\" (UID: \"3bdff849-753c-42b5-9e06-2e37dfe42666\") " pod="openstack/nova-scheduler-0" Nov 25 10:52:38 crc kubenswrapper[4821]: I1125 10:52:38.368775 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bdff849-753c-42b5-9e06-2e37dfe42666-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"3bdff849-753c-42b5-9e06-2e37dfe42666\") " pod="openstack/nova-scheduler-0" Nov 25 10:52:38 crc kubenswrapper[4821]: I1125 10:52:38.368806 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cvmr9\" (UniqueName: \"kubernetes.io/projected/3bdff849-753c-42b5-9e06-2e37dfe42666-kube-api-access-cvmr9\") pod \"nova-scheduler-0\" (UID: \"3bdff849-753c-42b5-9e06-2e37dfe42666\") " pod="openstack/nova-scheduler-0" Nov 25 10:52:38 crc kubenswrapper[4821]: I1125 10:52:38.376218 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3bdff849-753c-42b5-9e06-2e37dfe42666-config-data\") pod \"nova-scheduler-0\" (UID: \"3bdff849-753c-42b5-9e06-2e37dfe42666\") " pod="openstack/nova-scheduler-0" Nov 25 10:52:38 crc kubenswrapper[4821]: I1125 10:52:38.381736 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bdff849-753c-42b5-9e06-2e37dfe42666-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"3bdff849-753c-42b5-9e06-2e37dfe42666\") " pod="openstack/nova-scheduler-0" Nov 25 10:52:38 crc kubenswrapper[4821]: I1125 10:52:38.386760 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cvmr9\" (UniqueName: \"kubernetes.io/projected/3bdff849-753c-42b5-9e06-2e37dfe42666-kube-api-access-cvmr9\") pod \"nova-scheduler-0\" (UID: \"3bdff849-753c-42b5-9e06-2e37dfe42666\") " pod="openstack/nova-scheduler-0" Nov 25 10:52:38 crc kubenswrapper[4821]: I1125 10:52:38.467252 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 25 10:52:38 crc kubenswrapper[4821]: I1125 10:52:38.930112 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 10:52:38 crc kubenswrapper[4821]: I1125 10:52:38.937639 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 10:52:38 crc kubenswrapper[4821]: I1125 10:52:38.979377 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/802c0cd6-5dd0-41d9-ae10-12cd23aa0467-nova-metadata-tls-certs\") pod \"802c0cd6-5dd0-41d9-ae10-12cd23aa0467\" (UID: \"802c0cd6-5dd0-41d9-ae10-12cd23aa0467\") " Nov 25 10:52:38 crc kubenswrapper[4821]: I1125 10:52:38.979437 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e305fedf-6c46-4c0e-9f05-ce67a8b2668c-config-data\") pod \"e305fedf-6c46-4c0e-9f05-ce67a8b2668c\" (UID: \"e305fedf-6c46-4c0e-9f05-ce67a8b2668c\") " Nov 25 10:52:38 crc kubenswrapper[4821]: I1125 10:52:38.979466 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e305fedf-6c46-4c0e-9f05-ce67a8b2668c-logs\") pod \"e305fedf-6c46-4c0e-9f05-ce67a8b2668c\" (UID: \"e305fedf-6c46-4c0e-9f05-ce67a8b2668c\") " Nov 25 10:52:38 crc kubenswrapper[4821]: I1125 10:52:38.979498 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/802c0cd6-5dd0-41d9-ae10-12cd23aa0467-config-data\") pod \"802c0cd6-5dd0-41d9-ae10-12cd23aa0467\" (UID: \"802c0cd6-5dd0-41d9-ae10-12cd23aa0467\") " Nov 25 10:52:38 crc kubenswrapper[4821]: I1125 10:52:38.979566 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e305fedf-6c46-4c0e-9f05-ce67a8b2668c-internal-tls-certs\") pod \"e305fedf-6c46-4c0e-9f05-ce67a8b2668c\" (UID: \"e305fedf-6c46-4c0e-9f05-ce67a8b2668c\") " Nov 25 10:52:38 crc kubenswrapper[4821]: I1125 10:52:38.979604 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/802c0cd6-5dd0-41d9-ae10-12cd23aa0467-logs\") pod \"802c0cd6-5dd0-41d9-ae10-12cd23aa0467\" (UID: \"802c0cd6-5dd0-41d9-ae10-12cd23aa0467\") " Nov 25 10:52:38 crc kubenswrapper[4821]: I1125 10:52:38.979645 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/802c0cd6-5dd0-41d9-ae10-12cd23aa0467-combined-ca-bundle\") pod \"802c0cd6-5dd0-41d9-ae10-12cd23aa0467\" (UID: \"802c0cd6-5dd0-41d9-ae10-12cd23aa0467\") " Nov 25 10:52:38 crc kubenswrapper[4821]: I1125 10:52:38.979662 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e305fedf-6c46-4c0e-9f05-ce67a8b2668c-public-tls-certs\") pod \"e305fedf-6c46-4c0e-9f05-ce67a8b2668c\" (UID: \"e305fedf-6c46-4c0e-9f05-ce67a8b2668c\") " Nov 25 10:52:38 crc kubenswrapper[4821]: I1125 10:52:38.979706 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xgbbz\" (UniqueName: \"kubernetes.io/projected/e305fedf-6c46-4c0e-9f05-ce67a8b2668c-kube-api-access-xgbbz\") pod \"e305fedf-6c46-4c0e-9f05-ce67a8b2668c\" (UID: \"e305fedf-6c46-4c0e-9f05-ce67a8b2668c\") " Nov 25 10:52:38 crc kubenswrapper[4821]: I1125 10:52:38.979758 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e305fedf-6c46-4c0e-9f05-ce67a8b2668c-combined-ca-bundle\") pod \"e305fedf-6c46-4c0e-9f05-ce67a8b2668c\" (UID: \"e305fedf-6c46-4c0e-9f05-ce67a8b2668c\") " Nov 25 10:52:38 crc kubenswrapper[4821]: I1125 10:52:38.979800 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lbldn\" (UniqueName: \"kubernetes.io/projected/802c0cd6-5dd0-41d9-ae10-12cd23aa0467-kube-api-access-lbldn\") pod \"802c0cd6-5dd0-41d9-ae10-12cd23aa0467\" (UID: \"802c0cd6-5dd0-41d9-ae10-12cd23aa0467\") " Nov 25 10:52:38 crc kubenswrapper[4821]: I1125 10:52:38.981292 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/802c0cd6-5dd0-41d9-ae10-12cd23aa0467-logs" (OuterVolumeSpecName: "logs") pod "802c0cd6-5dd0-41d9-ae10-12cd23aa0467" (UID: "802c0cd6-5dd0-41d9-ae10-12cd23aa0467"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:52:38 crc kubenswrapper[4821]: I1125 10:52:38.981572 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e305fedf-6c46-4c0e-9f05-ce67a8b2668c-logs" (OuterVolumeSpecName: "logs") pod "e305fedf-6c46-4c0e-9f05-ce67a8b2668c" (UID: "e305fedf-6c46-4c0e-9f05-ce67a8b2668c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:52:38 crc kubenswrapper[4821]: I1125 10:52:38.988845 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e305fedf-6c46-4c0e-9f05-ce67a8b2668c-kube-api-access-xgbbz" (OuterVolumeSpecName: "kube-api-access-xgbbz") pod "e305fedf-6c46-4c0e-9f05-ce67a8b2668c" (UID: "e305fedf-6c46-4c0e-9f05-ce67a8b2668c"). InnerVolumeSpecName "kube-api-access-xgbbz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:52:38 crc kubenswrapper[4821]: I1125 10:52:38.988984 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/802c0cd6-5dd0-41d9-ae10-12cd23aa0467-kube-api-access-lbldn" (OuterVolumeSpecName: "kube-api-access-lbldn") pod "802c0cd6-5dd0-41d9-ae10-12cd23aa0467" (UID: "802c0cd6-5dd0-41d9-ae10-12cd23aa0467"). InnerVolumeSpecName "kube-api-access-lbldn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.014985 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e305fedf-6c46-4c0e-9f05-ce67a8b2668c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e305fedf-6c46-4c0e-9f05-ce67a8b2668c" (UID: "e305fedf-6c46-4c0e-9f05-ce67a8b2668c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.028413 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e305fedf-6c46-4c0e-9f05-ce67a8b2668c-config-data" (OuterVolumeSpecName: "config-data") pod "e305fedf-6c46-4c0e-9f05-ce67a8b2668c" (UID: "e305fedf-6c46-4c0e-9f05-ce67a8b2668c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.029809 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.033748 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/802c0cd6-5dd0-41d9-ae10-12cd23aa0467-config-data" (OuterVolumeSpecName: "config-data") pod "802c0cd6-5dd0-41d9-ae10-12cd23aa0467" (UID: "802c0cd6-5dd0-41d9-ae10-12cd23aa0467"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.039441 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/802c0cd6-5dd0-41d9-ae10-12cd23aa0467-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "802c0cd6-5dd0-41d9-ae10-12cd23aa0467" (UID: "802c0cd6-5dd0-41d9-ae10-12cd23aa0467"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.049837 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e305fedf-6c46-4c0e-9f05-ce67a8b2668c-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "e305fedf-6c46-4c0e-9f05-ce67a8b2668c" (UID: "e305fedf-6c46-4c0e-9f05-ce67a8b2668c"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.055779 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/802c0cd6-5dd0-41d9-ae10-12cd23aa0467-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "802c0cd6-5dd0-41d9-ae10-12cd23aa0467" (UID: "802c0cd6-5dd0-41d9-ae10-12cd23aa0467"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.066692 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e305fedf-6c46-4c0e-9f05-ce67a8b2668c-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "e305fedf-6c46-4c0e-9f05-ce67a8b2668c" (UID: "e305fedf-6c46-4c0e-9f05-ce67a8b2668c"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.078362 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"3bdff849-753c-42b5-9e06-2e37dfe42666","Type":"ContainerStarted","Data":"55a52fd22aacd625cd85492e65032bf849b97b12735d81ae591b81cc31eee19d"} Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.080140 4821 generic.go:334] "Generic (PLEG): container finished" podID="e305fedf-6c46-4c0e-9f05-ce67a8b2668c" containerID="4884242d869fcb68acbcbd8e4bdbb081b5a3c1dd10fa4dca42f46a2290bc70cc" exitCode=0 Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.080201 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e305fedf-6c46-4c0e-9f05-ce67a8b2668c","Type":"ContainerDied","Data":"4884242d869fcb68acbcbd8e4bdbb081b5a3c1dd10fa4dca42f46a2290bc70cc"} Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.080216 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.080234 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e305fedf-6c46-4c0e-9f05-ce67a8b2668c","Type":"ContainerDied","Data":"648a4a3e2d0645f1953bda347d075246f296aaef440150dcdf92f8db43a974db"} Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.080254 4821 scope.go:117] "RemoveContainer" containerID="4884242d869fcb68acbcbd8e4bdbb081b5a3c1dd10fa4dca42f46a2290bc70cc" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.082422 4821 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/802c0cd6-5dd0-41d9-ae10-12cd23aa0467-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.082466 4821 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e305fedf-6c46-4c0e-9f05-ce67a8b2668c-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.082480 4821 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e305fedf-6c46-4c0e-9f05-ce67a8b2668c-logs\") on node \"crc\" DevicePath \"\"" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.082489 4821 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/802c0cd6-5dd0-41d9-ae10-12cd23aa0467-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.082498 4821 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e305fedf-6c46-4c0e-9f05-ce67a8b2668c-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.082505 4821 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/802c0cd6-5dd0-41d9-ae10-12cd23aa0467-logs\") on node \"crc\" DevicePath \"\"" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.082513 4821 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/802c0cd6-5dd0-41d9-ae10-12cd23aa0467-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.082521 4821 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e305fedf-6c46-4c0e-9f05-ce67a8b2668c-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.082529 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xgbbz\" (UniqueName: \"kubernetes.io/projected/e305fedf-6c46-4c0e-9f05-ce67a8b2668c-kube-api-access-xgbbz\") on node \"crc\" DevicePath \"\"" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.082537 4821 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e305fedf-6c46-4c0e-9f05-ce67a8b2668c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.082545 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lbldn\" (UniqueName: \"kubernetes.io/projected/802c0cd6-5dd0-41d9-ae10-12cd23aa0467-kube-api-access-lbldn\") on node \"crc\" DevicePath \"\"" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.086580 4821 generic.go:334] "Generic (PLEG): container finished" podID="802c0cd6-5dd0-41d9-ae10-12cd23aa0467" containerID="9e2d63189a92dad7e806e26293da984c4e4eba7281abfb9f21111046c6cbe494" exitCode=0 Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.086613 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"802c0cd6-5dd0-41d9-ae10-12cd23aa0467","Type":"ContainerDied","Data":"9e2d63189a92dad7e806e26293da984c4e4eba7281abfb9f21111046c6cbe494"} Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.086757 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"802c0cd6-5dd0-41d9-ae10-12cd23aa0467","Type":"ContainerDied","Data":"ef3b6426ee31c32e17d75bc1dedc62c6896e26ab182a1e940d6bb8b42f77eb48"} Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.086721 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.104701 4821 scope.go:117] "RemoveContainer" containerID="efc640699b2f0c761ecc6cdd93bb7592387a23fd71b7c895e0a87b94aeeb0420" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.142298 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.152234 4821 scope.go:117] "RemoveContainer" containerID="4884242d869fcb68acbcbd8e4bdbb081b5a3c1dd10fa4dca42f46a2290bc70cc" Nov 25 10:52:39 crc kubenswrapper[4821]: E1125 10:52:39.155045 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4884242d869fcb68acbcbd8e4bdbb081b5a3c1dd10fa4dca42f46a2290bc70cc\": container with ID starting with 4884242d869fcb68acbcbd8e4bdbb081b5a3c1dd10fa4dca42f46a2290bc70cc not found: ID does not exist" containerID="4884242d869fcb68acbcbd8e4bdbb081b5a3c1dd10fa4dca42f46a2290bc70cc" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.155097 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4884242d869fcb68acbcbd8e4bdbb081b5a3c1dd10fa4dca42f46a2290bc70cc"} err="failed to get container status \"4884242d869fcb68acbcbd8e4bdbb081b5a3c1dd10fa4dca42f46a2290bc70cc\": rpc error: code = NotFound desc = could not find container \"4884242d869fcb68acbcbd8e4bdbb081b5a3c1dd10fa4dca42f46a2290bc70cc\": container with ID starting with 4884242d869fcb68acbcbd8e4bdbb081b5a3c1dd10fa4dca42f46a2290bc70cc not found: ID does not exist" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.155129 4821 scope.go:117] "RemoveContainer" containerID="efc640699b2f0c761ecc6cdd93bb7592387a23fd71b7c895e0a87b94aeeb0420" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.156638 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 25 10:52:39 crc kubenswrapper[4821]: E1125 10:52:39.159069 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"efc640699b2f0c761ecc6cdd93bb7592387a23fd71b7c895e0a87b94aeeb0420\": container with ID starting with efc640699b2f0c761ecc6cdd93bb7592387a23fd71b7c895e0a87b94aeeb0420 not found: ID does not exist" containerID="efc640699b2f0c761ecc6cdd93bb7592387a23fd71b7c895e0a87b94aeeb0420" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.159114 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"efc640699b2f0c761ecc6cdd93bb7592387a23fd71b7c895e0a87b94aeeb0420"} err="failed to get container status \"efc640699b2f0c761ecc6cdd93bb7592387a23fd71b7c895e0a87b94aeeb0420\": rpc error: code = NotFound desc = could not find container \"efc640699b2f0c761ecc6cdd93bb7592387a23fd71b7c895e0a87b94aeeb0420\": container with ID starting with efc640699b2f0c761ecc6cdd93bb7592387a23fd71b7c895e0a87b94aeeb0420 not found: ID does not exist" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.159141 4821 scope.go:117] "RemoveContainer" containerID="9e2d63189a92dad7e806e26293da984c4e4eba7281abfb9f21111046c6cbe494" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.180490 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.192266 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.210566 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 25 10:52:39 crc kubenswrapper[4821]: E1125 10:52:39.211255 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="802c0cd6-5dd0-41d9-ae10-12cd23aa0467" containerName="nova-metadata-log" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.211277 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="802c0cd6-5dd0-41d9-ae10-12cd23aa0467" containerName="nova-metadata-log" Nov 25 10:52:39 crc kubenswrapper[4821]: E1125 10:52:39.211300 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e305fedf-6c46-4c0e-9f05-ce67a8b2668c" containerName="nova-api-api" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.211309 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="e305fedf-6c46-4c0e-9f05-ce67a8b2668c" containerName="nova-api-api" Nov 25 10:52:39 crc kubenswrapper[4821]: E1125 10:52:39.211329 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e305fedf-6c46-4c0e-9f05-ce67a8b2668c" containerName="nova-api-log" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.211336 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="e305fedf-6c46-4c0e-9f05-ce67a8b2668c" containerName="nova-api-log" Nov 25 10:52:39 crc kubenswrapper[4821]: E1125 10:52:39.211364 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="802c0cd6-5dd0-41d9-ae10-12cd23aa0467" containerName="nova-metadata-metadata" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.211372 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="802c0cd6-5dd0-41d9-ae10-12cd23aa0467" containerName="nova-metadata-metadata" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.211584 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="802c0cd6-5dd0-41d9-ae10-12cd23aa0467" containerName="nova-metadata-log" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.211604 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="e305fedf-6c46-4c0e-9f05-ce67a8b2668c" containerName="nova-api-log" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.211617 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="802c0cd6-5dd0-41d9-ae10-12cd23aa0467" containerName="nova-metadata-metadata" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.211631 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="e305fedf-6c46-4c0e-9f05-ce67a8b2668c" containerName="nova-api-api" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.213375 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.215063 4821 scope.go:117] "RemoveContainer" containerID="078a3f7579d31eb42966404567599c09306bdc5804c0c47e780e7f31e77f7928" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.216568 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.216714 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.216870 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.219901 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.237352 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.237492 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.240146 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.240294 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.241279 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.265865 4821 scope.go:117] "RemoveContainer" containerID="9e2d63189a92dad7e806e26293da984c4e4eba7281abfb9f21111046c6cbe494" Nov 25 10:52:39 crc kubenswrapper[4821]: E1125 10:52:39.266323 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e2d63189a92dad7e806e26293da984c4e4eba7281abfb9f21111046c6cbe494\": container with ID starting with 9e2d63189a92dad7e806e26293da984c4e4eba7281abfb9f21111046c6cbe494 not found: ID does not exist" containerID="9e2d63189a92dad7e806e26293da984c4e4eba7281abfb9f21111046c6cbe494" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.266351 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e2d63189a92dad7e806e26293da984c4e4eba7281abfb9f21111046c6cbe494"} err="failed to get container status \"9e2d63189a92dad7e806e26293da984c4e4eba7281abfb9f21111046c6cbe494\": rpc error: code = NotFound desc = could not find container \"9e2d63189a92dad7e806e26293da984c4e4eba7281abfb9f21111046c6cbe494\": container with ID starting with 9e2d63189a92dad7e806e26293da984c4e4eba7281abfb9f21111046c6cbe494 not found: ID does not exist" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.266371 4821 scope.go:117] "RemoveContainer" containerID="078a3f7579d31eb42966404567599c09306bdc5804c0c47e780e7f31e77f7928" Nov 25 10:52:39 crc kubenswrapper[4821]: E1125 10:52:39.267498 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"078a3f7579d31eb42966404567599c09306bdc5804c0c47e780e7f31e77f7928\": container with ID starting with 078a3f7579d31eb42966404567599c09306bdc5804c0c47e780e7f31e77f7928 not found: ID does not exist" containerID="078a3f7579d31eb42966404567599c09306bdc5804c0c47e780e7f31e77f7928" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.267522 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"078a3f7579d31eb42966404567599c09306bdc5804c0c47e780e7f31e77f7928"} err="failed to get container status \"078a3f7579d31eb42966404567599c09306bdc5804c0c47e780e7f31e77f7928\": rpc error: code = NotFound desc = could not find container \"078a3f7579d31eb42966404567599c09306bdc5804c0c47e780e7f31e77f7928\": container with ID starting with 078a3f7579d31eb42966404567599c09306bdc5804c0c47e780e7f31e77f7928 not found: ID does not exist" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.287346 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2bd244d2-4836-4caf-a6f6-1a00aac5dbb3-public-tls-certs\") pod \"nova-api-0\" (UID: \"2bd244d2-4836-4caf-a6f6-1a00aac5dbb3\") " pod="openstack/nova-api-0" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.287469 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fs856\" (UniqueName: \"kubernetes.io/projected/39dff25e-5059-4eaf-bf98-93aea0eeb258-kube-api-access-fs856\") pod \"nova-metadata-0\" (UID: \"39dff25e-5059-4eaf-bf98-93aea0eeb258\") " pod="openstack/nova-metadata-0" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.287511 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bd244d2-4836-4caf-a6f6-1a00aac5dbb3-config-data\") pod \"nova-api-0\" (UID: \"2bd244d2-4836-4caf-a6f6-1a00aac5dbb3\") " pod="openstack/nova-api-0" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.287537 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39dff25e-5059-4eaf-bf98-93aea0eeb258-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"39dff25e-5059-4eaf-bf98-93aea0eeb258\") " pod="openstack/nova-metadata-0" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.287608 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/39dff25e-5059-4eaf-bf98-93aea0eeb258-logs\") pod \"nova-metadata-0\" (UID: \"39dff25e-5059-4eaf-bf98-93aea0eeb258\") " pod="openstack/nova-metadata-0" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.287686 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8dhdc\" (UniqueName: \"kubernetes.io/projected/2bd244d2-4836-4caf-a6f6-1a00aac5dbb3-kube-api-access-8dhdc\") pod \"nova-api-0\" (UID: \"2bd244d2-4836-4caf-a6f6-1a00aac5dbb3\") " pod="openstack/nova-api-0" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.287729 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/39dff25e-5059-4eaf-bf98-93aea0eeb258-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"39dff25e-5059-4eaf-bf98-93aea0eeb258\") " pod="openstack/nova-metadata-0" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.287783 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2bd244d2-4836-4caf-a6f6-1a00aac5dbb3-logs\") pod \"nova-api-0\" (UID: \"2bd244d2-4836-4caf-a6f6-1a00aac5dbb3\") " pod="openstack/nova-api-0" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.287818 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39dff25e-5059-4eaf-bf98-93aea0eeb258-config-data\") pod \"nova-metadata-0\" (UID: \"39dff25e-5059-4eaf-bf98-93aea0eeb258\") " pod="openstack/nova-metadata-0" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.287837 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bd244d2-4836-4caf-a6f6-1a00aac5dbb3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2bd244d2-4836-4caf-a6f6-1a00aac5dbb3\") " pod="openstack/nova-api-0" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.287855 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2bd244d2-4836-4caf-a6f6-1a00aac5dbb3-internal-tls-certs\") pod \"nova-api-0\" (UID: \"2bd244d2-4836-4caf-a6f6-1a00aac5dbb3\") " pod="openstack/nova-api-0" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.389044 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2bd244d2-4836-4caf-a6f6-1a00aac5dbb3-public-tls-certs\") pod \"nova-api-0\" (UID: \"2bd244d2-4836-4caf-a6f6-1a00aac5dbb3\") " pod="openstack/nova-api-0" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.389254 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fs856\" (UniqueName: \"kubernetes.io/projected/39dff25e-5059-4eaf-bf98-93aea0eeb258-kube-api-access-fs856\") pod \"nova-metadata-0\" (UID: \"39dff25e-5059-4eaf-bf98-93aea0eeb258\") " pod="openstack/nova-metadata-0" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.389281 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bd244d2-4836-4caf-a6f6-1a00aac5dbb3-config-data\") pod \"nova-api-0\" (UID: \"2bd244d2-4836-4caf-a6f6-1a00aac5dbb3\") " pod="openstack/nova-api-0" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.389301 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39dff25e-5059-4eaf-bf98-93aea0eeb258-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"39dff25e-5059-4eaf-bf98-93aea0eeb258\") " pod="openstack/nova-metadata-0" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.389340 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/39dff25e-5059-4eaf-bf98-93aea0eeb258-logs\") pod \"nova-metadata-0\" (UID: \"39dff25e-5059-4eaf-bf98-93aea0eeb258\") " pod="openstack/nova-metadata-0" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.389364 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8dhdc\" (UniqueName: \"kubernetes.io/projected/2bd244d2-4836-4caf-a6f6-1a00aac5dbb3-kube-api-access-8dhdc\") pod \"nova-api-0\" (UID: \"2bd244d2-4836-4caf-a6f6-1a00aac5dbb3\") " pod="openstack/nova-api-0" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.389397 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/39dff25e-5059-4eaf-bf98-93aea0eeb258-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"39dff25e-5059-4eaf-bf98-93aea0eeb258\") " pod="openstack/nova-metadata-0" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.389430 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2bd244d2-4836-4caf-a6f6-1a00aac5dbb3-logs\") pod \"nova-api-0\" (UID: \"2bd244d2-4836-4caf-a6f6-1a00aac5dbb3\") " pod="openstack/nova-api-0" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.389452 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39dff25e-5059-4eaf-bf98-93aea0eeb258-config-data\") pod \"nova-metadata-0\" (UID: \"39dff25e-5059-4eaf-bf98-93aea0eeb258\") " pod="openstack/nova-metadata-0" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.389476 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bd244d2-4836-4caf-a6f6-1a00aac5dbb3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2bd244d2-4836-4caf-a6f6-1a00aac5dbb3\") " pod="openstack/nova-api-0" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.389498 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2bd244d2-4836-4caf-a6f6-1a00aac5dbb3-internal-tls-certs\") pod \"nova-api-0\" (UID: \"2bd244d2-4836-4caf-a6f6-1a00aac5dbb3\") " pod="openstack/nova-api-0" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.391983 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/39dff25e-5059-4eaf-bf98-93aea0eeb258-logs\") pod \"nova-metadata-0\" (UID: \"39dff25e-5059-4eaf-bf98-93aea0eeb258\") " pod="openstack/nova-metadata-0" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.394543 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2bd244d2-4836-4caf-a6f6-1a00aac5dbb3-public-tls-certs\") pod \"nova-api-0\" (UID: \"2bd244d2-4836-4caf-a6f6-1a00aac5dbb3\") " pod="openstack/nova-api-0" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.394588 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2bd244d2-4836-4caf-a6f6-1a00aac5dbb3-logs\") pod \"nova-api-0\" (UID: \"2bd244d2-4836-4caf-a6f6-1a00aac5dbb3\") " pod="openstack/nova-api-0" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.394907 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bd244d2-4836-4caf-a6f6-1a00aac5dbb3-config-data\") pod \"nova-api-0\" (UID: \"2bd244d2-4836-4caf-a6f6-1a00aac5dbb3\") " pod="openstack/nova-api-0" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.395914 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/39dff25e-5059-4eaf-bf98-93aea0eeb258-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"39dff25e-5059-4eaf-bf98-93aea0eeb258\") " pod="openstack/nova-metadata-0" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.396011 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39dff25e-5059-4eaf-bf98-93aea0eeb258-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"39dff25e-5059-4eaf-bf98-93aea0eeb258\") " pod="openstack/nova-metadata-0" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.396592 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39dff25e-5059-4eaf-bf98-93aea0eeb258-config-data\") pod \"nova-metadata-0\" (UID: \"39dff25e-5059-4eaf-bf98-93aea0eeb258\") " pod="openstack/nova-metadata-0" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.401715 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bd244d2-4836-4caf-a6f6-1a00aac5dbb3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2bd244d2-4836-4caf-a6f6-1a00aac5dbb3\") " pod="openstack/nova-api-0" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.402278 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2bd244d2-4836-4caf-a6f6-1a00aac5dbb3-internal-tls-certs\") pod \"nova-api-0\" (UID: \"2bd244d2-4836-4caf-a6f6-1a00aac5dbb3\") " pod="openstack/nova-api-0" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.408840 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8dhdc\" (UniqueName: \"kubernetes.io/projected/2bd244d2-4836-4caf-a6f6-1a00aac5dbb3-kube-api-access-8dhdc\") pod \"nova-api-0\" (UID: \"2bd244d2-4836-4caf-a6f6-1a00aac5dbb3\") " pod="openstack/nova-api-0" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.410952 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fs856\" (UniqueName: \"kubernetes.io/projected/39dff25e-5059-4eaf-bf98-93aea0eeb258-kube-api-access-fs856\") pod \"nova-metadata-0\" (UID: \"39dff25e-5059-4eaf-bf98-93aea0eeb258\") " pod="openstack/nova-metadata-0" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.560954 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 10:52:39 crc kubenswrapper[4821]: I1125 10:52:39.572772 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 10:52:40 crc kubenswrapper[4821]: I1125 10:52:40.051207 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 25 10:52:40 crc kubenswrapper[4821]: W1125 10:52:40.073095 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2bd244d2_4836_4caf_a6f6_1a00aac5dbb3.slice/crio-569d4bea2cf8015cde88b8a9d08dfbb343d55d49a4c670cd04b35b4992631dcc WatchSource:0}: Error finding container 569d4bea2cf8015cde88b8a9d08dfbb343d55d49a4c670cd04b35b4992631dcc: Status 404 returned error can't find the container with id 569d4bea2cf8015cde88b8a9d08dfbb343d55d49a4c670cd04b35b4992631dcc Nov 25 10:52:40 crc kubenswrapper[4821]: I1125 10:52:40.107923 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"3bdff849-753c-42b5-9e06-2e37dfe42666","Type":"ContainerStarted","Data":"9d5d9c48d8755bb526707c1451997cc4cbfa47938f4cb2231d6ed6a8da6903c4"} Nov 25 10:52:40 crc kubenswrapper[4821]: W1125 10:52:40.142551 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod39dff25e_5059_4eaf_bf98_93aea0eeb258.slice/crio-ab4d4674825816985de39db50fccae51336e74f99498e4b77477aa0cd88a160c WatchSource:0}: Error finding container ab4d4674825816985de39db50fccae51336e74f99498e4b77477aa0cd88a160c: Status 404 returned error can't find the container with id ab4d4674825816985de39db50fccae51336e74f99498e4b77477aa0cd88a160c Nov 25 10:52:40 crc kubenswrapper[4821]: I1125 10:52:40.144529 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="802c0cd6-5dd0-41d9-ae10-12cd23aa0467" path="/var/lib/kubelet/pods/802c0cd6-5dd0-41d9-ae10-12cd23aa0467/volumes" Nov 25 10:52:40 crc kubenswrapper[4821]: I1125 10:52:40.149141 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c41109a7-e905-4087-9d04-ccd2cccb42b5" path="/var/lib/kubelet/pods/c41109a7-e905-4087-9d04-ccd2cccb42b5/volumes" Nov 25 10:52:40 crc kubenswrapper[4821]: I1125 10:52:40.151548 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.151519999 podStartE2EDuration="2.151519999s" podCreationTimestamp="2025-11-25 10:52:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:52:40.131974265 +0000 UTC m=+1230.668294132" watchObservedRunningTime="2025-11-25 10:52:40.151519999 +0000 UTC m=+1230.687839856" Nov 25 10:52:40 crc kubenswrapper[4821]: I1125 10:52:40.152496 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e305fedf-6c46-4c0e-9f05-ce67a8b2668c" path="/var/lib/kubelet/pods/e305fedf-6c46-4c0e-9f05-ce67a8b2668c/volumes" Nov 25 10:52:40 crc kubenswrapper[4821]: I1125 10:52:40.153336 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2bd244d2-4836-4caf-a6f6-1a00aac5dbb3","Type":"ContainerStarted","Data":"569d4bea2cf8015cde88b8a9d08dfbb343d55d49a4c670cd04b35b4992631dcc"} Nov 25 10:52:40 crc kubenswrapper[4821]: I1125 10:52:40.153476 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 10:52:41 crc kubenswrapper[4821]: I1125 10:52:41.131772 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2bd244d2-4836-4caf-a6f6-1a00aac5dbb3","Type":"ContainerStarted","Data":"9e0c398913a4f9c133c4df0056522b47fe960e0c61558a337cc66cf3d21e940b"} Nov 25 10:52:41 crc kubenswrapper[4821]: I1125 10:52:41.132366 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2bd244d2-4836-4caf-a6f6-1a00aac5dbb3","Type":"ContainerStarted","Data":"d768b36ca02c89cff2012c42186203cbde073816a104bf8cf000f786f08cd528"} Nov 25 10:52:41 crc kubenswrapper[4821]: I1125 10:52:41.134778 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"39dff25e-5059-4eaf-bf98-93aea0eeb258","Type":"ContainerStarted","Data":"cc998fa4c1e5d9af79a52a08a7a64056f2d83b7490bb123ad0fcbd0af6ee0c39"} Nov 25 10:52:41 crc kubenswrapper[4821]: I1125 10:52:41.134852 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"39dff25e-5059-4eaf-bf98-93aea0eeb258","Type":"ContainerStarted","Data":"dde68d37cc225a66466931ced3d3b4d3eaec9e1de1a7f0870209b06843412d55"} Nov 25 10:52:41 crc kubenswrapper[4821]: I1125 10:52:41.134868 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"39dff25e-5059-4eaf-bf98-93aea0eeb258","Type":"ContainerStarted","Data":"ab4d4674825816985de39db50fccae51336e74f99498e4b77477aa0cd88a160c"} Nov 25 10:52:41 crc kubenswrapper[4821]: I1125 10:52:41.161695 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.161673271 podStartE2EDuration="2.161673271s" podCreationTimestamp="2025-11-25 10:52:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:52:41.150633781 +0000 UTC m=+1231.686953648" watchObservedRunningTime="2025-11-25 10:52:41.161673271 +0000 UTC m=+1231.697993118" Nov 25 10:52:41 crc kubenswrapper[4821]: I1125 10:52:41.172403 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.172379412 podStartE2EDuration="2.172379412s" podCreationTimestamp="2025-11-25 10:52:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:52:41.168151941 +0000 UTC m=+1231.704471808" watchObservedRunningTime="2025-11-25 10:52:41.172379412 +0000 UTC m=+1231.708699259" Nov 25 10:52:43 crc kubenswrapper[4821]: I1125 10:52:43.468075 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 25 10:52:44 crc kubenswrapper[4821]: I1125 10:52:44.573361 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 25 10:52:44 crc kubenswrapper[4821]: I1125 10:52:44.573416 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 25 10:52:48 crc kubenswrapper[4821]: I1125 10:52:48.468497 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 25 10:52:48 crc kubenswrapper[4821]: I1125 10:52:48.500892 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 25 10:52:49 crc kubenswrapper[4821]: I1125 10:52:49.234498 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 25 10:52:49 crc kubenswrapper[4821]: I1125 10:52:49.561527 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 25 10:52:49 crc kubenswrapper[4821]: I1125 10:52:49.563114 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 25 10:52:49 crc kubenswrapper[4821]: I1125 10:52:49.573522 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 25 10:52:49 crc kubenswrapper[4821]: I1125 10:52:49.573561 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 25 10:52:50 crc kubenswrapper[4821]: I1125 10:52:50.572346 4821 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="2bd244d2-4836-4caf-a6f6-1a00aac5dbb3" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.204:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 25 10:52:50 crc kubenswrapper[4821]: I1125 10:52:50.572366 4821 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="2bd244d2-4836-4caf-a6f6-1a00aac5dbb3" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.204:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 25 10:52:50 crc kubenswrapper[4821]: I1125 10:52:50.587335 4821 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="39dff25e-5059-4eaf-bf98-93aea0eeb258" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.205:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 25 10:52:50 crc kubenswrapper[4821]: I1125 10:52:50.587368 4821 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="39dff25e-5059-4eaf-bf98-93aea0eeb258" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.205:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 25 10:52:51 crc kubenswrapper[4821]: I1125 10:52:51.502858 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 25 10:52:59 crc kubenswrapper[4821]: I1125 10:52:59.567747 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 25 10:52:59 crc kubenswrapper[4821]: I1125 10:52:59.568338 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 25 10:52:59 crc kubenswrapper[4821]: I1125 10:52:59.569146 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 25 10:52:59 crc kubenswrapper[4821]: I1125 10:52:59.569205 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 25 10:52:59 crc kubenswrapper[4821]: I1125 10:52:59.577680 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 25 10:52:59 crc kubenswrapper[4821]: I1125 10:52:59.580803 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 25 10:52:59 crc kubenswrapper[4821]: I1125 10:52:59.588450 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 25 10:52:59 crc kubenswrapper[4821]: I1125 10:52:59.589506 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 25 10:52:59 crc kubenswrapper[4821]: I1125 10:52:59.594641 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 25 10:53:00 crc kubenswrapper[4821]: I1125 10:53:00.321568 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 25 10:53:08 crc kubenswrapper[4821]: I1125 10:53:08.755861 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 25 10:53:09 crc kubenswrapper[4821]: I1125 10:53:09.789445 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 25 10:53:13 crc kubenswrapper[4821]: I1125 10:53:13.235819 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="2bbb754a-fff3-4ba2-ab75-68e48b626ab0" containerName="rabbitmq" containerID="cri-o://60497933ff5037a6512cb5d8a0394d48cdda3bb66db9431a0698858ee9f8eb43" gracePeriod=604796 Nov 25 10:53:13 crc kubenswrapper[4821]: I1125 10:53:13.934248 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="330a479e-3d76-4de3-bc84-1d7c2d65d265" containerName="rabbitmq" containerID="cri-o://7848d9adca99a2cef746a81eaafad2bfec909b4451ebcf92763c11f7448b1d15" gracePeriod=604796 Nov 25 10:53:19 crc kubenswrapper[4821]: I1125 10:53:19.471568 4821 generic.go:334] "Generic (PLEG): container finished" podID="2bbb754a-fff3-4ba2-ab75-68e48b626ab0" containerID="60497933ff5037a6512cb5d8a0394d48cdda3bb66db9431a0698858ee9f8eb43" exitCode=0 Nov 25 10:53:19 crc kubenswrapper[4821]: I1125 10:53:19.471665 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"2bbb754a-fff3-4ba2-ab75-68e48b626ab0","Type":"ContainerDied","Data":"60497933ff5037a6512cb5d8a0394d48cdda3bb66db9431a0698858ee9f8eb43"} Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.021778 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.169802 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2bbb754a-fff3-4ba2-ab75-68e48b626ab0-rabbitmq-confd\") pod \"2bbb754a-fff3-4ba2-ab75-68e48b626ab0\" (UID: \"2bbb754a-fff3-4ba2-ab75-68e48b626ab0\") " Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.169866 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2bbb754a-fff3-4ba2-ab75-68e48b626ab0-rabbitmq-plugins\") pod \"2bbb754a-fff3-4ba2-ab75-68e48b626ab0\" (UID: \"2bbb754a-fff3-4ba2-ab75-68e48b626ab0\") " Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.169883 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2bbb754a-fff3-4ba2-ab75-68e48b626ab0-server-conf\") pod \"2bbb754a-fff3-4ba2-ab75-68e48b626ab0\" (UID: \"2bbb754a-fff3-4ba2-ab75-68e48b626ab0\") " Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.169922 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2bbb754a-fff3-4ba2-ab75-68e48b626ab0-rabbitmq-tls\") pod \"2bbb754a-fff3-4ba2-ab75-68e48b626ab0\" (UID: \"2bbb754a-fff3-4ba2-ab75-68e48b626ab0\") " Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.169961 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2bbb754a-fff3-4ba2-ab75-68e48b626ab0-pod-info\") pod \"2bbb754a-fff3-4ba2-ab75-68e48b626ab0\" (UID: \"2bbb754a-fff3-4ba2-ab75-68e48b626ab0\") " Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.170001 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2bbb754a-fff3-4ba2-ab75-68e48b626ab0-rabbitmq-erlang-cookie\") pod \"2bbb754a-fff3-4ba2-ab75-68e48b626ab0\" (UID: \"2bbb754a-fff3-4ba2-ab75-68e48b626ab0\") " Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.170056 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2bbb754a-fff3-4ba2-ab75-68e48b626ab0-plugins-conf\") pod \"2bbb754a-fff3-4ba2-ab75-68e48b626ab0\" (UID: \"2bbb754a-fff3-4ba2-ab75-68e48b626ab0\") " Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.170070 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"2bbb754a-fff3-4ba2-ab75-68e48b626ab0\" (UID: \"2bbb754a-fff3-4ba2-ab75-68e48b626ab0\") " Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.170095 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gpq9f\" (UniqueName: \"kubernetes.io/projected/2bbb754a-fff3-4ba2-ab75-68e48b626ab0-kube-api-access-gpq9f\") pod \"2bbb754a-fff3-4ba2-ab75-68e48b626ab0\" (UID: \"2bbb754a-fff3-4ba2-ab75-68e48b626ab0\") " Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.170178 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2bbb754a-fff3-4ba2-ab75-68e48b626ab0-config-data\") pod \"2bbb754a-fff3-4ba2-ab75-68e48b626ab0\" (UID: \"2bbb754a-fff3-4ba2-ab75-68e48b626ab0\") " Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.170227 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2bbb754a-fff3-4ba2-ab75-68e48b626ab0-erlang-cookie-secret\") pod \"2bbb754a-fff3-4ba2-ab75-68e48b626ab0\" (UID: \"2bbb754a-fff3-4ba2-ab75-68e48b626ab0\") " Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.171059 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2bbb754a-fff3-4ba2-ab75-68e48b626ab0-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "2bbb754a-fff3-4ba2-ab75-68e48b626ab0" (UID: "2bbb754a-fff3-4ba2-ab75-68e48b626ab0"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.171143 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2bbb754a-fff3-4ba2-ab75-68e48b626ab0-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "2bbb754a-fff3-4ba2-ab75-68e48b626ab0" (UID: "2bbb754a-fff3-4ba2-ab75-68e48b626ab0"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.171239 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2bbb754a-fff3-4ba2-ab75-68e48b626ab0-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "2bbb754a-fff3-4ba2-ab75-68e48b626ab0" (UID: "2bbb754a-fff3-4ba2-ab75-68e48b626ab0"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.176357 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "persistence") pod "2bbb754a-fff3-4ba2-ab75-68e48b626ab0" (UID: "2bbb754a-fff3-4ba2-ab75-68e48b626ab0"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.176526 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2bbb754a-fff3-4ba2-ab75-68e48b626ab0-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "2bbb754a-fff3-4ba2-ab75-68e48b626ab0" (UID: "2bbb754a-fff3-4ba2-ab75-68e48b626ab0"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.177363 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/2bbb754a-fff3-4ba2-ab75-68e48b626ab0-pod-info" (OuterVolumeSpecName: "pod-info") pod "2bbb754a-fff3-4ba2-ab75-68e48b626ab0" (UID: "2bbb754a-fff3-4ba2-ab75-68e48b626ab0"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.180944 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2bbb754a-fff3-4ba2-ab75-68e48b626ab0-kube-api-access-gpq9f" (OuterVolumeSpecName: "kube-api-access-gpq9f") pod "2bbb754a-fff3-4ba2-ab75-68e48b626ab0" (UID: "2bbb754a-fff3-4ba2-ab75-68e48b626ab0"). InnerVolumeSpecName "kube-api-access-gpq9f". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.184765 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bbb754a-fff3-4ba2-ab75-68e48b626ab0-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "2bbb754a-fff3-4ba2-ab75-68e48b626ab0" (UID: "2bbb754a-fff3-4ba2-ab75-68e48b626ab0"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.221497 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2bbb754a-fff3-4ba2-ab75-68e48b626ab0-config-data" (OuterVolumeSpecName: "config-data") pod "2bbb754a-fff3-4ba2-ab75-68e48b626ab0" (UID: "2bbb754a-fff3-4ba2-ab75-68e48b626ab0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.238052 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2bbb754a-fff3-4ba2-ab75-68e48b626ab0-server-conf" (OuterVolumeSpecName: "server-conf") pod "2bbb754a-fff3-4ba2-ab75-68e48b626ab0" (UID: "2bbb754a-fff3-4ba2-ab75-68e48b626ab0"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.272055 4821 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2bbb754a-fff3-4ba2-ab75-68e48b626ab0-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.272084 4821 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2bbb754a-fff3-4ba2-ab75-68e48b626ab0-server-conf\") on node \"crc\" DevicePath \"\"" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.272093 4821 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2bbb754a-fff3-4ba2-ab75-68e48b626ab0-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.272101 4821 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2bbb754a-fff3-4ba2-ab75-68e48b626ab0-pod-info\") on node \"crc\" DevicePath \"\"" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.272109 4821 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2bbb754a-fff3-4ba2-ab75-68e48b626ab0-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.272119 4821 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2bbb754a-fff3-4ba2-ab75-68e48b626ab0-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.272135 4821 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.272144 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gpq9f\" (UniqueName: \"kubernetes.io/projected/2bbb754a-fff3-4ba2-ab75-68e48b626ab0-kube-api-access-gpq9f\") on node \"crc\" DevicePath \"\"" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.272153 4821 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2bbb754a-fff3-4ba2-ab75-68e48b626ab0-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.272176 4821 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2bbb754a-fff3-4ba2-ab75-68e48b626ab0-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.294442 4821 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.306379 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2bbb754a-fff3-4ba2-ab75-68e48b626ab0-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "2bbb754a-fff3-4ba2-ab75-68e48b626ab0" (UID: "2bbb754a-fff3-4ba2-ab75-68e48b626ab0"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.374045 4821 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2bbb754a-fff3-4ba2-ab75-68e48b626ab0-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.374116 4821 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.493270 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"2bbb754a-fff3-4ba2-ab75-68e48b626ab0","Type":"ContainerDied","Data":"e736050aac0d58ca110e71f33ba6e70d441ed7f3a727bd519a0b4115de93d021"} Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.493317 4821 scope.go:117] "RemoveContainer" containerID="60497933ff5037a6512cb5d8a0394d48cdda3bb66db9431a0698858ee9f8eb43" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.493438 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.509074 4821 generic.go:334] "Generic (PLEG): container finished" podID="330a479e-3d76-4de3-bc84-1d7c2d65d265" containerID="7848d9adca99a2cef746a81eaafad2bfec909b4451ebcf92763c11f7448b1d15" exitCode=0 Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.509126 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"330a479e-3d76-4de3-bc84-1d7c2d65d265","Type":"ContainerDied","Data":"7848d9adca99a2cef746a81eaafad2bfec909b4451ebcf92763c11f7448b1d15"} Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.546928 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.547333 4821 scope.go:117] "RemoveContainer" containerID="ef45f64d359b5165b93856160cad982dc3eb70d24d3b0aed64df4b49a699da2b" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.563952 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.574365 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Nov 25 10:53:20 crc kubenswrapper[4821]: E1125 10:53:20.575309 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bbb754a-fff3-4ba2-ab75-68e48b626ab0" containerName="rabbitmq" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.575334 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bbb754a-fff3-4ba2-ab75-68e48b626ab0" containerName="rabbitmq" Nov 25 10:53:20 crc kubenswrapper[4821]: E1125 10:53:20.575358 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bbb754a-fff3-4ba2-ab75-68e48b626ab0" containerName="setup-container" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.575366 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bbb754a-fff3-4ba2-ab75-68e48b626ab0" containerName="setup-container" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.575582 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bbb754a-fff3-4ba2-ab75-68e48b626ab0" containerName="rabbitmq" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.576886 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.583204 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.583633 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.584055 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-fwctx" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.584355 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.585526 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.585708 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.586086 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.588459 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.683536 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/50c27d16-2f67-4f90-83f0-83e16161de50-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"50c27d16-2f67-4f90-83f0-83e16161de50\") " pod="openstack/rabbitmq-server-0" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.683710 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/50c27d16-2f67-4f90-83f0-83e16161de50-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"50c27d16-2f67-4f90-83f0-83e16161de50\") " pod="openstack/rabbitmq-server-0" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.683749 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/50c27d16-2f67-4f90-83f0-83e16161de50-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"50c27d16-2f67-4f90-83f0-83e16161de50\") " pod="openstack/rabbitmq-server-0" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.683773 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/50c27d16-2f67-4f90-83f0-83e16161de50-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"50c27d16-2f67-4f90-83f0-83e16161de50\") " pod="openstack/rabbitmq-server-0" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.683793 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/50c27d16-2f67-4f90-83f0-83e16161de50-server-conf\") pod \"rabbitmq-server-0\" (UID: \"50c27d16-2f67-4f90-83f0-83e16161de50\") " pod="openstack/rabbitmq-server-0" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.683812 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"50c27d16-2f67-4f90-83f0-83e16161de50\") " pod="openstack/rabbitmq-server-0" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.683852 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/50c27d16-2f67-4f90-83f0-83e16161de50-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"50c27d16-2f67-4f90-83f0-83e16161de50\") " pod="openstack/rabbitmq-server-0" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.683872 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/50c27d16-2f67-4f90-83f0-83e16161de50-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"50c27d16-2f67-4f90-83f0-83e16161de50\") " pod="openstack/rabbitmq-server-0" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.683894 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/50c27d16-2f67-4f90-83f0-83e16161de50-config-data\") pod \"rabbitmq-server-0\" (UID: \"50c27d16-2f67-4f90-83f0-83e16161de50\") " pod="openstack/rabbitmq-server-0" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.683921 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/50c27d16-2f67-4f90-83f0-83e16161de50-pod-info\") pod \"rabbitmq-server-0\" (UID: \"50c27d16-2f67-4f90-83f0-83e16161de50\") " pod="openstack/rabbitmq-server-0" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.683953 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2qd7\" (UniqueName: \"kubernetes.io/projected/50c27d16-2f67-4f90-83f0-83e16161de50-kube-api-access-j2qd7\") pod \"rabbitmq-server-0\" (UID: \"50c27d16-2f67-4f90-83f0-83e16161de50\") " pod="openstack/rabbitmq-server-0" Nov 25 10:53:20 crc kubenswrapper[4821]: E1125 10:53:20.716890 4821 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2bbb754a_fff3_4ba2_ab75_68e48b626ab0.slice\": RecentStats: unable to find data in memory cache]" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.721712 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.786669 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/50c27d16-2f67-4f90-83f0-83e16161de50-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"50c27d16-2f67-4f90-83f0-83e16161de50\") " pod="openstack/rabbitmq-server-0" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.786740 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/50c27d16-2f67-4f90-83f0-83e16161de50-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"50c27d16-2f67-4f90-83f0-83e16161de50\") " pod="openstack/rabbitmq-server-0" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.786771 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/50c27d16-2f67-4f90-83f0-83e16161de50-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"50c27d16-2f67-4f90-83f0-83e16161de50\") " pod="openstack/rabbitmq-server-0" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.786798 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/50c27d16-2f67-4f90-83f0-83e16161de50-server-conf\") pod \"rabbitmq-server-0\" (UID: \"50c27d16-2f67-4f90-83f0-83e16161de50\") " pod="openstack/rabbitmq-server-0" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.786824 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"50c27d16-2f67-4f90-83f0-83e16161de50\") " pod="openstack/rabbitmq-server-0" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.786881 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/50c27d16-2f67-4f90-83f0-83e16161de50-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"50c27d16-2f67-4f90-83f0-83e16161de50\") " pod="openstack/rabbitmq-server-0" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.786906 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/50c27d16-2f67-4f90-83f0-83e16161de50-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"50c27d16-2f67-4f90-83f0-83e16161de50\") " pod="openstack/rabbitmq-server-0" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.786928 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/50c27d16-2f67-4f90-83f0-83e16161de50-config-data\") pod \"rabbitmq-server-0\" (UID: \"50c27d16-2f67-4f90-83f0-83e16161de50\") " pod="openstack/rabbitmq-server-0" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.786965 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/50c27d16-2f67-4f90-83f0-83e16161de50-pod-info\") pod \"rabbitmq-server-0\" (UID: \"50c27d16-2f67-4f90-83f0-83e16161de50\") " pod="openstack/rabbitmq-server-0" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.787003 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2qd7\" (UniqueName: \"kubernetes.io/projected/50c27d16-2f67-4f90-83f0-83e16161de50-kube-api-access-j2qd7\") pod \"rabbitmq-server-0\" (UID: \"50c27d16-2f67-4f90-83f0-83e16161de50\") " pod="openstack/rabbitmq-server-0" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.787085 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/50c27d16-2f67-4f90-83f0-83e16161de50-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"50c27d16-2f67-4f90-83f0-83e16161de50\") " pod="openstack/rabbitmq-server-0" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.787730 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/50c27d16-2f67-4f90-83f0-83e16161de50-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"50c27d16-2f67-4f90-83f0-83e16161de50\") " pod="openstack/rabbitmq-server-0" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.787841 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/50c27d16-2f67-4f90-83f0-83e16161de50-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"50c27d16-2f67-4f90-83f0-83e16161de50\") " pod="openstack/rabbitmq-server-0" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.788110 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/50c27d16-2f67-4f90-83f0-83e16161de50-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"50c27d16-2f67-4f90-83f0-83e16161de50\") " pod="openstack/rabbitmq-server-0" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.788120 4821 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"50c27d16-2f67-4f90-83f0-83e16161de50\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/rabbitmq-server-0" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.788664 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/50c27d16-2f67-4f90-83f0-83e16161de50-server-conf\") pod \"rabbitmq-server-0\" (UID: \"50c27d16-2f67-4f90-83f0-83e16161de50\") " pod="openstack/rabbitmq-server-0" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.789582 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/50c27d16-2f67-4f90-83f0-83e16161de50-config-data\") pod \"rabbitmq-server-0\" (UID: \"50c27d16-2f67-4f90-83f0-83e16161de50\") " pod="openstack/rabbitmq-server-0" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.806318 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/50c27d16-2f67-4f90-83f0-83e16161de50-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"50c27d16-2f67-4f90-83f0-83e16161de50\") " pod="openstack/rabbitmq-server-0" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.818704 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/50c27d16-2f67-4f90-83f0-83e16161de50-pod-info\") pod \"rabbitmq-server-0\" (UID: \"50c27d16-2f67-4f90-83f0-83e16161de50\") " pod="openstack/rabbitmq-server-0" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.819855 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/50c27d16-2f67-4f90-83f0-83e16161de50-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"50c27d16-2f67-4f90-83f0-83e16161de50\") " pod="openstack/rabbitmq-server-0" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.829675 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"50c27d16-2f67-4f90-83f0-83e16161de50\") " pod="openstack/rabbitmq-server-0" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.836578 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2qd7\" (UniqueName: \"kubernetes.io/projected/50c27d16-2f67-4f90-83f0-83e16161de50-kube-api-access-j2qd7\") pod \"rabbitmq-server-0\" (UID: \"50c27d16-2f67-4f90-83f0-83e16161de50\") " pod="openstack/rabbitmq-server-0" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.838516 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/50c27d16-2f67-4f90-83f0-83e16161de50-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"50c27d16-2f67-4f90-83f0-83e16161de50\") " pod="openstack/rabbitmq-server-0" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.888500 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/330a479e-3d76-4de3-bc84-1d7c2d65d265-erlang-cookie-secret\") pod \"330a479e-3d76-4de3-bc84-1d7c2d65d265\" (UID: \"330a479e-3d76-4de3-bc84-1d7c2d65d265\") " Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.888588 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/330a479e-3d76-4de3-bc84-1d7c2d65d265-rabbitmq-erlang-cookie\") pod \"330a479e-3d76-4de3-bc84-1d7c2d65d265\" (UID: \"330a479e-3d76-4de3-bc84-1d7c2d65d265\") " Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.888621 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t4tp7\" (UniqueName: \"kubernetes.io/projected/330a479e-3d76-4de3-bc84-1d7c2d65d265-kube-api-access-t4tp7\") pod \"330a479e-3d76-4de3-bc84-1d7c2d65d265\" (UID: \"330a479e-3d76-4de3-bc84-1d7c2d65d265\") " Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.888648 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/330a479e-3d76-4de3-bc84-1d7c2d65d265-rabbitmq-confd\") pod \"330a479e-3d76-4de3-bc84-1d7c2d65d265\" (UID: \"330a479e-3d76-4de3-bc84-1d7c2d65d265\") " Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.888673 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/330a479e-3d76-4de3-bc84-1d7c2d65d265-rabbitmq-plugins\") pod \"330a479e-3d76-4de3-bc84-1d7c2d65d265\" (UID: \"330a479e-3d76-4de3-bc84-1d7c2d65d265\") " Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.888741 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/330a479e-3d76-4de3-bc84-1d7c2d65d265-plugins-conf\") pod \"330a479e-3d76-4de3-bc84-1d7c2d65d265\" (UID: \"330a479e-3d76-4de3-bc84-1d7c2d65d265\") " Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.888786 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"330a479e-3d76-4de3-bc84-1d7c2d65d265\" (UID: \"330a479e-3d76-4de3-bc84-1d7c2d65d265\") " Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.888830 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/330a479e-3d76-4de3-bc84-1d7c2d65d265-rabbitmq-tls\") pod \"330a479e-3d76-4de3-bc84-1d7c2d65d265\" (UID: \"330a479e-3d76-4de3-bc84-1d7c2d65d265\") " Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.888872 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/330a479e-3d76-4de3-bc84-1d7c2d65d265-server-conf\") pod \"330a479e-3d76-4de3-bc84-1d7c2d65d265\" (UID: \"330a479e-3d76-4de3-bc84-1d7c2d65d265\") " Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.888925 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/330a479e-3d76-4de3-bc84-1d7c2d65d265-pod-info\") pod \"330a479e-3d76-4de3-bc84-1d7c2d65d265\" (UID: \"330a479e-3d76-4de3-bc84-1d7c2d65d265\") " Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.888988 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/330a479e-3d76-4de3-bc84-1d7c2d65d265-config-data\") pod \"330a479e-3d76-4de3-bc84-1d7c2d65d265\" (UID: \"330a479e-3d76-4de3-bc84-1d7c2d65d265\") " Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.889146 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/330a479e-3d76-4de3-bc84-1d7c2d65d265-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "330a479e-3d76-4de3-bc84-1d7c2d65d265" (UID: "330a479e-3d76-4de3-bc84-1d7c2d65d265"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.889249 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/330a479e-3d76-4de3-bc84-1d7c2d65d265-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "330a479e-3d76-4de3-bc84-1d7c2d65d265" (UID: "330a479e-3d76-4de3-bc84-1d7c2d65d265"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.889555 4821 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/330a479e-3d76-4de3-bc84-1d7c2d65d265-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.889581 4821 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/330a479e-3d76-4de3-bc84-1d7c2d65d265-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.891661 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/330a479e-3d76-4de3-bc84-1d7c2d65d265-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "330a479e-3d76-4de3-bc84-1d7c2d65d265" (UID: "330a479e-3d76-4de3-bc84-1d7c2d65d265"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.892592 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "persistence") pod "330a479e-3d76-4de3-bc84-1d7c2d65d265" (UID: "330a479e-3d76-4de3-bc84-1d7c2d65d265"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.892600 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/330a479e-3d76-4de3-bc84-1d7c2d65d265-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "330a479e-3d76-4de3-bc84-1d7c2d65d265" (UID: "330a479e-3d76-4de3-bc84-1d7c2d65d265"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.893231 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/330a479e-3d76-4de3-bc84-1d7c2d65d265-kube-api-access-t4tp7" (OuterVolumeSpecName: "kube-api-access-t4tp7") pod "330a479e-3d76-4de3-bc84-1d7c2d65d265" (UID: "330a479e-3d76-4de3-bc84-1d7c2d65d265"). InnerVolumeSpecName "kube-api-access-t4tp7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.893698 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/330a479e-3d76-4de3-bc84-1d7c2d65d265-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "330a479e-3d76-4de3-bc84-1d7c2d65d265" (UID: "330a479e-3d76-4de3-bc84-1d7c2d65d265"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.894391 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/330a479e-3d76-4de3-bc84-1d7c2d65d265-pod-info" (OuterVolumeSpecName: "pod-info") pod "330a479e-3d76-4de3-bc84-1d7c2d65d265" (UID: "330a479e-3d76-4de3-bc84-1d7c2d65d265"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.915440 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/330a479e-3d76-4de3-bc84-1d7c2d65d265-config-data" (OuterVolumeSpecName: "config-data") pod "330a479e-3d76-4de3-bc84-1d7c2d65d265" (UID: "330a479e-3d76-4de3-bc84-1d7c2d65d265"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.948067 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/330a479e-3d76-4de3-bc84-1d7c2d65d265-server-conf" (OuterVolumeSpecName: "server-conf") pod "330a479e-3d76-4de3-bc84-1d7c2d65d265" (UID: "330a479e-3d76-4de3-bc84-1d7c2d65d265"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.991426 4821 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/330a479e-3d76-4de3-bc84-1d7c2d65d265-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.991460 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t4tp7\" (UniqueName: \"kubernetes.io/projected/330a479e-3d76-4de3-bc84-1d7c2d65d265-kube-api-access-t4tp7\") on node \"crc\" DevicePath \"\"" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.991476 4821 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/330a479e-3d76-4de3-bc84-1d7c2d65d265-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.991506 4821 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.991518 4821 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/330a479e-3d76-4de3-bc84-1d7c2d65d265-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.991528 4821 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/330a479e-3d76-4de3-bc84-1d7c2d65d265-server-conf\") on node \"crc\" DevicePath \"\"" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.991537 4821 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/330a479e-3d76-4de3-bc84-1d7c2d65d265-pod-info\") on node \"crc\" DevicePath \"\"" Nov 25 10:53:20 crc kubenswrapper[4821]: I1125 10:53:20.991546 4821 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/330a479e-3d76-4de3-bc84-1d7c2d65d265-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:53:21 crc kubenswrapper[4821]: I1125 10:53:21.026690 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 25 10:53:21 crc kubenswrapper[4821]: I1125 10:53:21.041060 4821 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Nov 25 10:53:21 crc kubenswrapper[4821]: I1125 10:53:21.041422 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/330a479e-3d76-4de3-bc84-1d7c2d65d265-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "330a479e-3d76-4de3-bc84-1d7c2d65d265" (UID: "330a479e-3d76-4de3-bc84-1d7c2d65d265"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:53:21 crc kubenswrapper[4821]: I1125 10:53:21.093626 4821 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/330a479e-3d76-4de3-bc84-1d7c2d65d265-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 25 10:53:21 crc kubenswrapper[4821]: I1125 10:53:21.093929 4821 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Nov 25 10:53:21 crc kubenswrapper[4821]: I1125 10:53:21.510089 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 25 10:53:21 crc kubenswrapper[4821]: I1125 10:53:21.528891 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"330a479e-3d76-4de3-bc84-1d7c2d65d265","Type":"ContainerDied","Data":"79ce92a97646f24487ef3a12c92318542dd51e8f2b1e98173fcec59237602084"} Nov 25 10:53:21 crc kubenswrapper[4821]: I1125 10:53:21.528976 4821 scope.go:117] "RemoveContainer" containerID="7848d9adca99a2cef746a81eaafad2bfec909b4451ebcf92763c11f7448b1d15" Nov 25 10:53:21 crc kubenswrapper[4821]: I1125 10:53:21.528918 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:53:21 crc kubenswrapper[4821]: I1125 10:53:21.637066 4821 scope.go:117] "RemoveContainer" containerID="658c4ee0f18233ace6320ae015dedf80011737bf96f761266c9995e11d4f5dc7" Nov 25 10:53:21 crc kubenswrapper[4821]: I1125 10:53:21.680870 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 25 10:53:21 crc kubenswrapper[4821]: I1125 10:53:21.697838 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 25 10:53:21 crc kubenswrapper[4821]: I1125 10:53:21.716289 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 25 10:53:21 crc kubenswrapper[4821]: E1125 10:53:21.716741 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="330a479e-3d76-4de3-bc84-1d7c2d65d265" containerName="setup-container" Nov 25 10:53:21 crc kubenswrapper[4821]: I1125 10:53:21.716765 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="330a479e-3d76-4de3-bc84-1d7c2d65d265" containerName="setup-container" Nov 25 10:53:21 crc kubenswrapper[4821]: E1125 10:53:21.716785 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="330a479e-3d76-4de3-bc84-1d7c2d65d265" containerName="rabbitmq" Nov 25 10:53:21 crc kubenswrapper[4821]: I1125 10:53:21.716792 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="330a479e-3d76-4de3-bc84-1d7c2d65d265" containerName="rabbitmq" Nov 25 10:53:21 crc kubenswrapper[4821]: I1125 10:53:21.716973 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="330a479e-3d76-4de3-bc84-1d7c2d65d265" containerName="rabbitmq" Nov 25 10:53:21 crc kubenswrapper[4821]: I1125 10:53:21.718413 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:53:21 crc kubenswrapper[4821]: I1125 10:53:21.720104 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 25 10:53:21 crc kubenswrapper[4821]: I1125 10:53:21.721883 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 25 10:53:21 crc kubenswrapper[4821]: I1125 10:53:21.722131 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Nov 25 10:53:21 crc kubenswrapper[4821]: I1125 10:53:21.722269 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Nov 25 10:53:21 crc kubenswrapper[4821]: I1125 10:53:21.722400 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 25 10:53:21 crc kubenswrapper[4821]: I1125 10:53:21.722507 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 25 10:53:21 crc kubenswrapper[4821]: I1125 10:53:21.722636 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-vfvpv" Nov 25 10:53:21 crc kubenswrapper[4821]: I1125 10:53:21.729153 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 25 10:53:21 crc kubenswrapper[4821]: I1125 10:53:21.806337 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3556cc6b-7d73-4cbc-9c2e-378c6ba606e4-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"3556cc6b-7d73-4cbc-9c2e-378c6ba606e4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:53:21 crc kubenswrapper[4821]: I1125 10:53:21.806443 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3556cc6b-7d73-4cbc-9c2e-378c6ba606e4-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"3556cc6b-7d73-4cbc-9c2e-378c6ba606e4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:53:21 crc kubenswrapper[4821]: I1125 10:53:21.806472 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pntrl\" (UniqueName: \"kubernetes.io/projected/3556cc6b-7d73-4cbc-9c2e-378c6ba606e4-kube-api-access-pntrl\") pod \"rabbitmq-cell1-server-0\" (UID: \"3556cc6b-7d73-4cbc-9c2e-378c6ba606e4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:53:21 crc kubenswrapper[4821]: I1125 10:53:21.806531 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3556cc6b-7d73-4cbc-9c2e-378c6ba606e4-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3556cc6b-7d73-4cbc-9c2e-378c6ba606e4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:53:21 crc kubenswrapper[4821]: I1125 10:53:21.806664 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3556cc6b-7d73-4cbc-9c2e-378c6ba606e4-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"3556cc6b-7d73-4cbc-9c2e-378c6ba606e4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:53:21 crc kubenswrapper[4821]: I1125 10:53:21.806728 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3556cc6b-7d73-4cbc-9c2e-378c6ba606e4-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3556cc6b-7d73-4cbc-9c2e-378c6ba606e4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:53:21 crc kubenswrapper[4821]: I1125 10:53:21.806755 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3556cc6b-7d73-4cbc-9c2e-378c6ba606e4-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"3556cc6b-7d73-4cbc-9c2e-378c6ba606e4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:53:21 crc kubenswrapper[4821]: I1125 10:53:21.806837 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3556cc6b-7d73-4cbc-9c2e-378c6ba606e4-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"3556cc6b-7d73-4cbc-9c2e-378c6ba606e4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:53:21 crc kubenswrapper[4821]: I1125 10:53:21.806908 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3556cc6b-7d73-4cbc-9c2e-378c6ba606e4-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"3556cc6b-7d73-4cbc-9c2e-378c6ba606e4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:53:21 crc kubenswrapper[4821]: I1125 10:53:21.806943 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"3556cc6b-7d73-4cbc-9c2e-378c6ba606e4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:53:21 crc kubenswrapper[4821]: I1125 10:53:21.807181 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3556cc6b-7d73-4cbc-9c2e-378c6ba606e4-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"3556cc6b-7d73-4cbc-9c2e-378c6ba606e4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:53:21 crc kubenswrapper[4821]: I1125 10:53:21.908588 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3556cc6b-7d73-4cbc-9c2e-378c6ba606e4-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"3556cc6b-7d73-4cbc-9c2e-378c6ba606e4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:53:21 crc kubenswrapper[4821]: I1125 10:53:21.908643 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3556cc6b-7d73-4cbc-9c2e-378c6ba606e4-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"3556cc6b-7d73-4cbc-9c2e-378c6ba606e4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:53:21 crc kubenswrapper[4821]: I1125 10:53:21.908665 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"3556cc6b-7d73-4cbc-9c2e-378c6ba606e4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:53:21 crc kubenswrapper[4821]: I1125 10:53:21.908755 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3556cc6b-7d73-4cbc-9c2e-378c6ba606e4-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"3556cc6b-7d73-4cbc-9c2e-378c6ba606e4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:53:21 crc kubenswrapper[4821]: I1125 10:53:21.908779 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3556cc6b-7d73-4cbc-9c2e-378c6ba606e4-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"3556cc6b-7d73-4cbc-9c2e-378c6ba606e4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:53:21 crc kubenswrapper[4821]: I1125 10:53:21.908815 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3556cc6b-7d73-4cbc-9c2e-378c6ba606e4-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"3556cc6b-7d73-4cbc-9c2e-378c6ba606e4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:53:21 crc kubenswrapper[4821]: I1125 10:53:21.908830 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pntrl\" (UniqueName: \"kubernetes.io/projected/3556cc6b-7d73-4cbc-9c2e-378c6ba606e4-kube-api-access-pntrl\") pod \"rabbitmq-cell1-server-0\" (UID: \"3556cc6b-7d73-4cbc-9c2e-378c6ba606e4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:53:21 crc kubenswrapper[4821]: I1125 10:53:21.908873 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3556cc6b-7d73-4cbc-9c2e-378c6ba606e4-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3556cc6b-7d73-4cbc-9c2e-378c6ba606e4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:53:21 crc kubenswrapper[4821]: I1125 10:53:21.908900 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3556cc6b-7d73-4cbc-9c2e-378c6ba606e4-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"3556cc6b-7d73-4cbc-9c2e-378c6ba606e4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:53:21 crc kubenswrapper[4821]: I1125 10:53:21.908918 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3556cc6b-7d73-4cbc-9c2e-378c6ba606e4-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3556cc6b-7d73-4cbc-9c2e-378c6ba606e4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:53:21 crc kubenswrapper[4821]: I1125 10:53:21.908933 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3556cc6b-7d73-4cbc-9c2e-378c6ba606e4-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"3556cc6b-7d73-4cbc-9c2e-378c6ba606e4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:53:21 crc kubenswrapper[4821]: I1125 10:53:21.909214 4821 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"3556cc6b-7d73-4cbc-9c2e-378c6ba606e4\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:53:21 crc kubenswrapper[4821]: I1125 10:53:21.909411 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3556cc6b-7d73-4cbc-9c2e-378c6ba606e4-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"3556cc6b-7d73-4cbc-9c2e-378c6ba606e4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:53:21 crc kubenswrapper[4821]: I1125 10:53:21.909544 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3556cc6b-7d73-4cbc-9c2e-378c6ba606e4-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"3556cc6b-7d73-4cbc-9c2e-378c6ba606e4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:53:21 crc kubenswrapper[4821]: I1125 10:53:21.910335 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3556cc6b-7d73-4cbc-9c2e-378c6ba606e4-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"3556cc6b-7d73-4cbc-9c2e-378c6ba606e4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:53:21 crc kubenswrapper[4821]: I1125 10:53:21.910343 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3556cc6b-7d73-4cbc-9c2e-378c6ba606e4-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3556cc6b-7d73-4cbc-9c2e-378c6ba606e4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:53:21 crc kubenswrapper[4821]: I1125 10:53:21.910519 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3556cc6b-7d73-4cbc-9c2e-378c6ba606e4-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3556cc6b-7d73-4cbc-9c2e-378c6ba606e4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:53:21 crc kubenswrapper[4821]: I1125 10:53:21.912959 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3556cc6b-7d73-4cbc-9c2e-378c6ba606e4-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"3556cc6b-7d73-4cbc-9c2e-378c6ba606e4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:53:21 crc kubenswrapper[4821]: I1125 10:53:21.913002 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3556cc6b-7d73-4cbc-9c2e-378c6ba606e4-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"3556cc6b-7d73-4cbc-9c2e-378c6ba606e4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:53:21 crc kubenswrapper[4821]: I1125 10:53:21.914147 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3556cc6b-7d73-4cbc-9c2e-378c6ba606e4-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"3556cc6b-7d73-4cbc-9c2e-378c6ba606e4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:53:21 crc kubenswrapper[4821]: I1125 10:53:21.915244 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3556cc6b-7d73-4cbc-9c2e-378c6ba606e4-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"3556cc6b-7d73-4cbc-9c2e-378c6ba606e4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:53:21 crc kubenswrapper[4821]: I1125 10:53:21.932022 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pntrl\" (UniqueName: \"kubernetes.io/projected/3556cc6b-7d73-4cbc-9c2e-378c6ba606e4-kube-api-access-pntrl\") pod \"rabbitmq-cell1-server-0\" (UID: \"3556cc6b-7d73-4cbc-9c2e-378c6ba606e4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:53:21 crc kubenswrapper[4821]: I1125 10:53:21.940028 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"3556cc6b-7d73-4cbc-9c2e-378c6ba606e4\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:53:22 crc kubenswrapper[4821]: I1125 10:53:22.055763 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:53:22 crc kubenswrapper[4821]: I1125 10:53:22.127483 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2bbb754a-fff3-4ba2-ab75-68e48b626ab0" path="/var/lib/kubelet/pods/2bbb754a-fff3-4ba2-ab75-68e48b626ab0/volumes" Nov 25 10:53:22 crc kubenswrapper[4821]: I1125 10:53:22.128677 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="330a479e-3d76-4de3-bc84-1d7c2d65d265" path="/var/lib/kubelet/pods/330a479e-3d76-4de3-bc84-1d7c2d65d265/volumes" Nov 25 10:53:22 crc kubenswrapper[4821]: I1125 10:53:22.510321 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 25 10:53:22 crc kubenswrapper[4821]: I1125 10:53:22.543333 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"50c27d16-2f67-4f90-83f0-83e16161de50","Type":"ContainerStarted","Data":"1a3ddba1844984b0c72a271c3139db2e441463ebecc6118e5579350ed8092b09"} Nov 25 10:53:22 crc kubenswrapper[4821]: I1125 10:53:22.545346 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3556cc6b-7d73-4cbc-9c2e-378c6ba606e4","Type":"ContainerStarted","Data":"da2bd32846f40bb31953d122b1dbfc5c3546fd9ce5af61e7a806a8d531f4d907"} Nov 25 10:53:23 crc kubenswrapper[4821]: I1125 10:53:23.561684 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"50c27d16-2f67-4f90-83f0-83e16161de50","Type":"ContainerStarted","Data":"fbb495a1788e29f2e27e80fc8a10837f27ded913875b6651206cbe09b69b1cc6"} Nov 25 10:53:24 crc kubenswrapper[4821]: I1125 10:53:24.572098 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3556cc6b-7d73-4cbc-9c2e-378c6ba606e4","Type":"ContainerStarted","Data":"6a26c64b2bbe34526964a22113590ca4f5b4d5e3570e85c3d786f20becf68924"} Nov 25 10:53:24 crc kubenswrapper[4821]: I1125 10:53:24.650385 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-d558885bc-dbvrb"] Nov 25 10:53:24 crc kubenswrapper[4821]: I1125 10:53:24.651928 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d558885bc-dbvrb" Nov 25 10:53:24 crc kubenswrapper[4821]: I1125 10:53:24.655044 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Nov 25 10:53:24 crc kubenswrapper[4821]: I1125 10:53:24.664264 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-dbvrb"] Nov 25 10:53:24 crc kubenswrapper[4821]: I1125 10:53:24.757882 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ceba82fa-60b8-42d5-af83-7860a39b20a3-ovsdbserver-nb\") pod \"dnsmasq-dns-d558885bc-dbvrb\" (UID: \"ceba82fa-60b8-42d5-af83-7860a39b20a3\") " pod="openstack/dnsmasq-dns-d558885bc-dbvrb" Nov 25 10:53:24 crc kubenswrapper[4821]: I1125 10:53:24.758010 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ceba82fa-60b8-42d5-af83-7860a39b20a3-dns-svc\") pod \"dnsmasq-dns-d558885bc-dbvrb\" (UID: \"ceba82fa-60b8-42d5-af83-7860a39b20a3\") " pod="openstack/dnsmasq-dns-d558885bc-dbvrb" Nov 25 10:53:24 crc kubenswrapper[4821]: I1125 10:53:24.758071 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ceba82fa-60b8-42d5-af83-7860a39b20a3-ovsdbserver-sb\") pod \"dnsmasq-dns-d558885bc-dbvrb\" (UID: \"ceba82fa-60b8-42d5-af83-7860a39b20a3\") " pod="openstack/dnsmasq-dns-d558885bc-dbvrb" Nov 25 10:53:24 crc kubenswrapper[4821]: I1125 10:53:24.758096 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ceba82fa-60b8-42d5-af83-7860a39b20a3-openstack-edpm-ipam\") pod \"dnsmasq-dns-d558885bc-dbvrb\" (UID: \"ceba82fa-60b8-42d5-af83-7860a39b20a3\") " pod="openstack/dnsmasq-dns-d558885bc-dbvrb" Nov 25 10:53:24 crc kubenswrapper[4821]: I1125 10:53:24.758189 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ceba82fa-60b8-42d5-af83-7860a39b20a3-config\") pod \"dnsmasq-dns-d558885bc-dbvrb\" (UID: \"ceba82fa-60b8-42d5-af83-7860a39b20a3\") " pod="openstack/dnsmasq-dns-d558885bc-dbvrb" Nov 25 10:53:24 crc kubenswrapper[4821]: I1125 10:53:24.758232 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ceba82fa-60b8-42d5-af83-7860a39b20a3-dns-swift-storage-0\") pod \"dnsmasq-dns-d558885bc-dbvrb\" (UID: \"ceba82fa-60b8-42d5-af83-7860a39b20a3\") " pod="openstack/dnsmasq-dns-d558885bc-dbvrb" Nov 25 10:53:24 crc kubenswrapper[4821]: I1125 10:53:24.758266 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7wqz\" (UniqueName: \"kubernetes.io/projected/ceba82fa-60b8-42d5-af83-7860a39b20a3-kube-api-access-q7wqz\") pod \"dnsmasq-dns-d558885bc-dbvrb\" (UID: \"ceba82fa-60b8-42d5-af83-7860a39b20a3\") " pod="openstack/dnsmasq-dns-d558885bc-dbvrb" Nov 25 10:53:24 crc kubenswrapper[4821]: I1125 10:53:24.859835 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ceba82fa-60b8-42d5-af83-7860a39b20a3-dns-swift-storage-0\") pod \"dnsmasq-dns-d558885bc-dbvrb\" (UID: \"ceba82fa-60b8-42d5-af83-7860a39b20a3\") " pod="openstack/dnsmasq-dns-d558885bc-dbvrb" Nov 25 10:53:24 crc kubenswrapper[4821]: I1125 10:53:24.859915 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7wqz\" (UniqueName: \"kubernetes.io/projected/ceba82fa-60b8-42d5-af83-7860a39b20a3-kube-api-access-q7wqz\") pod \"dnsmasq-dns-d558885bc-dbvrb\" (UID: \"ceba82fa-60b8-42d5-af83-7860a39b20a3\") " pod="openstack/dnsmasq-dns-d558885bc-dbvrb" Nov 25 10:53:24 crc kubenswrapper[4821]: I1125 10:53:24.859972 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ceba82fa-60b8-42d5-af83-7860a39b20a3-ovsdbserver-nb\") pod \"dnsmasq-dns-d558885bc-dbvrb\" (UID: \"ceba82fa-60b8-42d5-af83-7860a39b20a3\") " pod="openstack/dnsmasq-dns-d558885bc-dbvrb" Nov 25 10:53:24 crc kubenswrapper[4821]: I1125 10:53:24.860051 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ceba82fa-60b8-42d5-af83-7860a39b20a3-dns-svc\") pod \"dnsmasq-dns-d558885bc-dbvrb\" (UID: \"ceba82fa-60b8-42d5-af83-7860a39b20a3\") " pod="openstack/dnsmasq-dns-d558885bc-dbvrb" Nov 25 10:53:24 crc kubenswrapper[4821]: I1125 10:53:24.860132 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ceba82fa-60b8-42d5-af83-7860a39b20a3-ovsdbserver-sb\") pod \"dnsmasq-dns-d558885bc-dbvrb\" (UID: \"ceba82fa-60b8-42d5-af83-7860a39b20a3\") " pod="openstack/dnsmasq-dns-d558885bc-dbvrb" Nov 25 10:53:24 crc kubenswrapper[4821]: I1125 10:53:24.860185 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ceba82fa-60b8-42d5-af83-7860a39b20a3-openstack-edpm-ipam\") pod \"dnsmasq-dns-d558885bc-dbvrb\" (UID: \"ceba82fa-60b8-42d5-af83-7860a39b20a3\") " pod="openstack/dnsmasq-dns-d558885bc-dbvrb" Nov 25 10:53:24 crc kubenswrapper[4821]: I1125 10:53:24.860245 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ceba82fa-60b8-42d5-af83-7860a39b20a3-config\") pod \"dnsmasq-dns-d558885bc-dbvrb\" (UID: \"ceba82fa-60b8-42d5-af83-7860a39b20a3\") " pod="openstack/dnsmasq-dns-d558885bc-dbvrb" Nov 25 10:53:24 crc kubenswrapper[4821]: I1125 10:53:24.860939 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ceba82fa-60b8-42d5-af83-7860a39b20a3-dns-swift-storage-0\") pod \"dnsmasq-dns-d558885bc-dbvrb\" (UID: \"ceba82fa-60b8-42d5-af83-7860a39b20a3\") " pod="openstack/dnsmasq-dns-d558885bc-dbvrb" Nov 25 10:53:24 crc kubenswrapper[4821]: I1125 10:53:24.860962 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ceba82fa-60b8-42d5-af83-7860a39b20a3-ovsdbserver-sb\") pod \"dnsmasq-dns-d558885bc-dbvrb\" (UID: \"ceba82fa-60b8-42d5-af83-7860a39b20a3\") " pod="openstack/dnsmasq-dns-d558885bc-dbvrb" Nov 25 10:53:24 crc kubenswrapper[4821]: I1125 10:53:24.861131 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ceba82fa-60b8-42d5-af83-7860a39b20a3-dns-svc\") pod \"dnsmasq-dns-d558885bc-dbvrb\" (UID: \"ceba82fa-60b8-42d5-af83-7860a39b20a3\") " pod="openstack/dnsmasq-dns-d558885bc-dbvrb" Nov 25 10:53:24 crc kubenswrapper[4821]: I1125 10:53:24.861732 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ceba82fa-60b8-42d5-af83-7860a39b20a3-openstack-edpm-ipam\") pod \"dnsmasq-dns-d558885bc-dbvrb\" (UID: \"ceba82fa-60b8-42d5-af83-7860a39b20a3\") " pod="openstack/dnsmasq-dns-d558885bc-dbvrb" Nov 25 10:53:24 crc kubenswrapper[4821]: I1125 10:53:24.861936 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ceba82fa-60b8-42d5-af83-7860a39b20a3-config\") pod \"dnsmasq-dns-d558885bc-dbvrb\" (UID: \"ceba82fa-60b8-42d5-af83-7860a39b20a3\") " pod="openstack/dnsmasq-dns-d558885bc-dbvrb" Nov 25 10:53:24 crc kubenswrapper[4821]: I1125 10:53:24.861982 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ceba82fa-60b8-42d5-af83-7860a39b20a3-ovsdbserver-nb\") pod \"dnsmasq-dns-d558885bc-dbvrb\" (UID: \"ceba82fa-60b8-42d5-af83-7860a39b20a3\") " pod="openstack/dnsmasq-dns-d558885bc-dbvrb" Nov 25 10:53:24 crc kubenswrapper[4821]: I1125 10:53:24.881556 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7wqz\" (UniqueName: \"kubernetes.io/projected/ceba82fa-60b8-42d5-af83-7860a39b20a3-kube-api-access-q7wqz\") pod \"dnsmasq-dns-d558885bc-dbvrb\" (UID: \"ceba82fa-60b8-42d5-af83-7860a39b20a3\") " pod="openstack/dnsmasq-dns-d558885bc-dbvrb" Nov 25 10:53:24 crc kubenswrapper[4821]: I1125 10:53:24.971297 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d558885bc-dbvrb" Nov 25 10:53:25 crc kubenswrapper[4821]: I1125 10:53:25.431687 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-dbvrb"] Nov 25 10:53:25 crc kubenswrapper[4821]: I1125 10:53:25.582495 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-dbvrb" event={"ID":"ceba82fa-60b8-42d5-af83-7860a39b20a3","Type":"ContainerStarted","Data":"efe3efa1d24b0654ee47ab50d280aee842d3aabd85b35e9d4f1a65717a5eb8b4"} Nov 25 10:53:27 crc kubenswrapper[4821]: I1125 10:53:27.601502 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-dbvrb" event={"ID":"ceba82fa-60b8-42d5-af83-7860a39b20a3","Type":"ContainerStarted","Data":"9c700324b5af2a723f00cc2a38925a7dbe3ead006a6a8a654a815f1d9f258055"} Nov 25 10:53:28 crc kubenswrapper[4821]: I1125 10:53:28.617575 4821 generic.go:334] "Generic (PLEG): container finished" podID="ceba82fa-60b8-42d5-af83-7860a39b20a3" containerID="9c700324b5af2a723f00cc2a38925a7dbe3ead006a6a8a654a815f1d9f258055" exitCode=0 Nov 25 10:53:28 crc kubenswrapper[4821]: I1125 10:53:28.617733 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-dbvrb" event={"ID":"ceba82fa-60b8-42d5-af83-7860a39b20a3","Type":"ContainerDied","Data":"9c700324b5af2a723f00cc2a38925a7dbe3ead006a6a8a654a815f1d9f258055"} Nov 25 10:53:29 crc kubenswrapper[4821]: I1125 10:53:29.628092 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-dbvrb" event={"ID":"ceba82fa-60b8-42d5-af83-7860a39b20a3","Type":"ContainerStarted","Data":"355620a153321dcfe92de2ea0a81d2ba2b3470b4d901dd9c1ce6faeded09a7cd"} Nov 25 10:53:29 crc kubenswrapper[4821]: I1125 10:53:29.628497 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-d558885bc-dbvrb" Nov 25 10:53:29 crc kubenswrapper[4821]: I1125 10:53:29.651117 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-d558885bc-dbvrb" podStartSLOduration=5.651094757 podStartE2EDuration="5.651094757s" podCreationTimestamp="2025-11-25 10:53:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:53:29.645353686 +0000 UTC m=+1280.181673533" watchObservedRunningTime="2025-11-25 10:53:29.651094757 +0000 UTC m=+1280.187414604" Nov 25 10:53:34 crc kubenswrapper[4821]: I1125 10:53:34.973604 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-d558885bc-dbvrb" Nov 25 10:53:35 crc kubenswrapper[4821]: I1125 10:53:35.045829 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-xd4pf"] Nov 25 10:53:35 crc kubenswrapper[4821]: I1125 10:53:35.046125 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-cd5cbd7b9-xd4pf" podUID="db92740e-3cd9-4b81-b153-3f866fd3db62" containerName="dnsmasq-dns" containerID="cri-o://f1f4e312d2979b1204af23dde344c8ca9c9547dab34f7e7fdd9fb0184f5f582f" gracePeriod=10 Nov 25 10:53:35 crc kubenswrapper[4821]: I1125 10:53:35.170305 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78c64bc9c5-8r5jj"] Nov 25 10:53:35 crc kubenswrapper[4821]: I1125 10:53:35.172275 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78c64bc9c5-8r5jj" Nov 25 10:53:35 crc kubenswrapper[4821]: I1125 10:53:35.190016 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78c64bc9c5-8r5jj"] Nov 25 10:53:35 crc kubenswrapper[4821]: I1125 10:53:35.352303 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bda8a37b-363b-446a-8cbb-a708cb3f716f-dns-svc\") pod \"dnsmasq-dns-78c64bc9c5-8r5jj\" (UID: \"bda8a37b-363b-446a-8cbb-a708cb3f716f\") " pod="openstack/dnsmasq-dns-78c64bc9c5-8r5jj" Nov 25 10:53:35 crc kubenswrapper[4821]: I1125 10:53:35.352864 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bda8a37b-363b-446a-8cbb-a708cb3f716f-ovsdbserver-sb\") pod \"dnsmasq-dns-78c64bc9c5-8r5jj\" (UID: \"bda8a37b-363b-446a-8cbb-a708cb3f716f\") " pod="openstack/dnsmasq-dns-78c64bc9c5-8r5jj" Nov 25 10:53:35 crc kubenswrapper[4821]: I1125 10:53:35.352958 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jjvdb\" (UniqueName: \"kubernetes.io/projected/bda8a37b-363b-446a-8cbb-a708cb3f716f-kube-api-access-jjvdb\") pod \"dnsmasq-dns-78c64bc9c5-8r5jj\" (UID: \"bda8a37b-363b-446a-8cbb-a708cb3f716f\") " pod="openstack/dnsmasq-dns-78c64bc9c5-8r5jj" Nov 25 10:53:35 crc kubenswrapper[4821]: I1125 10:53:35.352987 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bda8a37b-363b-446a-8cbb-a708cb3f716f-ovsdbserver-nb\") pod \"dnsmasq-dns-78c64bc9c5-8r5jj\" (UID: \"bda8a37b-363b-446a-8cbb-a708cb3f716f\") " pod="openstack/dnsmasq-dns-78c64bc9c5-8r5jj" Nov 25 10:53:35 crc kubenswrapper[4821]: I1125 10:53:35.353071 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bda8a37b-363b-446a-8cbb-a708cb3f716f-dns-swift-storage-0\") pod \"dnsmasq-dns-78c64bc9c5-8r5jj\" (UID: \"bda8a37b-363b-446a-8cbb-a708cb3f716f\") " pod="openstack/dnsmasq-dns-78c64bc9c5-8r5jj" Nov 25 10:53:35 crc kubenswrapper[4821]: I1125 10:53:35.353403 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/bda8a37b-363b-446a-8cbb-a708cb3f716f-openstack-edpm-ipam\") pod \"dnsmasq-dns-78c64bc9c5-8r5jj\" (UID: \"bda8a37b-363b-446a-8cbb-a708cb3f716f\") " pod="openstack/dnsmasq-dns-78c64bc9c5-8r5jj" Nov 25 10:53:35 crc kubenswrapper[4821]: I1125 10:53:35.353510 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bda8a37b-363b-446a-8cbb-a708cb3f716f-config\") pod \"dnsmasq-dns-78c64bc9c5-8r5jj\" (UID: \"bda8a37b-363b-446a-8cbb-a708cb3f716f\") " pod="openstack/dnsmasq-dns-78c64bc9c5-8r5jj" Nov 25 10:53:35 crc kubenswrapper[4821]: I1125 10:53:35.456138 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bda8a37b-363b-446a-8cbb-a708cb3f716f-dns-svc\") pod \"dnsmasq-dns-78c64bc9c5-8r5jj\" (UID: \"bda8a37b-363b-446a-8cbb-a708cb3f716f\") " pod="openstack/dnsmasq-dns-78c64bc9c5-8r5jj" Nov 25 10:53:35 crc kubenswrapper[4821]: I1125 10:53:35.456234 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bda8a37b-363b-446a-8cbb-a708cb3f716f-ovsdbserver-sb\") pod \"dnsmasq-dns-78c64bc9c5-8r5jj\" (UID: \"bda8a37b-363b-446a-8cbb-a708cb3f716f\") " pod="openstack/dnsmasq-dns-78c64bc9c5-8r5jj" Nov 25 10:53:35 crc kubenswrapper[4821]: I1125 10:53:35.456275 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jjvdb\" (UniqueName: \"kubernetes.io/projected/bda8a37b-363b-446a-8cbb-a708cb3f716f-kube-api-access-jjvdb\") pod \"dnsmasq-dns-78c64bc9c5-8r5jj\" (UID: \"bda8a37b-363b-446a-8cbb-a708cb3f716f\") " pod="openstack/dnsmasq-dns-78c64bc9c5-8r5jj" Nov 25 10:53:35 crc kubenswrapper[4821]: I1125 10:53:35.456303 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bda8a37b-363b-446a-8cbb-a708cb3f716f-ovsdbserver-nb\") pod \"dnsmasq-dns-78c64bc9c5-8r5jj\" (UID: \"bda8a37b-363b-446a-8cbb-a708cb3f716f\") " pod="openstack/dnsmasq-dns-78c64bc9c5-8r5jj" Nov 25 10:53:35 crc kubenswrapper[4821]: I1125 10:53:35.456337 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bda8a37b-363b-446a-8cbb-a708cb3f716f-dns-swift-storage-0\") pod \"dnsmasq-dns-78c64bc9c5-8r5jj\" (UID: \"bda8a37b-363b-446a-8cbb-a708cb3f716f\") " pod="openstack/dnsmasq-dns-78c64bc9c5-8r5jj" Nov 25 10:53:35 crc kubenswrapper[4821]: I1125 10:53:35.456431 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/bda8a37b-363b-446a-8cbb-a708cb3f716f-openstack-edpm-ipam\") pod \"dnsmasq-dns-78c64bc9c5-8r5jj\" (UID: \"bda8a37b-363b-446a-8cbb-a708cb3f716f\") " pod="openstack/dnsmasq-dns-78c64bc9c5-8r5jj" Nov 25 10:53:35 crc kubenswrapper[4821]: I1125 10:53:35.456481 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bda8a37b-363b-446a-8cbb-a708cb3f716f-config\") pod \"dnsmasq-dns-78c64bc9c5-8r5jj\" (UID: \"bda8a37b-363b-446a-8cbb-a708cb3f716f\") " pod="openstack/dnsmasq-dns-78c64bc9c5-8r5jj" Nov 25 10:53:35 crc kubenswrapper[4821]: I1125 10:53:35.457757 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bda8a37b-363b-446a-8cbb-a708cb3f716f-dns-svc\") pod \"dnsmasq-dns-78c64bc9c5-8r5jj\" (UID: \"bda8a37b-363b-446a-8cbb-a708cb3f716f\") " pod="openstack/dnsmasq-dns-78c64bc9c5-8r5jj" Nov 25 10:53:35 crc kubenswrapper[4821]: I1125 10:53:35.457994 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bda8a37b-363b-446a-8cbb-a708cb3f716f-ovsdbserver-sb\") pod \"dnsmasq-dns-78c64bc9c5-8r5jj\" (UID: \"bda8a37b-363b-446a-8cbb-a708cb3f716f\") " pod="openstack/dnsmasq-dns-78c64bc9c5-8r5jj" Nov 25 10:53:35 crc kubenswrapper[4821]: I1125 10:53:35.460005 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bda8a37b-363b-446a-8cbb-a708cb3f716f-dns-swift-storage-0\") pod \"dnsmasq-dns-78c64bc9c5-8r5jj\" (UID: \"bda8a37b-363b-446a-8cbb-a708cb3f716f\") " pod="openstack/dnsmasq-dns-78c64bc9c5-8r5jj" Nov 25 10:53:35 crc kubenswrapper[4821]: I1125 10:53:35.461833 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/bda8a37b-363b-446a-8cbb-a708cb3f716f-openstack-edpm-ipam\") pod \"dnsmasq-dns-78c64bc9c5-8r5jj\" (UID: \"bda8a37b-363b-446a-8cbb-a708cb3f716f\") " pod="openstack/dnsmasq-dns-78c64bc9c5-8r5jj" Nov 25 10:53:35 crc kubenswrapper[4821]: I1125 10:53:35.461849 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bda8a37b-363b-446a-8cbb-a708cb3f716f-config\") pod \"dnsmasq-dns-78c64bc9c5-8r5jj\" (UID: \"bda8a37b-363b-446a-8cbb-a708cb3f716f\") " pod="openstack/dnsmasq-dns-78c64bc9c5-8r5jj" Nov 25 10:53:35 crc kubenswrapper[4821]: I1125 10:53:35.463881 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bda8a37b-363b-446a-8cbb-a708cb3f716f-ovsdbserver-nb\") pod \"dnsmasq-dns-78c64bc9c5-8r5jj\" (UID: \"bda8a37b-363b-446a-8cbb-a708cb3f716f\") " pod="openstack/dnsmasq-dns-78c64bc9c5-8r5jj" Nov 25 10:53:35 crc kubenswrapper[4821]: I1125 10:53:35.484897 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jjvdb\" (UniqueName: \"kubernetes.io/projected/bda8a37b-363b-446a-8cbb-a708cb3f716f-kube-api-access-jjvdb\") pod \"dnsmasq-dns-78c64bc9c5-8r5jj\" (UID: \"bda8a37b-363b-446a-8cbb-a708cb3f716f\") " pod="openstack/dnsmasq-dns-78c64bc9c5-8r5jj" Nov 25 10:53:35 crc kubenswrapper[4821]: I1125 10:53:35.519890 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78c64bc9c5-8r5jj" Nov 25 10:53:35 crc kubenswrapper[4821]: I1125 10:53:35.669143 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-xd4pf" Nov 25 10:53:35 crc kubenswrapper[4821]: I1125 10:53:35.689945 4821 generic.go:334] "Generic (PLEG): container finished" podID="db92740e-3cd9-4b81-b153-3f866fd3db62" containerID="f1f4e312d2979b1204af23dde344c8ca9c9547dab34f7e7fdd9fb0184f5f582f" exitCode=0 Nov 25 10:53:35 crc kubenswrapper[4821]: I1125 10:53:35.689987 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-xd4pf" event={"ID":"db92740e-3cd9-4b81-b153-3f866fd3db62","Type":"ContainerDied","Data":"f1f4e312d2979b1204af23dde344c8ca9c9547dab34f7e7fdd9fb0184f5f582f"} Nov 25 10:53:35 crc kubenswrapper[4821]: I1125 10:53:35.690014 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-xd4pf" event={"ID":"db92740e-3cd9-4b81-b153-3f866fd3db62","Type":"ContainerDied","Data":"57c57fb91904c23186e70ab11e5d85b3e5886f55850beade32a56686f77fbd0c"} Nov 25 10:53:35 crc kubenswrapper[4821]: I1125 10:53:35.690032 4821 scope.go:117] "RemoveContainer" containerID="f1f4e312d2979b1204af23dde344c8ca9c9547dab34f7e7fdd9fb0184f5f582f" Nov 25 10:53:35 crc kubenswrapper[4821]: I1125 10:53:35.690033 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-xd4pf" Nov 25 10:53:35 crc kubenswrapper[4821]: I1125 10:53:35.730558 4821 scope.go:117] "RemoveContainer" containerID="f7a503d2e982596c45acc041ea6221e887ebb186f6f3e48e63644ac930367f7d" Nov 25 10:53:35 crc kubenswrapper[4821]: I1125 10:53:35.762289 4821 scope.go:117] "RemoveContainer" containerID="f1f4e312d2979b1204af23dde344c8ca9c9547dab34f7e7fdd9fb0184f5f582f" Nov 25 10:53:35 crc kubenswrapper[4821]: E1125 10:53:35.763047 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f1f4e312d2979b1204af23dde344c8ca9c9547dab34f7e7fdd9fb0184f5f582f\": container with ID starting with f1f4e312d2979b1204af23dde344c8ca9c9547dab34f7e7fdd9fb0184f5f582f not found: ID does not exist" containerID="f1f4e312d2979b1204af23dde344c8ca9c9547dab34f7e7fdd9fb0184f5f582f" Nov 25 10:53:35 crc kubenswrapper[4821]: I1125 10:53:35.763085 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1f4e312d2979b1204af23dde344c8ca9c9547dab34f7e7fdd9fb0184f5f582f"} err="failed to get container status \"f1f4e312d2979b1204af23dde344c8ca9c9547dab34f7e7fdd9fb0184f5f582f\": rpc error: code = NotFound desc = could not find container \"f1f4e312d2979b1204af23dde344c8ca9c9547dab34f7e7fdd9fb0184f5f582f\": container with ID starting with f1f4e312d2979b1204af23dde344c8ca9c9547dab34f7e7fdd9fb0184f5f582f not found: ID does not exist" Nov 25 10:53:35 crc kubenswrapper[4821]: I1125 10:53:35.763106 4821 scope.go:117] "RemoveContainer" containerID="f7a503d2e982596c45acc041ea6221e887ebb186f6f3e48e63644ac930367f7d" Nov 25 10:53:35 crc kubenswrapper[4821]: I1125 10:53:35.763871 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db92740e-3cd9-4b81-b153-3f866fd3db62-config\") pod \"db92740e-3cd9-4b81-b153-3f866fd3db62\" (UID: \"db92740e-3cd9-4b81-b153-3f866fd3db62\") " Nov 25 10:53:35 crc kubenswrapper[4821]: I1125 10:53:35.764072 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/db92740e-3cd9-4b81-b153-3f866fd3db62-dns-svc\") pod \"db92740e-3cd9-4b81-b153-3f866fd3db62\" (UID: \"db92740e-3cd9-4b81-b153-3f866fd3db62\") " Nov 25 10:53:35 crc kubenswrapper[4821]: I1125 10:53:35.764298 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/db92740e-3cd9-4b81-b153-3f866fd3db62-dns-swift-storage-0\") pod \"db92740e-3cd9-4b81-b153-3f866fd3db62\" (UID: \"db92740e-3cd9-4b81-b153-3f866fd3db62\") " Nov 25 10:53:35 crc kubenswrapper[4821]: E1125 10:53:35.765781 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f7a503d2e982596c45acc041ea6221e887ebb186f6f3e48e63644ac930367f7d\": container with ID starting with f7a503d2e982596c45acc041ea6221e887ebb186f6f3e48e63644ac930367f7d not found: ID does not exist" containerID="f7a503d2e982596c45acc041ea6221e887ebb186f6f3e48e63644ac930367f7d" Nov 25 10:53:35 crc kubenswrapper[4821]: I1125 10:53:35.765832 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7a503d2e982596c45acc041ea6221e887ebb186f6f3e48e63644ac930367f7d"} err="failed to get container status \"f7a503d2e982596c45acc041ea6221e887ebb186f6f3e48e63644ac930367f7d\": rpc error: code = NotFound desc = could not find container \"f7a503d2e982596c45acc041ea6221e887ebb186f6f3e48e63644ac930367f7d\": container with ID starting with f7a503d2e982596c45acc041ea6221e887ebb186f6f3e48e63644ac930367f7d not found: ID does not exist" Nov 25 10:53:35 crc kubenswrapper[4821]: I1125 10:53:35.813032 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db92740e-3cd9-4b81-b153-3f866fd3db62-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "db92740e-3cd9-4b81-b153-3f866fd3db62" (UID: "db92740e-3cd9-4b81-b153-3f866fd3db62"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:53:35 crc kubenswrapper[4821]: I1125 10:53:35.821300 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db92740e-3cd9-4b81-b153-3f866fd3db62-config" (OuterVolumeSpecName: "config") pod "db92740e-3cd9-4b81-b153-3f866fd3db62" (UID: "db92740e-3cd9-4b81-b153-3f866fd3db62"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:53:35 crc kubenswrapper[4821]: I1125 10:53:35.822513 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db92740e-3cd9-4b81-b153-3f866fd3db62-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "db92740e-3cd9-4b81-b153-3f866fd3db62" (UID: "db92740e-3cd9-4b81-b153-3f866fd3db62"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:53:35 crc kubenswrapper[4821]: I1125 10:53:35.865541 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/db92740e-3cd9-4b81-b153-3f866fd3db62-ovsdbserver-nb\") pod \"db92740e-3cd9-4b81-b153-3f866fd3db62\" (UID: \"db92740e-3cd9-4b81-b153-3f866fd3db62\") " Nov 25 10:53:35 crc kubenswrapper[4821]: I1125 10:53:35.865594 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-42sws\" (UniqueName: \"kubernetes.io/projected/db92740e-3cd9-4b81-b153-3f866fd3db62-kube-api-access-42sws\") pod \"db92740e-3cd9-4b81-b153-3f866fd3db62\" (UID: \"db92740e-3cd9-4b81-b153-3f866fd3db62\") " Nov 25 10:53:35 crc kubenswrapper[4821]: I1125 10:53:35.868334 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/db92740e-3cd9-4b81-b153-3f866fd3db62-ovsdbserver-sb\") pod \"db92740e-3cd9-4b81-b153-3f866fd3db62\" (UID: \"db92740e-3cd9-4b81-b153-3f866fd3db62\") " Nov 25 10:53:35 crc kubenswrapper[4821]: I1125 10:53:35.869522 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db92740e-3cd9-4b81-b153-3f866fd3db62-kube-api-access-42sws" (OuterVolumeSpecName: "kube-api-access-42sws") pod "db92740e-3cd9-4b81-b153-3f866fd3db62" (UID: "db92740e-3cd9-4b81-b153-3f866fd3db62"). InnerVolumeSpecName "kube-api-access-42sws". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:53:35 crc kubenswrapper[4821]: I1125 10:53:35.869563 4821 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db92740e-3cd9-4b81-b153-3f866fd3db62-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:53:35 crc kubenswrapper[4821]: I1125 10:53:35.869610 4821 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/db92740e-3cd9-4b81-b153-3f866fd3db62-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 10:53:35 crc kubenswrapper[4821]: I1125 10:53:35.869624 4821 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/db92740e-3cd9-4b81-b153-3f866fd3db62-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 25 10:53:35 crc kubenswrapper[4821]: I1125 10:53:35.921606 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db92740e-3cd9-4b81-b153-3f866fd3db62-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "db92740e-3cd9-4b81-b153-3f866fd3db62" (UID: "db92740e-3cd9-4b81-b153-3f866fd3db62"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:53:35 crc kubenswrapper[4821]: I1125 10:53:35.922038 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db92740e-3cd9-4b81-b153-3f866fd3db62-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "db92740e-3cd9-4b81-b153-3f866fd3db62" (UID: "db92740e-3cd9-4b81-b153-3f866fd3db62"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:53:35 crc kubenswrapper[4821]: I1125 10:53:35.970761 4821 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/db92740e-3cd9-4b81-b153-3f866fd3db62-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 10:53:35 crc kubenswrapper[4821]: I1125 10:53:35.970797 4821 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/db92740e-3cd9-4b81-b153-3f866fd3db62-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 10:53:35 crc kubenswrapper[4821]: I1125 10:53:35.970810 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-42sws\" (UniqueName: \"kubernetes.io/projected/db92740e-3cd9-4b81-b153-3f866fd3db62-kube-api-access-42sws\") on node \"crc\" DevicePath \"\"" Nov 25 10:53:36 crc kubenswrapper[4821]: W1125 10:53:36.019948 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbda8a37b_363b_446a_8cbb_a708cb3f716f.slice/crio-07d4587d521242463d5652325b4f30d230106288b3c5163312cbca05ba4e1119 WatchSource:0}: Error finding container 07d4587d521242463d5652325b4f30d230106288b3c5163312cbca05ba4e1119: Status 404 returned error can't find the container with id 07d4587d521242463d5652325b4f30d230106288b3c5163312cbca05ba4e1119 Nov 25 10:53:36 crc kubenswrapper[4821]: I1125 10:53:36.021411 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78c64bc9c5-8r5jj"] Nov 25 10:53:36 crc kubenswrapper[4821]: I1125 10:53:36.030073 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-xd4pf"] Nov 25 10:53:36 crc kubenswrapper[4821]: I1125 10:53:36.037820 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-xd4pf"] Nov 25 10:53:36 crc kubenswrapper[4821]: I1125 10:53:36.129925 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db92740e-3cd9-4b81-b153-3f866fd3db62" path="/var/lib/kubelet/pods/db92740e-3cd9-4b81-b153-3f866fd3db62/volumes" Nov 25 10:53:36 crc kubenswrapper[4821]: I1125 10:53:36.702633 4821 generic.go:334] "Generic (PLEG): container finished" podID="bda8a37b-363b-446a-8cbb-a708cb3f716f" containerID="426977fff44702faafd7096c037dfaf82daa0fe760dd40b58d840f01e9cd7a60" exitCode=0 Nov 25 10:53:36 crc kubenswrapper[4821]: I1125 10:53:36.702681 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78c64bc9c5-8r5jj" event={"ID":"bda8a37b-363b-446a-8cbb-a708cb3f716f","Type":"ContainerDied","Data":"426977fff44702faafd7096c037dfaf82daa0fe760dd40b58d840f01e9cd7a60"} Nov 25 10:53:36 crc kubenswrapper[4821]: I1125 10:53:36.702705 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78c64bc9c5-8r5jj" event={"ID":"bda8a37b-363b-446a-8cbb-a708cb3f716f","Type":"ContainerStarted","Data":"07d4587d521242463d5652325b4f30d230106288b3c5163312cbca05ba4e1119"} Nov 25 10:53:37 crc kubenswrapper[4821]: I1125 10:53:37.714466 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78c64bc9c5-8r5jj" event={"ID":"bda8a37b-363b-446a-8cbb-a708cb3f716f","Type":"ContainerStarted","Data":"b7663f0756e6e8b08e819b4e0ad7858f16f6e5ed18ae22b811ab6c13753b84dd"} Nov 25 10:53:37 crc kubenswrapper[4821]: I1125 10:53:37.715015 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-78c64bc9c5-8r5jj" Nov 25 10:53:37 crc kubenswrapper[4821]: I1125 10:53:37.742637 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-78c64bc9c5-8r5jj" podStartSLOduration=2.7426202809999998 podStartE2EDuration="2.742620281s" podCreationTimestamp="2025-11-25 10:53:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:53:37.734655253 +0000 UTC m=+1288.270975110" watchObservedRunningTime="2025-11-25 10:53:37.742620281 +0000 UTC m=+1288.278940118" Nov 25 10:53:45 crc kubenswrapper[4821]: I1125 10:53:45.521203 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-78c64bc9c5-8r5jj" Nov 25 10:53:45 crc kubenswrapper[4821]: I1125 10:53:45.575222 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-dbvrb"] Nov 25 10:53:45 crc kubenswrapper[4821]: I1125 10:53:45.575522 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-d558885bc-dbvrb" podUID="ceba82fa-60b8-42d5-af83-7860a39b20a3" containerName="dnsmasq-dns" containerID="cri-o://355620a153321dcfe92de2ea0a81d2ba2b3470b4d901dd9c1ce6faeded09a7cd" gracePeriod=10 Nov 25 10:53:45 crc kubenswrapper[4821]: I1125 10:53:45.784585 4821 generic.go:334] "Generic (PLEG): container finished" podID="ceba82fa-60b8-42d5-af83-7860a39b20a3" containerID="355620a153321dcfe92de2ea0a81d2ba2b3470b4d901dd9c1ce6faeded09a7cd" exitCode=0 Nov 25 10:53:45 crc kubenswrapper[4821]: I1125 10:53:45.784635 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-dbvrb" event={"ID":"ceba82fa-60b8-42d5-af83-7860a39b20a3","Type":"ContainerDied","Data":"355620a153321dcfe92de2ea0a81d2ba2b3470b4d901dd9c1ce6faeded09a7cd"} Nov 25 10:53:46 crc kubenswrapper[4821]: I1125 10:53:46.545293 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d558885bc-dbvrb" Nov 25 10:53:46 crc kubenswrapper[4821]: I1125 10:53:46.687952 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ceba82fa-60b8-42d5-af83-7860a39b20a3-openstack-edpm-ipam\") pod \"ceba82fa-60b8-42d5-af83-7860a39b20a3\" (UID: \"ceba82fa-60b8-42d5-af83-7860a39b20a3\") " Nov 25 10:53:46 crc kubenswrapper[4821]: I1125 10:53:46.688032 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ceba82fa-60b8-42d5-af83-7860a39b20a3-dns-svc\") pod \"ceba82fa-60b8-42d5-af83-7860a39b20a3\" (UID: \"ceba82fa-60b8-42d5-af83-7860a39b20a3\") " Nov 25 10:53:46 crc kubenswrapper[4821]: I1125 10:53:46.688115 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ceba82fa-60b8-42d5-af83-7860a39b20a3-ovsdbserver-sb\") pod \"ceba82fa-60b8-42d5-af83-7860a39b20a3\" (UID: \"ceba82fa-60b8-42d5-af83-7860a39b20a3\") " Nov 25 10:53:46 crc kubenswrapper[4821]: I1125 10:53:46.688146 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ceba82fa-60b8-42d5-af83-7860a39b20a3-ovsdbserver-nb\") pod \"ceba82fa-60b8-42d5-af83-7860a39b20a3\" (UID: \"ceba82fa-60b8-42d5-af83-7860a39b20a3\") " Nov 25 10:53:46 crc kubenswrapper[4821]: I1125 10:53:46.688863 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ceba82fa-60b8-42d5-af83-7860a39b20a3-config\") pod \"ceba82fa-60b8-42d5-af83-7860a39b20a3\" (UID: \"ceba82fa-60b8-42d5-af83-7860a39b20a3\") " Nov 25 10:53:46 crc kubenswrapper[4821]: I1125 10:53:46.688941 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ceba82fa-60b8-42d5-af83-7860a39b20a3-dns-swift-storage-0\") pod \"ceba82fa-60b8-42d5-af83-7860a39b20a3\" (UID: \"ceba82fa-60b8-42d5-af83-7860a39b20a3\") " Nov 25 10:53:46 crc kubenswrapper[4821]: I1125 10:53:46.689030 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q7wqz\" (UniqueName: \"kubernetes.io/projected/ceba82fa-60b8-42d5-af83-7860a39b20a3-kube-api-access-q7wqz\") pod \"ceba82fa-60b8-42d5-af83-7860a39b20a3\" (UID: \"ceba82fa-60b8-42d5-af83-7860a39b20a3\") " Nov 25 10:53:46 crc kubenswrapper[4821]: I1125 10:53:46.693601 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ceba82fa-60b8-42d5-af83-7860a39b20a3-kube-api-access-q7wqz" (OuterVolumeSpecName: "kube-api-access-q7wqz") pod "ceba82fa-60b8-42d5-af83-7860a39b20a3" (UID: "ceba82fa-60b8-42d5-af83-7860a39b20a3"). InnerVolumeSpecName "kube-api-access-q7wqz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:53:46 crc kubenswrapper[4821]: I1125 10:53:46.744735 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ceba82fa-60b8-42d5-af83-7860a39b20a3-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ceba82fa-60b8-42d5-af83-7860a39b20a3" (UID: "ceba82fa-60b8-42d5-af83-7860a39b20a3"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:53:46 crc kubenswrapper[4821]: I1125 10:53:46.746827 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ceba82fa-60b8-42d5-af83-7860a39b20a3-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "ceba82fa-60b8-42d5-af83-7860a39b20a3" (UID: "ceba82fa-60b8-42d5-af83-7860a39b20a3"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:53:46 crc kubenswrapper[4821]: I1125 10:53:46.748050 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ceba82fa-60b8-42d5-af83-7860a39b20a3-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ceba82fa-60b8-42d5-af83-7860a39b20a3" (UID: "ceba82fa-60b8-42d5-af83-7860a39b20a3"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:53:46 crc kubenswrapper[4821]: I1125 10:53:46.753983 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ceba82fa-60b8-42d5-af83-7860a39b20a3-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ceba82fa-60b8-42d5-af83-7860a39b20a3" (UID: "ceba82fa-60b8-42d5-af83-7860a39b20a3"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:53:46 crc kubenswrapper[4821]: I1125 10:53:46.754736 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ceba82fa-60b8-42d5-af83-7860a39b20a3-config" (OuterVolumeSpecName: "config") pod "ceba82fa-60b8-42d5-af83-7860a39b20a3" (UID: "ceba82fa-60b8-42d5-af83-7860a39b20a3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:53:46 crc kubenswrapper[4821]: I1125 10:53:46.755447 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ceba82fa-60b8-42d5-af83-7860a39b20a3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ceba82fa-60b8-42d5-af83-7860a39b20a3" (UID: "ceba82fa-60b8-42d5-af83-7860a39b20a3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:53:46 crc kubenswrapper[4821]: I1125 10:53:46.793281 4821 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ceba82fa-60b8-42d5-af83-7860a39b20a3-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Nov 25 10:53:46 crc kubenswrapper[4821]: I1125 10:53:46.793318 4821 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ceba82fa-60b8-42d5-af83-7860a39b20a3-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 10:53:46 crc kubenswrapper[4821]: I1125 10:53:46.793333 4821 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ceba82fa-60b8-42d5-af83-7860a39b20a3-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 10:53:46 crc kubenswrapper[4821]: I1125 10:53:46.793350 4821 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ceba82fa-60b8-42d5-af83-7860a39b20a3-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 10:53:46 crc kubenswrapper[4821]: I1125 10:53:46.793367 4821 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ceba82fa-60b8-42d5-af83-7860a39b20a3-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:53:46 crc kubenswrapper[4821]: I1125 10:53:46.793380 4821 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ceba82fa-60b8-42d5-af83-7860a39b20a3-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 25 10:53:46 crc kubenswrapper[4821]: I1125 10:53:46.793392 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q7wqz\" (UniqueName: \"kubernetes.io/projected/ceba82fa-60b8-42d5-af83-7860a39b20a3-kube-api-access-q7wqz\") on node \"crc\" DevicePath \"\"" Nov 25 10:53:46 crc kubenswrapper[4821]: I1125 10:53:46.798480 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-dbvrb" event={"ID":"ceba82fa-60b8-42d5-af83-7860a39b20a3","Type":"ContainerDied","Data":"efe3efa1d24b0654ee47ab50d280aee842d3aabd85b35e9d4f1a65717a5eb8b4"} Nov 25 10:53:46 crc kubenswrapper[4821]: I1125 10:53:46.798531 4821 scope.go:117] "RemoveContainer" containerID="355620a153321dcfe92de2ea0a81d2ba2b3470b4d901dd9c1ce6faeded09a7cd" Nov 25 10:53:46 crc kubenswrapper[4821]: I1125 10:53:46.798552 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d558885bc-dbvrb" Nov 25 10:53:46 crc kubenswrapper[4821]: I1125 10:53:46.871856 4821 scope.go:117] "RemoveContainer" containerID="9c700324b5af2a723f00cc2a38925a7dbe3ead006a6a8a654a815f1d9f258055" Nov 25 10:53:46 crc kubenswrapper[4821]: I1125 10:53:46.881206 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-dbvrb"] Nov 25 10:53:46 crc kubenswrapper[4821]: I1125 10:53:46.889574 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-dbvrb"] Nov 25 10:53:48 crc kubenswrapper[4821]: I1125 10:53:48.140758 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ceba82fa-60b8-42d5-af83-7860a39b20a3" path="/var/lib/kubelet/pods/ceba82fa-60b8-42d5-af83-7860a39b20a3/volumes" Nov 25 10:53:55 crc kubenswrapper[4821]: I1125 10:53:55.886709 4821 generic.go:334] "Generic (PLEG): container finished" podID="50c27d16-2f67-4f90-83f0-83e16161de50" containerID="fbb495a1788e29f2e27e80fc8a10837f27ded913875b6651206cbe09b69b1cc6" exitCode=0 Nov 25 10:53:55 crc kubenswrapper[4821]: I1125 10:53:55.886805 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"50c27d16-2f67-4f90-83f0-83e16161de50","Type":"ContainerDied","Data":"fbb495a1788e29f2e27e80fc8a10837f27ded913875b6651206cbe09b69b1cc6"} Nov 25 10:53:56 crc kubenswrapper[4821]: I1125 10:53:56.898591 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"50c27d16-2f67-4f90-83f0-83e16161de50","Type":"ContainerStarted","Data":"8e40ad5324717d57fc7301da9e81eccd5742dfd1f61a7c983a7c59e8c86132bb"} Nov 25 10:53:56 crc kubenswrapper[4821]: I1125 10:53:56.900009 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 25 10:53:56 crc kubenswrapper[4821]: I1125 10:53:56.901093 4821 generic.go:334] "Generic (PLEG): container finished" podID="3556cc6b-7d73-4cbc-9c2e-378c6ba606e4" containerID="6a26c64b2bbe34526964a22113590ca4f5b4d5e3570e85c3d786f20becf68924" exitCode=0 Nov 25 10:53:56 crc kubenswrapper[4821]: I1125 10:53:56.901142 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3556cc6b-7d73-4cbc-9c2e-378c6ba606e4","Type":"ContainerDied","Data":"6a26c64b2bbe34526964a22113590ca4f5b4d5e3570e85c3d786f20becf68924"} Nov 25 10:53:56 crc kubenswrapper[4821]: I1125 10:53:56.929037 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.929021664 podStartE2EDuration="36.929021664s" podCreationTimestamp="2025-11-25 10:53:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:53:56.922745276 +0000 UTC m=+1307.459065133" watchObservedRunningTime="2025-11-25 10:53:56.929021664 +0000 UTC m=+1307.465341511" Nov 25 10:53:57 crc kubenswrapper[4821]: I1125 10:53:57.911995 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3556cc6b-7d73-4cbc-9c2e-378c6ba606e4","Type":"ContainerStarted","Data":"1c7162aa7b7801d9d5abdb98bdc77da21ac509e9a151bf89cd06e3306ea5a805"} Nov 25 10:53:57 crc kubenswrapper[4821]: I1125 10:53:57.912806 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:53:57 crc kubenswrapper[4821]: I1125 10:53:57.946039 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=36.946017094 podStartE2EDuration="36.946017094s" podCreationTimestamp="2025-11-25 10:53:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:53:57.936115146 +0000 UTC m=+1308.472434993" watchObservedRunningTime="2025-11-25 10:53:57.946017094 +0000 UTC m=+1308.482336941" Nov 25 10:53:58 crc kubenswrapper[4821]: I1125 10:53:58.831539 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rdcx6"] Nov 25 10:53:58 crc kubenswrapper[4821]: E1125 10:53:58.832241 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ceba82fa-60b8-42d5-af83-7860a39b20a3" containerName="init" Nov 25 10:53:58 crc kubenswrapper[4821]: I1125 10:53:58.832332 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="ceba82fa-60b8-42d5-af83-7860a39b20a3" containerName="init" Nov 25 10:53:58 crc kubenswrapper[4821]: E1125 10:53:58.832399 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db92740e-3cd9-4b81-b153-3f866fd3db62" containerName="dnsmasq-dns" Nov 25 10:53:58 crc kubenswrapper[4821]: I1125 10:53:58.832463 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="db92740e-3cd9-4b81-b153-3f866fd3db62" containerName="dnsmasq-dns" Nov 25 10:53:58 crc kubenswrapper[4821]: E1125 10:53:58.832549 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db92740e-3cd9-4b81-b153-3f866fd3db62" containerName="init" Nov 25 10:53:58 crc kubenswrapper[4821]: I1125 10:53:58.832602 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="db92740e-3cd9-4b81-b153-3f866fd3db62" containerName="init" Nov 25 10:53:58 crc kubenswrapper[4821]: E1125 10:53:58.832665 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ceba82fa-60b8-42d5-af83-7860a39b20a3" containerName="dnsmasq-dns" Nov 25 10:53:58 crc kubenswrapper[4821]: I1125 10:53:58.832720 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="ceba82fa-60b8-42d5-af83-7860a39b20a3" containerName="dnsmasq-dns" Nov 25 10:53:58 crc kubenswrapper[4821]: I1125 10:53:58.832948 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="db92740e-3cd9-4b81-b153-3f866fd3db62" containerName="dnsmasq-dns" Nov 25 10:53:58 crc kubenswrapper[4821]: I1125 10:53:58.833010 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="ceba82fa-60b8-42d5-af83-7860a39b20a3" containerName="dnsmasq-dns" Nov 25 10:53:58 crc kubenswrapper[4821]: I1125 10:53:58.833856 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rdcx6" Nov 25 10:53:58 crc kubenswrapper[4821]: I1125 10:53:58.835683 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 10:53:58 crc kubenswrapper[4821]: I1125 10:53:58.836215 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 10:53:58 crc kubenswrapper[4821]: I1125 10:53:58.836774 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 10:53:58 crc kubenswrapper[4821]: I1125 10:53:58.837683 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-z764x" Nov 25 10:53:58 crc kubenswrapper[4821]: I1125 10:53:58.848573 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rdcx6"] Nov 25 10:53:58 crc kubenswrapper[4821]: I1125 10:53:58.916507 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lcl6z\" (UniqueName: \"kubernetes.io/projected/a15db0db-c211-40aa-bb3e-8f222669b005-kube-api-access-lcl6z\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-rdcx6\" (UID: \"a15db0db-c211-40aa-bb3e-8f222669b005\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rdcx6" Nov 25 10:53:58 crc kubenswrapper[4821]: I1125 10:53:58.916593 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a15db0db-c211-40aa-bb3e-8f222669b005-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-rdcx6\" (UID: \"a15db0db-c211-40aa-bb3e-8f222669b005\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rdcx6" Nov 25 10:53:58 crc kubenswrapper[4821]: I1125 10:53:58.916617 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a15db0db-c211-40aa-bb3e-8f222669b005-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-rdcx6\" (UID: \"a15db0db-c211-40aa-bb3e-8f222669b005\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rdcx6" Nov 25 10:53:58 crc kubenswrapper[4821]: I1125 10:53:58.916710 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a15db0db-c211-40aa-bb3e-8f222669b005-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-rdcx6\" (UID: \"a15db0db-c211-40aa-bb3e-8f222669b005\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rdcx6" Nov 25 10:53:59 crc kubenswrapper[4821]: I1125 10:53:59.017852 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a15db0db-c211-40aa-bb3e-8f222669b005-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-rdcx6\" (UID: \"a15db0db-c211-40aa-bb3e-8f222669b005\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rdcx6" Nov 25 10:53:59 crc kubenswrapper[4821]: I1125 10:53:59.017970 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lcl6z\" (UniqueName: \"kubernetes.io/projected/a15db0db-c211-40aa-bb3e-8f222669b005-kube-api-access-lcl6z\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-rdcx6\" (UID: \"a15db0db-c211-40aa-bb3e-8f222669b005\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rdcx6" Nov 25 10:53:59 crc kubenswrapper[4821]: I1125 10:53:59.018051 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a15db0db-c211-40aa-bb3e-8f222669b005-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-rdcx6\" (UID: \"a15db0db-c211-40aa-bb3e-8f222669b005\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rdcx6" Nov 25 10:53:59 crc kubenswrapper[4821]: I1125 10:53:59.018079 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a15db0db-c211-40aa-bb3e-8f222669b005-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-rdcx6\" (UID: \"a15db0db-c211-40aa-bb3e-8f222669b005\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rdcx6" Nov 25 10:53:59 crc kubenswrapper[4821]: I1125 10:53:59.022983 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a15db0db-c211-40aa-bb3e-8f222669b005-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-rdcx6\" (UID: \"a15db0db-c211-40aa-bb3e-8f222669b005\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rdcx6" Nov 25 10:53:59 crc kubenswrapper[4821]: I1125 10:53:59.023429 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a15db0db-c211-40aa-bb3e-8f222669b005-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-rdcx6\" (UID: \"a15db0db-c211-40aa-bb3e-8f222669b005\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rdcx6" Nov 25 10:53:59 crc kubenswrapper[4821]: I1125 10:53:59.025274 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a15db0db-c211-40aa-bb3e-8f222669b005-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-rdcx6\" (UID: \"a15db0db-c211-40aa-bb3e-8f222669b005\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rdcx6" Nov 25 10:53:59 crc kubenswrapper[4821]: I1125 10:53:59.033644 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lcl6z\" (UniqueName: \"kubernetes.io/projected/a15db0db-c211-40aa-bb3e-8f222669b005-kube-api-access-lcl6z\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-rdcx6\" (UID: \"a15db0db-c211-40aa-bb3e-8f222669b005\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rdcx6" Nov 25 10:53:59 crc kubenswrapper[4821]: I1125 10:53:59.151010 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rdcx6" Nov 25 10:53:59 crc kubenswrapper[4821]: I1125 10:53:59.669031 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rdcx6"] Nov 25 10:53:59 crc kubenswrapper[4821]: I1125 10:53:59.933647 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rdcx6" event={"ID":"a15db0db-c211-40aa-bb3e-8f222669b005","Type":"ContainerStarted","Data":"ed6dc6cc0f54b82c357dbc787dd7d6808c89b361779062e6a34dee5315252d4f"} Nov 25 10:54:10 crc kubenswrapper[4821]: I1125 10:54:10.725937 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 10:54:11 crc kubenswrapper[4821]: I1125 10:54:11.029180 4821 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="50c27d16-2f67-4f90-83f0-83e16161de50" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.206:5671: connect: connection refused" Nov 25 10:54:11 crc kubenswrapper[4821]: I1125 10:54:11.043223 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rdcx6" event={"ID":"a15db0db-c211-40aa-bb3e-8f222669b005","Type":"ContainerStarted","Data":"0c33be614ff777d0d6810b87f93157d573b43783807b727ff0e5ffa7edcdd9e0"} Nov 25 10:54:12 crc kubenswrapper[4821]: I1125 10:54:12.060592 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:54:12 crc kubenswrapper[4821]: I1125 10:54:12.103463 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rdcx6" podStartSLOduration=3.070523289 podStartE2EDuration="14.103436981s" podCreationTimestamp="2025-11-25 10:53:58 +0000 UTC" firstStartedPulling="2025-11-25 10:53:59.690705588 +0000 UTC m=+1310.227025435" lastFinishedPulling="2025-11-25 10:54:10.72361928 +0000 UTC m=+1321.259939127" observedRunningTime="2025-11-25 10:54:12.067659555 +0000 UTC m=+1322.603979402" watchObservedRunningTime="2025-11-25 10:54:12.103436981 +0000 UTC m=+1322.639756828" Nov 25 10:54:21 crc kubenswrapper[4821]: I1125 10:54:21.030418 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Nov 25 10:54:22 crc kubenswrapper[4821]: I1125 10:54:22.057702 4821 scope.go:117] "RemoveContainer" containerID="6950178aad59d19cd80bb0a741c400699828134cafedb603197165656121141c" Nov 25 10:54:23 crc kubenswrapper[4821]: I1125 10:54:23.166176 4821 generic.go:334] "Generic (PLEG): container finished" podID="a15db0db-c211-40aa-bb3e-8f222669b005" containerID="0c33be614ff777d0d6810b87f93157d573b43783807b727ff0e5ffa7edcdd9e0" exitCode=0 Nov 25 10:54:23 crc kubenswrapper[4821]: I1125 10:54:23.166332 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rdcx6" event={"ID":"a15db0db-c211-40aa-bb3e-8f222669b005","Type":"ContainerDied","Data":"0c33be614ff777d0d6810b87f93157d573b43783807b727ff0e5ffa7edcdd9e0"} Nov 25 10:54:24 crc kubenswrapper[4821]: I1125 10:54:24.613436 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rdcx6" Nov 25 10:54:24 crc kubenswrapper[4821]: I1125 10:54:24.758551 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a15db0db-c211-40aa-bb3e-8f222669b005-inventory\") pod \"a15db0db-c211-40aa-bb3e-8f222669b005\" (UID: \"a15db0db-c211-40aa-bb3e-8f222669b005\") " Nov 25 10:54:24 crc kubenswrapper[4821]: I1125 10:54:24.758615 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a15db0db-c211-40aa-bb3e-8f222669b005-repo-setup-combined-ca-bundle\") pod \"a15db0db-c211-40aa-bb3e-8f222669b005\" (UID: \"a15db0db-c211-40aa-bb3e-8f222669b005\") " Nov 25 10:54:24 crc kubenswrapper[4821]: I1125 10:54:24.758698 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lcl6z\" (UniqueName: \"kubernetes.io/projected/a15db0db-c211-40aa-bb3e-8f222669b005-kube-api-access-lcl6z\") pod \"a15db0db-c211-40aa-bb3e-8f222669b005\" (UID: \"a15db0db-c211-40aa-bb3e-8f222669b005\") " Nov 25 10:54:24 crc kubenswrapper[4821]: I1125 10:54:24.758825 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a15db0db-c211-40aa-bb3e-8f222669b005-ssh-key\") pod \"a15db0db-c211-40aa-bb3e-8f222669b005\" (UID: \"a15db0db-c211-40aa-bb3e-8f222669b005\") " Nov 25 10:54:24 crc kubenswrapper[4821]: I1125 10:54:24.768918 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a15db0db-c211-40aa-bb3e-8f222669b005-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "a15db0db-c211-40aa-bb3e-8f222669b005" (UID: "a15db0db-c211-40aa-bb3e-8f222669b005"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:54:24 crc kubenswrapper[4821]: I1125 10:54:24.768970 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a15db0db-c211-40aa-bb3e-8f222669b005-kube-api-access-lcl6z" (OuterVolumeSpecName: "kube-api-access-lcl6z") pod "a15db0db-c211-40aa-bb3e-8f222669b005" (UID: "a15db0db-c211-40aa-bb3e-8f222669b005"). InnerVolumeSpecName "kube-api-access-lcl6z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:54:24 crc kubenswrapper[4821]: I1125 10:54:24.794595 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a15db0db-c211-40aa-bb3e-8f222669b005-inventory" (OuterVolumeSpecName: "inventory") pod "a15db0db-c211-40aa-bb3e-8f222669b005" (UID: "a15db0db-c211-40aa-bb3e-8f222669b005"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:54:24 crc kubenswrapper[4821]: I1125 10:54:24.801605 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a15db0db-c211-40aa-bb3e-8f222669b005-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a15db0db-c211-40aa-bb3e-8f222669b005" (UID: "a15db0db-c211-40aa-bb3e-8f222669b005"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:54:24 crc kubenswrapper[4821]: I1125 10:54:24.861599 4821 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a15db0db-c211-40aa-bb3e-8f222669b005-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:24 crc kubenswrapper[4821]: I1125 10:54:24.861646 4821 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a15db0db-c211-40aa-bb3e-8f222669b005-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:24 crc kubenswrapper[4821]: I1125 10:54:24.861660 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lcl6z\" (UniqueName: \"kubernetes.io/projected/a15db0db-c211-40aa-bb3e-8f222669b005-kube-api-access-lcl6z\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:24 crc kubenswrapper[4821]: I1125 10:54:24.861669 4821 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a15db0db-c211-40aa-bb3e-8f222669b005-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:25 crc kubenswrapper[4821]: I1125 10:54:25.184902 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rdcx6" event={"ID":"a15db0db-c211-40aa-bb3e-8f222669b005","Type":"ContainerDied","Data":"ed6dc6cc0f54b82c357dbc787dd7d6808c89b361779062e6a34dee5315252d4f"} Nov 25 10:54:25 crc kubenswrapper[4821]: I1125 10:54:25.184953 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-rdcx6" Nov 25 10:54:25 crc kubenswrapper[4821]: I1125 10:54:25.184964 4821 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ed6dc6cc0f54b82c357dbc787dd7d6808c89b361779062e6a34dee5315252d4f" Nov 25 10:54:25 crc kubenswrapper[4821]: I1125 10:54:25.275729 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-9tzcw"] Nov 25 10:54:25 crc kubenswrapper[4821]: E1125 10:54:25.276265 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a15db0db-c211-40aa-bb3e-8f222669b005" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 25 10:54:25 crc kubenswrapper[4821]: I1125 10:54:25.276287 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="a15db0db-c211-40aa-bb3e-8f222669b005" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 25 10:54:25 crc kubenswrapper[4821]: I1125 10:54:25.276523 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="a15db0db-c211-40aa-bb3e-8f222669b005" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 25 10:54:25 crc kubenswrapper[4821]: I1125 10:54:25.277366 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-9tzcw" Nov 25 10:54:25 crc kubenswrapper[4821]: I1125 10:54:25.279543 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 10:54:25 crc kubenswrapper[4821]: I1125 10:54:25.280660 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 10:54:25 crc kubenswrapper[4821]: I1125 10:54:25.280929 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 10:54:25 crc kubenswrapper[4821]: I1125 10:54:25.281750 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-z764x" Nov 25 10:54:25 crc kubenswrapper[4821]: I1125 10:54:25.299661 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-9tzcw"] Nov 25 10:54:25 crc kubenswrapper[4821]: I1125 10:54:25.372235 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/74c5a130-3a40-4920-8e84-e90e9ee4b39f-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-9tzcw\" (UID: \"74c5a130-3a40-4920-8e84-e90e9ee4b39f\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-9tzcw" Nov 25 10:54:25 crc kubenswrapper[4821]: I1125 10:54:25.372289 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/74c5a130-3a40-4920-8e84-e90e9ee4b39f-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-9tzcw\" (UID: \"74c5a130-3a40-4920-8e84-e90e9ee4b39f\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-9tzcw" Nov 25 10:54:25 crc kubenswrapper[4821]: I1125 10:54:25.372349 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rv5mx\" (UniqueName: \"kubernetes.io/projected/74c5a130-3a40-4920-8e84-e90e9ee4b39f-kube-api-access-rv5mx\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-9tzcw\" (UID: \"74c5a130-3a40-4920-8e84-e90e9ee4b39f\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-9tzcw" Nov 25 10:54:25 crc kubenswrapper[4821]: I1125 10:54:25.475002 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/74c5a130-3a40-4920-8e84-e90e9ee4b39f-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-9tzcw\" (UID: \"74c5a130-3a40-4920-8e84-e90e9ee4b39f\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-9tzcw" Nov 25 10:54:25 crc kubenswrapper[4821]: I1125 10:54:25.475047 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/74c5a130-3a40-4920-8e84-e90e9ee4b39f-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-9tzcw\" (UID: \"74c5a130-3a40-4920-8e84-e90e9ee4b39f\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-9tzcw" Nov 25 10:54:25 crc kubenswrapper[4821]: I1125 10:54:25.475120 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rv5mx\" (UniqueName: \"kubernetes.io/projected/74c5a130-3a40-4920-8e84-e90e9ee4b39f-kube-api-access-rv5mx\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-9tzcw\" (UID: \"74c5a130-3a40-4920-8e84-e90e9ee4b39f\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-9tzcw" Nov 25 10:54:25 crc kubenswrapper[4821]: I1125 10:54:25.482192 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/74c5a130-3a40-4920-8e84-e90e9ee4b39f-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-9tzcw\" (UID: \"74c5a130-3a40-4920-8e84-e90e9ee4b39f\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-9tzcw" Nov 25 10:54:25 crc kubenswrapper[4821]: I1125 10:54:25.483003 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/74c5a130-3a40-4920-8e84-e90e9ee4b39f-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-9tzcw\" (UID: \"74c5a130-3a40-4920-8e84-e90e9ee4b39f\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-9tzcw" Nov 25 10:54:25 crc kubenswrapper[4821]: I1125 10:54:25.494962 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rv5mx\" (UniqueName: \"kubernetes.io/projected/74c5a130-3a40-4920-8e84-e90e9ee4b39f-kube-api-access-rv5mx\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-9tzcw\" (UID: \"74c5a130-3a40-4920-8e84-e90e9ee4b39f\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-9tzcw" Nov 25 10:54:25 crc kubenswrapper[4821]: I1125 10:54:25.611554 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-9tzcw" Nov 25 10:54:26 crc kubenswrapper[4821]: I1125 10:54:26.162048 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-9tzcw"] Nov 25 10:54:26 crc kubenswrapper[4821]: W1125 10:54:26.167997 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod74c5a130_3a40_4920_8e84_e90e9ee4b39f.slice/crio-312f1d7ad8aeb0806d4c4c1f28a8a709ffd982e61b9c041923ea64162d0156b4 WatchSource:0}: Error finding container 312f1d7ad8aeb0806d4c4c1f28a8a709ffd982e61b9c041923ea64162d0156b4: Status 404 returned error can't find the container with id 312f1d7ad8aeb0806d4c4c1f28a8a709ffd982e61b9c041923ea64162d0156b4 Nov 25 10:54:26 crc kubenswrapper[4821]: I1125 10:54:26.195720 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-9tzcw" event={"ID":"74c5a130-3a40-4920-8e84-e90e9ee4b39f","Type":"ContainerStarted","Data":"312f1d7ad8aeb0806d4c4c1f28a8a709ffd982e61b9c041923ea64162d0156b4"} Nov 25 10:54:28 crc kubenswrapper[4821]: I1125 10:54:28.215886 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-9tzcw" event={"ID":"74c5a130-3a40-4920-8e84-e90e9ee4b39f","Type":"ContainerStarted","Data":"1fcf6908ab66c9eee6913ed7b780f3bbcfcf753f1954c4797040fad031ebcc7e"} Nov 25 10:54:28 crc kubenswrapper[4821]: I1125 10:54:28.237922 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-9tzcw" podStartSLOduration=2.227348344 podStartE2EDuration="3.237898474s" podCreationTimestamp="2025-11-25 10:54:25 +0000 UTC" firstStartedPulling="2025-11-25 10:54:26.170419193 +0000 UTC m=+1336.706739040" lastFinishedPulling="2025-11-25 10:54:27.180969323 +0000 UTC m=+1337.717289170" observedRunningTime="2025-11-25 10:54:28.233849373 +0000 UTC m=+1338.770169220" watchObservedRunningTime="2025-11-25 10:54:28.237898474 +0000 UTC m=+1338.774218321" Nov 25 10:54:30 crc kubenswrapper[4821]: I1125 10:54:30.239474 4821 generic.go:334] "Generic (PLEG): container finished" podID="74c5a130-3a40-4920-8e84-e90e9ee4b39f" containerID="1fcf6908ab66c9eee6913ed7b780f3bbcfcf753f1954c4797040fad031ebcc7e" exitCode=0 Nov 25 10:54:30 crc kubenswrapper[4821]: I1125 10:54:30.239586 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-9tzcw" event={"ID":"74c5a130-3a40-4920-8e84-e90e9ee4b39f","Type":"ContainerDied","Data":"1fcf6908ab66c9eee6913ed7b780f3bbcfcf753f1954c4797040fad031ebcc7e"} Nov 25 10:54:31 crc kubenswrapper[4821]: I1125 10:54:31.673326 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-9tzcw" Nov 25 10:54:31 crc kubenswrapper[4821]: I1125 10:54:31.801382 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rv5mx\" (UniqueName: \"kubernetes.io/projected/74c5a130-3a40-4920-8e84-e90e9ee4b39f-kube-api-access-rv5mx\") pod \"74c5a130-3a40-4920-8e84-e90e9ee4b39f\" (UID: \"74c5a130-3a40-4920-8e84-e90e9ee4b39f\") " Nov 25 10:54:31 crc kubenswrapper[4821]: I1125 10:54:31.801601 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/74c5a130-3a40-4920-8e84-e90e9ee4b39f-ssh-key\") pod \"74c5a130-3a40-4920-8e84-e90e9ee4b39f\" (UID: \"74c5a130-3a40-4920-8e84-e90e9ee4b39f\") " Nov 25 10:54:31 crc kubenswrapper[4821]: I1125 10:54:31.801682 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/74c5a130-3a40-4920-8e84-e90e9ee4b39f-inventory\") pod \"74c5a130-3a40-4920-8e84-e90e9ee4b39f\" (UID: \"74c5a130-3a40-4920-8e84-e90e9ee4b39f\") " Nov 25 10:54:31 crc kubenswrapper[4821]: I1125 10:54:31.807917 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74c5a130-3a40-4920-8e84-e90e9ee4b39f-kube-api-access-rv5mx" (OuterVolumeSpecName: "kube-api-access-rv5mx") pod "74c5a130-3a40-4920-8e84-e90e9ee4b39f" (UID: "74c5a130-3a40-4920-8e84-e90e9ee4b39f"). InnerVolumeSpecName "kube-api-access-rv5mx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:54:31 crc kubenswrapper[4821]: I1125 10:54:31.834340 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74c5a130-3a40-4920-8e84-e90e9ee4b39f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "74c5a130-3a40-4920-8e84-e90e9ee4b39f" (UID: "74c5a130-3a40-4920-8e84-e90e9ee4b39f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:54:31 crc kubenswrapper[4821]: I1125 10:54:31.835399 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74c5a130-3a40-4920-8e84-e90e9ee4b39f-inventory" (OuterVolumeSpecName: "inventory") pod "74c5a130-3a40-4920-8e84-e90e9ee4b39f" (UID: "74c5a130-3a40-4920-8e84-e90e9ee4b39f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:54:31 crc kubenswrapper[4821]: I1125 10:54:31.904240 4821 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/74c5a130-3a40-4920-8e84-e90e9ee4b39f-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:31 crc kubenswrapper[4821]: I1125 10:54:31.904275 4821 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/74c5a130-3a40-4920-8e84-e90e9ee4b39f-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:31 crc kubenswrapper[4821]: I1125 10:54:31.904284 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rv5mx\" (UniqueName: \"kubernetes.io/projected/74c5a130-3a40-4920-8e84-e90e9ee4b39f-kube-api-access-rv5mx\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:32 crc kubenswrapper[4821]: I1125 10:54:32.258257 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-9tzcw" event={"ID":"74c5a130-3a40-4920-8e84-e90e9ee4b39f","Type":"ContainerDied","Data":"312f1d7ad8aeb0806d4c4c1f28a8a709ffd982e61b9c041923ea64162d0156b4"} Nov 25 10:54:32 crc kubenswrapper[4821]: I1125 10:54:32.258301 4821 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="312f1d7ad8aeb0806d4c4c1f28a8a709ffd982e61b9c041923ea64162d0156b4" Nov 25 10:54:32 crc kubenswrapper[4821]: I1125 10:54:32.258308 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-9tzcw" Nov 25 10:54:32 crc kubenswrapper[4821]: I1125 10:54:32.343894 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pkx5r"] Nov 25 10:54:32 crc kubenswrapper[4821]: E1125 10:54:32.345400 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74c5a130-3a40-4920-8e84-e90e9ee4b39f" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Nov 25 10:54:32 crc kubenswrapper[4821]: I1125 10:54:32.345469 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="74c5a130-3a40-4920-8e84-e90e9ee4b39f" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Nov 25 10:54:32 crc kubenswrapper[4821]: I1125 10:54:32.346034 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="74c5a130-3a40-4920-8e84-e90e9ee4b39f" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Nov 25 10:54:32 crc kubenswrapper[4821]: I1125 10:54:32.347750 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pkx5r" Nov 25 10:54:32 crc kubenswrapper[4821]: I1125 10:54:32.350609 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 10:54:32 crc kubenswrapper[4821]: I1125 10:54:32.350625 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-z764x" Nov 25 10:54:32 crc kubenswrapper[4821]: I1125 10:54:32.353827 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 10:54:32 crc kubenswrapper[4821]: I1125 10:54:32.354603 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 10:54:32 crc kubenswrapper[4821]: I1125 10:54:32.356516 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pkx5r"] Nov 25 10:54:32 crc kubenswrapper[4821]: I1125 10:54:32.515553 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b93320f-a41f-4c0e-874f-b4f945f14c90-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-pkx5r\" (UID: \"0b93320f-a41f-4c0e-874f-b4f945f14c90\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pkx5r" Nov 25 10:54:32 crc kubenswrapper[4821]: I1125 10:54:32.515661 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdg26\" (UniqueName: \"kubernetes.io/projected/0b93320f-a41f-4c0e-874f-b4f945f14c90-kube-api-access-vdg26\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-pkx5r\" (UID: \"0b93320f-a41f-4c0e-874f-b4f945f14c90\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pkx5r" Nov 25 10:54:32 crc kubenswrapper[4821]: I1125 10:54:32.515714 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0b93320f-a41f-4c0e-874f-b4f945f14c90-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-pkx5r\" (UID: \"0b93320f-a41f-4c0e-874f-b4f945f14c90\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pkx5r" Nov 25 10:54:32 crc kubenswrapper[4821]: I1125 10:54:32.515867 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0b93320f-a41f-4c0e-874f-b4f945f14c90-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-pkx5r\" (UID: \"0b93320f-a41f-4c0e-874f-b4f945f14c90\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pkx5r" Nov 25 10:54:32 crc kubenswrapper[4821]: I1125 10:54:32.618434 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b93320f-a41f-4c0e-874f-b4f945f14c90-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-pkx5r\" (UID: \"0b93320f-a41f-4c0e-874f-b4f945f14c90\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pkx5r" Nov 25 10:54:32 crc kubenswrapper[4821]: I1125 10:54:32.618514 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdg26\" (UniqueName: \"kubernetes.io/projected/0b93320f-a41f-4c0e-874f-b4f945f14c90-kube-api-access-vdg26\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-pkx5r\" (UID: \"0b93320f-a41f-4c0e-874f-b4f945f14c90\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pkx5r" Nov 25 10:54:32 crc kubenswrapper[4821]: I1125 10:54:32.618570 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0b93320f-a41f-4c0e-874f-b4f945f14c90-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-pkx5r\" (UID: \"0b93320f-a41f-4c0e-874f-b4f945f14c90\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pkx5r" Nov 25 10:54:32 crc kubenswrapper[4821]: I1125 10:54:32.618650 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0b93320f-a41f-4c0e-874f-b4f945f14c90-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-pkx5r\" (UID: \"0b93320f-a41f-4c0e-874f-b4f945f14c90\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pkx5r" Nov 25 10:54:32 crc kubenswrapper[4821]: I1125 10:54:32.622862 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0b93320f-a41f-4c0e-874f-b4f945f14c90-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-pkx5r\" (UID: \"0b93320f-a41f-4c0e-874f-b4f945f14c90\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pkx5r" Nov 25 10:54:32 crc kubenswrapper[4821]: I1125 10:54:32.624866 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0b93320f-a41f-4c0e-874f-b4f945f14c90-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-pkx5r\" (UID: \"0b93320f-a41f-4c0e-874f-b4f945f14c90\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pkx5r" Nov 25 10:54:32 crc kubenswrapper[4821]: I1125 10:54:32.626029 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b93320f-a41f-4c0e-874f-b4f945f14c90-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-pkx5r\" (UID: \"0b93320f-a41f-4c0e-874f-b4f945f14c90\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pkx5r" Nov 25 10:54:32 crc kubenswrapper[4821]: I1125 10:54:32.643199 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdg26\" (UniqueName: \"kubernetes.io/projected/0b93320f-a41f-4c0e-874f-b4f945f14c90-kube-api-access-vdg26\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-pkx5r\" (UID: \"0b93320f-a41f-4c0e-874f-b4f945f14c90\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pkx5r" Nov 25 10:54:32 crc kubenswrapper[4821]: I1125 10:54:32.667736 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pkx5r" Nov 25 10:54:33 crc kubenswrapper[4821]: I1125 10:54:33.342556 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pkx5r"] Nov 25 10:54:34 crc kubenswrapper[4821]: I1125 10:54:34.281446 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pkx5r" event={"ID":"0b93320f-a41f-4c0e-874f-b4f945f14c90","Type":"ContainerStarted","Data":"d7f4a85585c2000173582fdcd64f7070858b4170674fd53298a6c7eacd454786"} Nov 25 10:54:34 crc kubenswrapper[4821]: I1125 10:54:34.714661 4821 patch_prober.go:28] interesting pod/machine-config-daemon-2krbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:54:34 crc kubenswrapper[4821]: I1125 10:54:34.715040 4821 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:54:35 crc kubenswrapper[4821]: I1125 10:54:35.289865 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pkx5r" event={"ID":"0b93320f-a41f-4c0e-874f-b4f945f14c90","Type":"ContainerStarted","Data":"39033e10b4d59c74ae7b891bd41f3fdc0693d727c43e7840f122c90546b40d80"} Nov 25 10:54:35 crc kubenswrapper[4821]: I1125 10:54:35.319239 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pkx5r" podStartSLOduration=2.631474778 podStartE2EDuration="3.31921492s" podCreationTimestamp="2025-11-25 10:54:32 +0000 UTC" firstStartedPulling="2025-11-25 10:54:33.352220956 +0000 UTC m=+1343.888540803" lastFinishedPulling="2025-11-25 10:54:34.039961098 +0000 UTC m=+1344.576280945" observedRunningTime="2025-11-25 10:54:35.310495141 +0000 UTC m=+1345.846814988" watchObservedRunningTime="2025-11-25 10:54:35.31921492 +0000 UTC m=+1345.855534767" Nov 25 10:55:04 crc kubenswrapper[4821]: I1125 10:55:04.714998 4821 patch_prober.go:28] interesting pod/machine-config-daemon-2krbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:55:04 crc kubenswrapper[4821]: I1125 10:55:04.715630 4821 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:55:22 crc kubenswrapper[4821]: I1125 10:55:22.192277 4821 scope.go:117] "RemoveContainer" containerID="ac7e7a6e7fa3583d81e45e8d909f8b18b87b37f18db9fdb40be515f4278bf153" Nov 25 10:55:34 crc kubenswrapper[4821]: I1125 10:55:34.714429 4821 patch_prober.go:28] interesting pod/machine-config-daemon-2krbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:55:34 crc kubenswrapper[4821]: I1125 10:55:34.714889 4821 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:55:34 crc kubenswrapper[4821]: I1125 10:55:34.714933 4821 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" Nov 25 10:55:34 crc kubenswrapper[4821]: I1125 10:55:34.715585 4821 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1cba28a33f40c656b6dedea9b9df6fa03370b5e6b1aff86ef7fcc571ea800429"} pod="openshift-machine-config-operator/machine-config-daemon-2krbf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 10:55:34 crc kubenswrapper[4821]: I1125 10:55:34.715638 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerName="machine-config-daemon" containerID="cri-o://1cba28a33f40c656b6dedea9b9df6fa03370b5e6b1aff86ef7fcc571ea800429" gracePeriod=600 Nov 25 10:55:35 crc kubenswrapper[4821]: I1125 10:55:35.875443 4821 generic.go:334] "Generic (PLEG): container finished" podID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerID="1cba28a33f40c656b6dedea9b9df6fa03370b5e6b1aff86ef7fcc571ea800429" exitCode=0 Nov 25 10:55:35 crc kubenswrapper[4821]: I1125 10:55:35.875514 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" event={"ID":"5f948b87-ac86-4de6-ad64-c2ef947f84d4","Type":"ContainerDied","Data":"1cba28a33f40c656b6dedea9b9df6fa03370b5e6b1aff86ef7fcc571ea800429"} Nov 25 10:55:35 crc kubenswrapper[4821]: I1125 10:55:35.876024 4821 scope.go:117] "RemoveContainer" containerID="8e91d42cd72af932de083af71bf0c37ad5242c32b692e67d8b88f724fd1c26a8" Nov 25 10:55:36 crc kubenswrapper[4821]: I1125 10:55:36.888142 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" event={"ID":"5f948b87-ac86-4de6-ad64-c2ef947f84d4","Type":"ContainerStarted","Data":"fc16b2ea0cd36da4cef842d21755ec621134a4c21c5c964f8856fb8e6a102b2d"} Nov 25 10:55:48 crc kubenswrapper[4821]: I1125 10:55:48.532351 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-n2t4m"] Nov 25 10:55:48 crc kubenswrapper[4821]: I1125 10:55:48.535253 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n2t4m" Nov 25 10:55:48 crc kubenswrapper[4821]: I1125 10:55:48.541749 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-n2t4m"] Nov 25 10:55:48 crc kubenswrapper[4821]: I1125 10:55:48.628040 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-76g2g\" (UniqueName: \"kubernetes.io/projected/e811afdd-3d66-4b5c-a4f6-9266f2238ee5-kube-api-access-76g2g\") pod \"redhat-operators-n2t4m\" (UID: \"e811afdd-3d66-4b5c-a4f6-9266f2238ee5\") " pod="openshift-marketplace/redhat-operators-n2t4m" Nov 25 10:55:48 crc kubenswrapper[4821]: I1125 10:55:48.628362 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e811afdd-3d66-4b5c-a4f6-9266f2238ee5-catalog-content\") pod \"redhat-operators-n2t4m\" (UID: \"e811afdd-3d66-4b5c-a4f6-9266f2238ee5\") " pod="openshift-marketplace/redhat-operators-n2t4m" Nov 25 10:55:48 crc kubenswrapper[4821]: I1125 10:55:48.628460 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e811afdd-3d66-4b5c-a4f6-9266f2238ee5-utilities\") pod \"redhat-operators-n2t4m\" (UID: \"e811afdd-3d66-4b5c-a4f6-9266f2238ee5\") " pod="openshift-marketplace/redhat-operators-n2t4m" Nov 25 10:55:48 crc kubenswrapper[4821]: I1125 10:55:48.730068 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e811afdd-3d66-4b5c-a4f6-9266f2238ee5-utilities\") pod \"redhat-operators-n2t4m\" (UID: \"e811afdd-3d66-4b5c-a4f6-9266f2238ee5\") " pod="openshift-marketplace/redhat-operators-n2t4m" Nov 25 10:55:48 crc kubenswrapper[4821]: I1125 10:55:48.730147 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-76g2g\" (UniqueName: \"kubernetes.io/projected/e811afdd-3d66-4b5c-a4f6-9266f2238ee5-kube-api-access-76g2g\") pod \"redhat-operators-n2t4m\" (UID: \"e811afdd-3d66-4b5c-a4f6-9266f2238ee5\") " pod="openshift-marketplace/redhat-operators-n2t4m" Nov 25 10:55:48 crc kubenswrapper[4821]: I1125 10:55:48.730249 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e811afdd-3d66-4b5c-a4f6-9266f2238ee5-catalog-content\") pod \"redhat-operators-n2t4m\" (UID: \"e811afdd-3d66-4b5c-a4f6-9266f2238ee5\") " pod="openshift-marketplace/redhat-operators-n2t4m" Nov 25 10:55:48 crc kubenswrapper[4821]: I1125 10:55:48.730711 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e811afdd-3d66-4b5c-a4f6-9266f2238ee5-catalog-content\") pod \"redhat-operators-n2t4m\" (UID: \"e811afdd-3d66-4b5c-a4f6-9266f2238ee5\") " pod="openshift-marketplace/redhat-operators-n2t4m" Nov 25 10:55:48 crc kubenswrapper[4821]: I1125 10:55:48.730845 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e811afdd-3d66-4b5c-a4f6-9266f2238ee5-utilities\") pod \"redhat-operators-n2t4m\" (UID: \"e811afdd-3d66-4b5c-a4f6-9266f2238ee5\") " pod="openshift-marketplace/redhat-operators-n2t4m" Nov 25 10:55:48 crc kubenswrapper[4821]: I1125 10:55:48.748636 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-76g2g\" (UniqueName: \"kubernetes.io/projected/e811afdd-3d66-4b5c-a4f6-9266f2238ee5-kube-api-access-76g2g\") pod \"redhat-operators-n2t4m\" (UID: \"e811afdd-3d66-4b5c-a4f6-9266f2238ee5\") " pod="openshift-marketplace/redhat-operators-n2t4m" Nov 25 10:55:48 crc kubenswrapper[4821]: I1125 10:55:48.856870 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n2t4m" Nov 25 10:55:49 crc kubenswrapper[4821]: I1125 10:55:49.372415 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-n2t4m"] Nov 25 10:55:50 crc kubenswrapper[4821]: I1125 10:55:50.029487 4821 generic.go:334] "Generic (PLEG): container finished" podID="e811afdd-3d66-4b5c-a4f6-9266f2238ee5" containerID="373d515201409fab2b41442a7f933b5abc0c76d9d89e13798d75f8d0a0cc2d7c" exitCode=0 Nov 25 10:55:50 crc kubenswrapper[4821]: I1125 10:55:50.029543 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n2t4m" event={"ID":"e811afdd-3d66-4b5c-a4f6-9266f2238ee5","Type":"ContainerDied","Data":"373d515201409fab2b41442a7f933b5abc0c76d9d89e13798d75f8d0a0cc2d7c"} Nov 25 10:55:50 crc kubenswrapper[4821]: I1125 10:55:50.029868 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n2t4m" event={"ID":"e811afdd-3d66-4b5c-a4f6-9266f2238ee5","Type":"ContainerStarted","Data":"9330b889c8110754fb197f9cc6b3d8786cde11cc179fd9eeec407a9564afe3fa"} Nov 25 10:55:52 crc kubenswrapper[4821]: I1125 10:55:52.050079 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n2t4m" event={"ID":"e811afdd-3d66-4b5c-a4f6-9266f2238ee5","Type":"ContainerStarted","Data":"ef5f8590b3b0eb6535efe701c5a49f4b42709a107e7d5c648e288110afd3f075"} Nov 25 10:55:53 crc kubenswrapper[4821]: I1125 10:55:53.060648 4821 generic.go:334] "Generic (PLEG): container finished" podID="e811afdd-3d66-4b5c-a4f6-9266f2238ee5" containerID="ef5f8590b3b0eb6535efe701c5a49f4b42709a107e7d5c648e288110afd3f075" exitCode=0 Nov 25 10:55:53 crc kubenswrapper[4821]: I1125 10:55:53.060763 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n2t4m" event={"ID":"e811afdd-3d66-4b5c-a4f6-9266f2238ee5","Type":"ContainerDied","Data":"ef5f8590b3b0eb6535efe701c5a49f4b42709a107e7d5c648e288110afd3f075"} Nov 25 10:55:55 crc kubenswrapper[4821]: I1125 10:55:55.079669 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n2t4m" event={"ID":"e811afdd-3d66-4b5c-a4f6-9266f2238ee5","Type":"ContainerStarted","Data":"27e4c0a0ff9e9f613e065ee7515e4cf26b973b3c121e87dba2e9f543db189f7e"} Nov 25 10:55:55 crc kubenswrapper[4821]: I1125 10:55:55.105896 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-n2t4m" podStartSLOduration=3.225543424 podStartE2EDuration="7.105879337s" podCreationTimestamp="2025-11-25 10:55:48 +0000 UTC" firstStartedPulling="2025-11-25 10:55:50.031709182 +0000 UTC m=+1420.568029029" lastFinishedPulling="2025-11-25 10:55:53.912045095 +0000 UTC m=+1424.448364942" observedRunningTime="2025-11-25 10:55:55.096896363 +0000 UTC m=+1425.633216220" watchObservedRunningTime="2025-11-25 10:55:55.105879337 +0000 UTC m=+1425.642199184" Nov 25 10:55:58 crc kubenswrapper[4821]: I1125 10:55:58.857331 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-n2t4m" Nov 25 10:55:58 crc kubenswrapper[4821]: I1125 10:55:58.857906 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-n2t4m" Nov 25 10:55:59 crc kubenswrapper[4821]: I1125 10:55:59.914494 4821 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-n2t4m" podUID="e811afdd-3d66-4b5c-a4f6-9266f2238ee5" containerName="registry-server" probeResult="failure" output=< Nov 25 10:55:59 crc kubenswrapper[4821]: timeout: failed to connect service ":50051" within 1s Nov 25 10:55:59 crc kubenswrapper[4821]: > Nov 25 10:56:08 crc kubenswrapper[4821]: I1125 10:56:08.914071 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-n2t4m" Nov 25 10:56:08 crc kubenswrapper[4821]: I1125 10:56:08.967794 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-n2t4m" Nov 25 10:56:09 crc kubenswrapper[4821]: I1125 10:56:09.150326 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-n2t4m"] Nov 25 10:56:10 crc kubenswrapper[4821]: I1125 10:56:10.210910 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-n2t4m" podUID="e811afdd-3d66-4b5c-a4f6-9266f2238ee5" containerName="registry-server" containerID="cri-o://27e4c0a0ff9e9f613e065ee7515e4cf26b973b3c121e87dba2e9f543db189f7e" gracePeriod=2 Nov 25 10:56:10 crc kubenswrapper[4821]: I1125 10:56:10.719824 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n2t4m" Nov 25 10:56:10 crc kubenswrapper[4821]: I1125 10:56:10.867982 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-76g2g\" (UniqueName: \"kubernetes.io/projected/e811afdd-3d66-4b5c-a4f6-9266f2238ee5-kube-api-access-76g2g\") pod \"e811afdd-3d66-4b5c-a4f6-9266f2238ee5\" (UID: \"e811afdd-3d66-4b5c-a4f6-9266f2238ee5\") " Nov 25 10:56:10 crc kubenswrapper[4821]: I1125 10:56:10.868052 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e811afdd-3d66-4b5c-a4f6-9266f2238ee5-catalog-content\") pod \"e811afdd-3d66-4b5c-a4f6-9266f2238ee5\" (UID: \"e811afdd-3d66-4b5c-a4f6-9266f2238ee5\") " Nov 25 10:56:10 crc kubenswrapper[4821]: I1125 10:56:10.868147 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e811afdd-3d66-4b5c-a4f6-9266f2238ee5-utilities\") pod \"e811afdd-3d66-4b5c-a4f6-9266f2238ee5\" (UID: \"e811afdd-3d66-4b5c-a4f6-9266f2238ee5\") " Nov 25 10:56:10 crc kubenswrapper[4821]: I1125 10:56:10.869139 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e811afdd-3d66-4b5c-a4f6-9266f2238ee5-utilities" (OuterVolumeSpecName: "utilities") pod "e811afdd-3d66-4b5c-a4f6-9266f2238ee5" (UID: "e811afdd-3d66-4b5c-a4f6-9266f2238ee5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:56:10 crc kubenswrapper[4821]: I1125 10:56:10.874058 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e811afdd-3d66-4b5c-a4f6-9266f2238ee5-kube-api-access-76g2g" (OuterVolumeSpecName: "kube-api-access-76g2g") pod "e811afdd-3d66-4b5c-a4f6-9266f2238ee5" (UID: "e811afdd-3d66-4b5c-a4f6-9266f2238ee5"). InnerVolumeSpecName "kube-api-access-76g2g". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:56:10 crc kubenswrapper[4821]: I1125 10:56:10.970554 4821 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e811afdd-3d66-4b5c-a4f6-9266f2238ee5-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:10 crc kubenswrapper[4821]: I1125 10:56:10.970594 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-76g2g\" (UniqueName: \"kubernetes.io/projected/e811afdd-3d66-4b5c-a4f6-9266f2238ee5-kube-api-access-76g2g\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:10 crc kubenswrapper[4821]: I1125 10:56:10.973708 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e811afdd-3d66-4b5c-a4f6-9266f2238ee5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e811afdd-3d66-4b5c-a4f6-9266f2238ee5" (UID: "e811afdd-3d66-4b5c-a4f6-9266f2238ee5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:56:11 crc kubenswrapper[4821]: I1125 10:56:11.072069 4821 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e811afdd-3d66-4b5c-a4f6-9266f2238ee5-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:11 crc kubenswrapper[4821]: I1125 10:56:11.225525 4821 generic.go:334] "Generic (PLEG): container finished" podID="e811afdd-3d66-4b5c-a4f6-9266f2238ee5" containerID="27e4c0a0ff9e9f613e065ee7515e4cf26b973b3c121e87dba2e9f543db189f7e" exitCode=0 Nov 25 10:56:11 crc kubenswrapper[4821]: I1125 10:56:11.225587 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n2t4m" Nov 25 10:56:11 crc kubenswrapper[4821]: I1125 10:56:11.225636 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n2t4m" event={"ID":"e811afdd-3d66-4b5c-a4f6-9266f2238ee5","Type":"ContainerDied","Data":"27e4c0a0ff9e9f613e065ee7515e4cf26b973b3c121e87dba2e9f543db189f7e"} Nov 25 10:56:11 crc kubenswrapper[4821]: I1125 10:56:11.225982 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n2t4m" event={"ID":"e811afdd-3d66-4b5c-a4f6-9266f2238ee5","Type":"ContainerDied","Data":"9330b889c8110754fb197f9cc6b3d8786cde11cc179fd9eeec407a9564afe3fa"} Nov 25 10:56:11 crc kubenswrapper[4821]: I1125 10:56:11.226009 4821 scope.go:117] "RemoveContainer" containerID="27e4c0a0ff9e9f613e065ee7515e4cf26b973b3c121e87dba2e9f543db189f7e" Nov 25 10:56:11 crc kubenswrapper[4821]: I1125 10:56:11.255147 4821 scope.go:117] "RemoveContainer" containerID="ef5f8590b3b0eb6535efe701c5a49f4b42709a107e7d5c648e288110afd3f075" Nov 25 10:56:11 crc kubenswrapper[4821]: I1125 10:56:11.287103 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-n2t4m"] Nov 25 10:56:11 crc kubenswrapper[4821]: I1125 10:56:11.293867 4821 scope.go:117] "RemoveContainer" containerID="373d515201409fab2b41442a7f933b5abc0c76d9d89e13798d75f8d0a0cc2d7c" Nov 25 10:56:11 crc kubenswrapper[4821]: I1125 10:56:11.297424 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-n2t4m"] Nov 25 10:56:11 crc kubenswrapper[4821]: I1125 10:56:11.332412 4821 scope.go:117] "RemoveContainer" containerID="27e4c0a0ff9e9f613e065ee7515e4cf26b973b3c121e87dba2e9f543db189f7e" Nov 25 10:56:11 crc kubenswrapper[4821]: E1125 10:56:11.336413 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"27e4c0a0ff9e9f613e065ee7515e4cf26b973b3c121e87dba2e9f543db189f7e\": container with ID starting with 27e4c0a0ff9e9f613e065ee7515e4cf26b973b3c121e87dba2e9f543db189f7e not found: ID does not exist" containerID="27e4c0a0ff9e9f613e065ee7515e4cf26b973b3c121e87dba2e9f543db189f7e" Nov 25 10:56:11 crc kubenswrapper[4821]: I1125 10:56:11.336483 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27e4c0a0ff9e9f613e065ee7515e4cf26b973b3c121e87dba2e9f543db189f7e"} err="failed to get container status \"27e4c0a0ff9e9f613e065ee7515e4cf26b973b3c121e87dba2e9f543db189f7e\": rpc error: code = NotFound desc = could not find container \"27e4c0a0ff9e9f613e065ee7515e4cf26b973b3c121e87dba2e9f543db189f7e\": container with ID starting with 27e4c0a0ff9e9f613e065ee7515e4cf26b973b3c121e87dba2e9f543db189f7e not found: ID does not exist" Nov 25 10:56:11 crc kubenswrapper[4821]: I1125 10:56:11.336520 4821 scope.go:117] "RemoveContainer" containerID="ef5f8590b3b0eb6535efe701c5a49f4b42709a107e7d5c648e288110afd3f075" Nov 25 10:56:11 crc kubenswrapper[4821]: E1125 10:56:11.338495 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef5f8590b3b0eb6535efe701c5a49f4b42709a107e7d5c648e288110afd3f075\": container with ID starting with ef5f8590b3b0eb6535efe701c5a49f4b42709a107e7d5c648e288110afd3f075 not found: ID does not exist" containerID="ef5f8590b3b0eb6535efe701c5a49f4b42709a107e7d5c648e288110afd3f075" Nov 25 10:56:11 crc kubenswrapper[4821]: I1125 10:56:11.338541 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef5f8590b3b0eb6535efe701c5a49f4b42709a107e7d5c648e288110afd3f075"} err="failed to get container status \"ef5f8590b3b0eb6535efe701c5a49f4b42709a107e7d5c648e288110afd3f075\": rpc error: code = NotFound desc = could not find container \"ef5f8590b3b0eb6535efe701c5a49f4b42709a107e7d5c648e288110afd3f075\": container with ID starting with ef5f8590b3b0eb6535efe701c5a49f4b42709a107e7d5c648e288110afd3f075 not found: ID does not exist" Nov 25 10:56:11 crc kubenswrapper[4821]: I1125 10:56:11.338572 4821 scope.go:117] "RemoveContainer" containerID="373d515201409fab2b41442a7f933b5abc0c76d9d89e13798d75f8d0a0cc2d7c" Nov 25 10:56:11 crc kubenswrapper[4821]: E1125 10:56:11.339114 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"373d515201409fab2b41442a7f933b5abc0c76d9d89e13798d75f8d0a0cc2d7c\": container with ID starting with 373d515201409fab2b41442a7f933b5abc0c76d9d89e13798d75f8d0a0cc2d7c not found: ID does not exist" containerID="373d515201409fab2b41442a7f933b5abc0c76d9d89e13798d75f8d0a0cc2d7c" Nov 25 10:56:11 crc kubenswrapper[4821]: I1125 10:56:11.339154 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"373d515201409fab2b41442a7f933b5abc0c76d9d89e13798d75f8d0a0cc2d7c"} err="failed to get container status \"373d515201409fab2b41442a7f933b5abc0c76d9d89e13798d75f8d0a0cc2d7c\": rpc error: code = NotFound desc = could not find container \"373d515201409fab2b41442a7f933b5abc0c76d9d89e13798d75f8d0a0cc2d7c\": container with ID starting with 373d515201409fab2b41442a7f933b5abc0c76d9d89e13798d75f8d0a0cc2d7c not found: ID does not exist" Nov 25 10:56:12 crc kubenswrapper[4821]: I1125 10:56:12.128200 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e811afdd-3d66-4b5c-a4f6-9266f2238ee5" path="/var/lib/kubelet/pods/e811afdd-3d66-4b5c-a4f6-9266f2238ee5/volumes" Nov 25 10:56:13 crc kubenswrapper[4821]: I1125 10:56:13.567298 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-scglz"] Nov 25 10:56:13 crc kubenswrapper[4821]: E1125 10:56:13.568610 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e811afdd-3d66-4b5c-a4f6-9266f2238ee5" containerName="extract-content" Nov 25 10:56:13 crc kubenswrapper[4821]: I1125 10:56:13.568636 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="e811afdd-3d66-4b5c-a4f6-9266f2238ee5" containerName="extract-content" Nov 25 10:56:13 crc kubenswrapper[4821]: E1125 10:56:13.568665 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e811afdd-3d66-4b5c-a4f6-9266f2238ee5" containerName="registry-server" Nov 25 10:56:13 crc kubenswrapper[4821]: I1125 10:56:13.568674 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="e811afdd-3d66-4b5c-a4f6-9266f2238ee5" containerName="registry-server" Nov 25 10:56:13 crc kubenswrapper[4821]: E1125 10:56:13.568689 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e811afdd-3d66-4b5c-a4f6-9266f2238ee5" containerName="extract-utilities" Nov 25 10:56:13 crc kubenswrapper[4821]: I1125 10:56:13.568695 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="e811afdd-3d66-4b5c-a4f6-9266f2238ee5" containerName="extract-utilities" Nov 25 10:56:13 crc kubenswrapper[4821]: I1125 10:56:13.568877 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="e811afdd-3d66-4b5c-a4f6-9266f2238ee5" containerName="registry-server" Nov 25 10:56:13 crc kubenswrapper[4821]: I1125 10:56:13.570211 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-scglz" Nov 25 10:56:13 crc kubenswrapper[4821]: I1125 10:56:13.582292 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-scglz"] Nov 25 10:56:13 crc kubenswrapper[4821]: I1125 10:56:13.725433 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0566c97e-c605-497d-a28f-080d8729a619-catalog-content\") pod \"redhat-marketplace-scglz\" (UID: \"0566c97e-c605-497d-a28f-080d8729a619\") " pod="openshift-marketplace/redhat-marketplace-scglz" Nov 25 10:56:13 crc kubenswrapper[4821]: I1125 10:56:13.725556 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nhz8g\" (UniqueName: \"kubernetes.io/projected/0566c97e-c605-497d-a28f-080d8729a619-kube-api-access-nhz8g\") pod \"redhat-marketplace-scglz\" (UID: \"0566c97e-c605-497d-a28f-080d8729a619\") " pod="openshift-marketplace/redhat-marketplace-scglz" Nov 25 10:56:13 crc kubenswrapper[4821]: I1125 10:56:13.725596 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0566c97e-c605-497d-a28f-080d8729a619-utilities\") pod \"redhat-marketplace-scglz\" (UID: \"0566c97e-c605-497d-a28f-080d8729a619\") " pod="openshift-marketplace/redhat-marketplace-scglz" Nov 25 10:56:13 crc kubenswrapper[4821]: I1125 10:56:13.827094 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nhz8g\" (UniqueName: \"kubernetes.io/projected/0566c97e-c605-497d-a28f-080d8729a619-kube-api-access-nhz8g\") pod \"redhat-marketplace-scglz\" (UID: \"0566c97e-c605-497d-a28f-080d8729a619\") " pod="openshift-marketplace/redhat-marketplace-scglz" Nov 25 10:56:13 crc kubenswrapper[4821]: I1125 10:56:13.827230 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0566c97e-c605-497d-a28f-080d8729a619-utilities\") pod \"redhat-marketplace-scglz\" (UID: \"0566c97e-c605-497d-a28f-080d8729a619\") " pod="openshift-marketplace/redhat-marketplace-scglz" Nov 25 10:56:13 crc kubenswrapper[4821]: I1125 10:56:13.827314 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0566c97e-c605-497d-a28f-080d8729a619-catalog-content\") pod \"redhat-marketplace-scglz\" (UID: \"0566c97e-c605-497d-a28f-080d8729a619\") " pod="openshift-marketplace/redhat-marketplace-scglz" Nov 25 10:56:13 crc kubenswrapper[4821]: I1125 10:56:13.827773 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0566c97e-c605-497d-a28f-080d8729a619-utilities\") pod \"redhat-marketplace-scglz\" (UID: \"0566c97e-c605-497d-a28f-080d8729a619\") " pod="openshift-marketplace/redhat-marketplace-scglz" Nov 25 10:56:13 crc kubenswrapper[4821]: I1125 10:56:13.827826 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0566c97e-c605-497d-a28f-080d8729a619-catalog-content\") pod \"redhat-marketplace-scglz\" (UID: \"0566c97e-c605-497d-a28f-080d8729a619\") " pod="openshift-marketplace/redhat-marketplace-scglz" Nov 25 10:56:13 crc kubenswrapper[4821]: I1125 10:56:13.845427 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nhz8g\" (UniqueName: \"kubernetes.io/projected/0566c97e-c605-497d-a28f-080d8729a619-kube-api-access-nhz8g\") pod \"redhat-marketplace-scglz\" (UID: \"0566c97e-c605-497d-a28f-080d8729a619\") " pod="openshift-marketplace/redhat-marketplace-scglz" Nov 25 10:56:13 crc kubenswrapper[4821]: I1125 10:56:13.936199 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-scglz" Nov 25 10:56:14 crc kubenswrapper[4821]: I1125 10:56:14.436262 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-scglz"] Nov 25 10:56:15 crc kubenswrapper[4821]: E1125 10:56:15.082483 4821 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0566c97e_c605_497d_a28f_080d8729a619.slice/crio-conmon-891891a71d293924f52ef6b536c95967b30508c6c3316b15b7d2af69dee5e51d.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0566c97e_c605_497d_a28f_080d8729a619.slice/crio-891891a71d293924f52ef6b536c95967b30508c6c3316b15b7d2af69dee5e51d.scope\": RecentStats: unable to find data in memory cache]" Nov 25 10:56:15 crc kubenswrapper[4821]: I1125 10:56:15.274567 4821 generic.go:334] "Generic (PLEG): container finished" podID="0566c97e-c605-497d-a28f-080d8729a619" containerID="891891a71d293924f52ef6b536c95967b30508c6c3316b15b7d2af69dee5e51d" exitCode=0 Nov 25 10:56:15 crc kubenswrapper[4821]: I1125 10:56:15.274643 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-scglz" event={"ID":"0566c97e-c605-497d-a28f-080d8729a619","Type":"ContainerDied","Data":"891891a71d293924f52ef6b536c95967b30508c6c3316b15b7d2af69dee5e51d"} Nov 25 10:56:15 crc kubenswrapper[4821]: I1125 10:56:15.274887 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-scglz" event={"ID":"0566c97e-c605-497d-a28f-080d8729a619","Type":"ContainerStarted","Data":"e09e19e1bde8498d804cf0ef3a80f7386e0f7863f653241ed643bdc6da6d8951"} Nov 25 10:56:16 crc kubenswrapper[4821]: I1125 10:56:16.285697 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-scglz" event={"ID":"0566c97e-c605-497d-a28f-080d8729a619","Type":"ContainerStarted","Data":"b9e454209c725087d9e9a040ae76ca2e23d92b406d961fe8bc95a787022990df"} Nov 25 10:56:17 crc kubenswrapper[4821]: I1125 10:56:17.297860 4821 generic.go:334] "Generic (PLEG): container finished" podID="0566c97e-c605-497d-a28f-080d8729a619" containerID="b9e454209c725087d9e9a040ae76ca2e23d92b406d961fe8bc95a787022990df" exitCode=0 Nov 25 10:56:17 crc kubenswrapper[4821]: I1125 10:56:17.297908 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-scglz" event={"ID":"0566c97e-c605-497d-a28f-080d8729a619","Type":"ContainerDied","Data":"b9e454209c725087d9e9a040ae76ca2e23d92b406d961fe8bc95a787022990df"} Nov 25 10:56:19 crc kubenswrapper[4821]: I1125 10:56:19.318994 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-scglz" event={"ID":"0566c97e-c605-497d-a28f-080d8729a619","Type":"ContainerStarted","Data":"2b887572a97721d33c88e139ad8cb2079a710a45308efbb5df8b86ff99417848"} Nov 25 10:56:19 crc kubenswrapper[4821]: I1125 10:56:19.340104 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-scglz" podStartSLOduration=3.393077076 podStartE2EDuration="6.340080554s" podCreationTimestamp="2025-11-25 10:56:13 +0000 UTC" firstStartedPulling="2025-11-25 10:56:15.27652675 +0000 UTC m=+1445.812846597" lastFinishedPulling="2025-11-25 10:56:18.223530228 +0000 UTC m=+1448.759850075" observedRunningTime="2025-11-25 10:56:19.334221986 +0000 UTC m=+1449.870541833" watchObservedRunningTime="2025-11-25 10:56:19.340080554 +0000 UTC m=+1449.876400401" Nov 25 10:56:22 crc kubenswrapper[4821]: I1125 10:56:22.278679 4821 scope.go:117] "RemoveContainer" containerID="bfcd58ab3bfcac6d66961f3ccea2866bf6fd3ae05317283d51b75ec0f336edee" Nov 25 10:56:22 crc kubenswrapper[4821]: I1125 10:56:22.315972 4821 scope.go:117] "RemoveContainer" containerID="dd3b43875d28c2b6e542e602fa80e40c8a008021f40ec73bda730d66d4f16f20" Nov 25 10:56:23 crc kubenswrapper[4821]: I1125 10:56:23.937279 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-scglz" Nov 25 10:56:23 crc kubenswrapper[4821]: I1125 10:56:23.937634 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-scglz" Nov 25 10:56:23 crc kubenswrapper[4821]: I1125 10:56:23.982672 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-scglz" Nov 25 10:56:24 crc kubenswrapper[4821]: I1125 10:56:24.414900 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-scglz" Nov 25 10:56:24 crc kubenswrapper[4821]: I1125 10:56:24.477455 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-scglz"] Nov 25 10:56:26 crc kubenswrapper[4821]: I1125 10:56:26.377380 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-scglz" podUID="0566c97e-c605-497d-a28f-080d8729a619" containerName="registry-server" containerID="cri-o://2b887572a97721d33c88e139ad8cb2079a710a45308efbb5df8b86ff99417848" gracePeriod=2 Nov 25 10:56:27 crc kubenswrapper[4821]: I1125 10:56:27.389698 4821 generic.go:334] "Generic (PLEG): container finished" podID="0566c97e-c605-497d-a28f-080d8729a619" containerID="2b887572a97721d33c88e139ad8cb2079a710a45308efbb5df8b86ff99417848" exitCode=0 Nov 25 10:56:27 crc kubenswrapper[4821]: I1125 10:56:27.389804 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-scglz" event={"ID":"0566c97e-c605-497d-a28f-080d8729a619","Type":"ContainerDied","Data":"2b887572a97721d33c88e139ad8cb2079a710a45308efbb5df8b86ff99417848"} Nov 25 10:56:28 crc kubenswrapper[4821]: I1125 10:56:28.015087 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-scglz" Nov 25 10:56:28 crc kubenswrapper[4821]: I1125 10:56:28.201513 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nhz8g\" (UniqueName: \"kubernetes.io/projected/0566c97e-c605-497d-a28f-080d8729a619-kube-api-access-nhz8g\") pod \"0566c97e-c605-497d-a28f-080d8729a619\" (UID: \"0566c97e-c605-497d-a28f-080d8729a619\") " Nov 25 10:56:28 crc kubenswrapper[4821]: I1125 10:56:28.201750 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0566c97e-c605-497d-a28f-080d8729a619-utilities\") pod \"0566c97e-c605-497d-a28f-080d8729a619\" (UID: \"0566c97e-c605-497d-a28f-080d8729a619\") " Nov 25 10:56:28 crc kubenswrapper[4821]: I1125 10:56:28.202150 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0566c97e-c605-497d-a28f-080d8729a619-catalog-content\") pod \"0566c97e-c605-497d-a28f-080d8729a619\" (UID: \"0566c97e-c605-497d-a28f-080d8729a619\") " Nov 25 10:56:28 crc kubenswrapper[4821]: I1125 10:56:28.202794 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0566c97e-c605-497d-a28f-080d8729a619-utilities" (OuterVolumeSpecName: "utilities") pod "0566c97e-c605-497d-a28f-080d8729a619" (UID: "0566c97e-c605-497d-a28f-080d8729a619"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:56:28 crc kubenswrapper[4821]: I1125 10:56:28.207851 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0566c97e-c605-497d-a28f-080d8729a619-kube-api-access-nhz8g" (OuterVolumeSpecName: "kube-api-access-nhz8g") pod "0566c97e-c605-497d-a28f-080d8729a619" (UID: "0566c97e-c605-497d-a28f-080d8729a619"). InnerVolumeSpecName "kube-api-access-nhz8g". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:56:28 crc kubenswrapper[4821]: I1125 10:56:28.222873 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0566c97e-c605-497d-a28f-080d8729a619-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0566c97e-c605-497d-a28f-080d8729a619" (UID: "0566c97e-c605-497d-a28f-080d8729a619"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:56:28 crc kubenswrapper[4821]: I1125 10:56:28.304384 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nhz8g\" (UniqueName: \"kubernetes.io/projected/0566c97e-c605-497d-a28f-080d8729a619-kube-api-access-nhz8g\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:28 crc kubenswrapper[4821]: I1125 10:56:28.304419 4821 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0566c97e-c605-497d-a28f-080d8729a619-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:28 crc kubenswrapper[4821]: I1125 10:56:28.304431 4821 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0566c97e-c605-497d-a28f-080d8729a619-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:28 crc kubenswrapper[4821]: I1125 10:56:28.401465 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-scglz" event={"ID":"0566c97e-c605-497d-a28f-080d8729a619","Type":"ContainerDied","Data":"e09e19e1bde8498d804cf0ef3a80f7386e0f7863f653241ed643bdc6da6d8951"} Nov 25 10:56:28 crc kubenswrapper[4821]: I1125 10:56:28.401514 4821 scope.go:117] "RemoveContainer" containerID="2b887572a97721d33c88e139ad8cb2079a710a45308efbb5df8b86ff99417848" Nov 25 10:56:28 crc kubenswrapper[4821]: I1125 10:56:28.401638 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-scglz" Nov 25 10:56:28 crc kubenswrapper[4821]: I1125 10:56:28.423264 4821 scope.go:117] "RemoveContainer" containerID="b9e454209c725087d9e9a040ae76ca2e23d92b406d961fe8bc95a787022990df" Nov 25 10:56:28 crc kubenswrapper[4821]: I1125 10:56:28.449472 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-scglz"] Nov 25 10:56:28 crc kubenswrapper[4821]: I1125 10:56:28.469260 4821 scope.go:117] "RemoveContainer" containerID="891891a71d293924f52ef6b536c95967b30508c6c3316b15b7d2af69dee5e51d" Nov 25 10:56:28 crc kubenswrapper[4821]: I1125 10:56:28.484315 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-scglz"] Nov 25 10:56:30 crc kubenswrapper[4821]: I1125 10:56:30.125624 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0566c97e-c605-497d-a28f-080d8729a619" path="/var/lib/kubelet/pods/0566c97e-c605-497d-a28f-080d8729a619/volumes" Nov 25 10:56:40 crc kubenswrapper[4821]: I1125 10:56:40.153503 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-6xj89"] Nov 25 10:56:40 crc kubenswrapper[4821]: E1125 10:56:40.154673 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0566c97e-c605-497d-a28f-080d8729a619" containerName="extract-utilities" Nov 25 10:56:40 crc kubenswrapper[4821]: I1125 10:56:40.154692 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="0566c97e-c605-497d-a28f-080d8729a619" containerName="extract-utilities" Nov 25 10:56:40 crc kubenswrapper[4821]: E1125 10:56:40.154715 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0566c97e-c605-497d-a28f-080d8729a619" containerName="extract-content" Nov 25 10:56:40 crc kubenswrapper[4821]: I1125 10:56:40.154723 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="0566c97e-c605-497d-a28f-080d8729a619" containerName="extract-content" Nov 25 10:56:40 crc kubenswrapper[4821]: E1125 10:56:40.154742 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0566c97e-c605-497d-a28f-080d8729a619" containerName="registry-server" Nov 25 10:56:40 crc kubenswrapper[4821]: I1125 10:56:40.154750 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="0566c97e-c605-497d-a28f-080d8729a619" containerName="registry-server" Nov 25 10:56:40 crc kubenswrapper[4821]: I1125 10:56:40.154954 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="0566c97e-c605-497d-a28f-080d8729a619" containerName="registry-server" Nov 25 10:56:40 crc kubenswrapper[4821]: I1125 10:56:40.156601 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6xj89" Nov 25 10:56:40 crc kubenswrapper[4821]: I1125 10:56:40.179889 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6xj89"] Nov 25 10:56:40 crc kubenswrapper[4821]: I1125 10:56:40.184640 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6b93d4c-3924-4d54-b6bd-a78ef59fad97-catalog-content\") pod \"community-operators-6xj89\" (UID: \"b6b93d4c-3924-4d54-b6bd-a78ef59fad97\") " pod="openshift-marketplace/community-operators-6xj89" Nov 25 10:56:40 crc kubenswrapper[4821]: I1125 10:56:40.184701 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n26bc\" (UniqueName: \"kubernetes.io/projected/b6b93d4c-3924-4d54-b6bd-a78ef59fad97-kube-api-access-n26bc\") pod \"community-operators-6xj89\" (UID: \"b6b93d4c-3924-4d54-b6bd-a78ef59fad97\") " pod="openshift-marketplace/community-operators-6xj89" Nov 25 10:56:40 crc kubenswrapper[4821]: I1125 10:56:40.184774 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6b93d4c-3924-4d54-b6bd-a78ef59fad97-utilities\") pod \"community-operators-6xj89\" (UID: \"b6b93d4c-3924-4d54-b6bd-a78ef59fad97\") " pod="openshift-marketplace/community-operators-6xj89" Nov 25 10:56:40 crc kubenswrapper[4821]: I1125 10:56:40.286922 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6b93d4c-3924-4d54-b6bd-a78ef59fad97-utilities\") pod \"community-operators-6xj89\" (UID: \"b6b93d4c-3924-4d54-b6bd-a78ef59fad97\") " pod="openshift-marketplace/community-operators-6xj89" Nov 25 10:56:40 crc kubenswrapper[4821]: I1125 10:56:40.287079 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6b93d4c-3924-4d54-b6bd-a78ef59fad97-catalog-content\") pod \"community-operators-6xj89\" (UID: \"b6b93d4c-3924-4d54-b6bd-a78ef59fad97\") " pod="openshift-marketplace/community-operators-6xj89" Nov 25 10:56:40 crc kubenswrapper[4821]: I1125 10:56:40.287284 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n26bc\" (UniqueName: \"kubernetes.io/projected/b6b93d4c-3924-4d54-b6bd-a78ef59fad97-kube-api-access-n26bc\") pod \"community-operators-6xj89\" (UID: \"b6b93d4c-3924-4d54-b6bd-a78ef59fad97\") " pod="openshift-marketplace/community-operators-6xj89" Nov 25 10:56:40 crc kubenswrapper[4821]: I1125 10:56:40.287501 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6b93d4c-3924-4d54-b6bd-a78ef59fad97-utilities\") pod \"community-operators-6xj89\" (UID: \"b6b93d4c-3924-4d54-b6bd-a78ef59fad97\") " pod="openshift-marketplace/community-operators-6xj89" Nov 25 10:56:40 crc kubenswrapper[4821]: I1125 10:56:40.287590 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6b93d4c-3924-4d54-b6bd-a78ef59fad97-catalog-content\") pod \"community-operators-6xj89\" (UID: \"b6b93d4c-3924-4d54-b6bd-a78ef59fad97\") " pod="openshift-marketplace/community-operators-6xj89" Nov 25 10:56:40 crc kubenswrapper[4821]: I1125 10:56:40.309712 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n26bc\" (UniqueName: \"kubernetes.io/projected/b6b93d4c-3924-4d54-b6bd-a78ef59fad97-kube-api-access-n26bc\") pod \"community-operators-6xj89\" (UID: \"b6b93d4c-3924-4d54-b6bd-a78ef59fad97\") " pod="openshift-marketplace/community-operators-6xj89" Nov 25 10:56:40 crc kubenswrapper[4821]: I1125 10:56:40.490906 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6xj89" Nov 25 10:56:41 crc kubenswrapper[4821]: I1125 10:56:41.045506 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6xj89"] Nov 25 10:56:41 crc kubenswrapper[4821]: I1125 10:56:41.586456 4821 generic.go:334] "Generic (PLEG): container finished" podID="b6b93d4c-3924-4d54-b6bd-a78ef59fad97" containerID="c9eb1689192a7cfdcff6fa9e6dcfa26c3245efd5e918673864c45d4f7ac203f9" exitCode=0 Nov 25 10:56:41 crc kubenswrapper[4821]: I1125 10:56:41.586558 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6xj89" event={"ID":"b6b93d4c-3924-4d54-b6bd-a78ef59fad97","Type":"ContainerDied","Data":"c9eb1689192a7cfdcff6fa9e6dcfa26c3245efd5e918673864c45d4f7ac203f9"} Nov 25 10:56:41 crc kubenswrapper[4821]: I1125 10:56:41.586884 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6xj89" event={"ID":"b6b93d4c-3924-4d54-b6bd-a78ef59fad97","Type":"ContainerStarted","Data":"272c672294406fdac0da7b35fd1d7699289dfdbe71d3f73b7b4c632072f01bd3"} Nov 25 10:56:43 crc kubenswrapper[4821]: I1125 10:56:43.605016 4821 generic.go:334] "Generic (PLEG): container finished" podID="b6b93d4c-3924-4d54-b6bd-a78ef59fad97" containerID="a68c8f885d2fbe4ef950154567d94650454f9305bcfe2f18aebffe5d7bda05a4" exitCode=0 Nov 25 10:56:43 crc kubenswrapper[4821]: I1125 10:56:43.605059 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6xj89" event={"ID":"b6b93d4c-3924-4d54-b6bd-a78ef59fad97","Type":"ContainerDied","Data":"a68c8f885d2fbe4ef950154567d94650454f9305bcfe2f18aebffe5d7bda05a4"} Nov 25 10:56:44 crc kubenswrapper[4821]: I1125 10:56:44.615942 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6xj89" event={"ID":"b6b93d4c-3924-4d54-b6bd-a78ef59fad97","Type":"ContainerStarted","Data":"c7bf8f86030c3c381dae2aeffaaff8df4c7db1ed9f73aa2946e58743cfd3e6fe"} Nov 25 10:56:44 crc kubenswrapper[4821]: I1125 10:56:44.646839 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-6xj89" podStartSLOduration=2.176577888 podStartE2EDuration="4.646819621s" podCreationTimestamp="2025-11-25 10:56:40 +0000 UTC" firstStartedPulling="2025-11-25 10:56:41.589597811 +0000 UTC m=+1472.125917658" lastFinishedPulling="2025-11-25 10:56:44.059839544 +0000 UTC m=+1474.596159391" observedRunningTime="2025-11-25 10:56:44.637271681 +0000 UTC m=+1475.173591548" watchObservedRunningTime="2025-11-25 10:56:44.646819621 +0000 UTC m=+1475.183139468" Nov 25 10:56:50 crc kubenswrapper[4821]: I1125 10:56:50.492291 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-6xj89" Nov 25 10:56:50 crc kubenswrapper[4821]: I1125 10:56:50.492883 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-6xj89" Nov 25 10:56:50 crc kubenswrapper[4821]: I1125 10:56:50.539336 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-6xj89" Nov 25 10:56:50 crc kubenswrapper[4821]: I1125 10:56:50.718664 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-6xj89" Nov 25 10:56:50 crc kubenswrapper[4821]: I1125 10:56:50.772048 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6xj89"] Nov 25 10:56:52 crc kubenswrapper[4821]: I1125 10:56:52.682759 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-6xj89" podUID="b6b93d4c-3924-4d54-b6bd-a78ef59fad97" containerName="registry-server" containerID="cri-o://c7bf8f86030c3c381dae2aeffaaff8df4c7db1ed9f73aa2946e58743cfd3e6fe" gracePeriod=2 Nov 25 10:56:53 crc kubenswrapper[4821]: I1125 10:56:53.152393 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6xj89" Nov 25 10:56:53 crc kubenswrapper[4821]: I1125 10:56:53.331011 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n26bc\" (UniqueName: \"kubernetes.io/projected/b6b93d4c-3924-4d54-b6bd-a78ef59fad97-kube-api-access-n26bc\") pod \"b6b93d4c-3924-4d54-b6bd-a78ef59fad97\" (UID: \"b6b93d4c-3924-4d54-b6bd-a78ef59fad97\") " Nov 25 10:56:53 crc kubenswrapper[4821]: I1125 10:56:53.331296 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6b93d4c-3924-4d54-b6bd-a78ef59fad97-catalog-content\") pod \"b6b93d4c-3924-4d54-b6bd-a78ef59fad97\" (UID: \"b6b93d4c-3924-4d54-b6bd-a78ef59fad97\") " Nov 25 10:56:53 crc kubenswrapper[4821]: I1125 10:56:53.331384 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6b93d4c-3924-4d54-b6bd-a78ef59fad97-utilities\") pod \"b6b93d4c-3924-4d54-b6bd-a78ef59fad97\" (UID: \"b6b93d4c-3924-4d54-b6bd-a78ef59fad97\") " Nov 25 10:56:53 crc kubenswrapper[4821]: I1125 10:56:53.332456 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b6b93d4c-3924-4d54-b6bd-a78ef59fad97-utilities" (OuterVolumeSpecName: "utilities") pod "b6b93d4c-3924-4d54-b6bd-a78ef59fad97" (UID: "b6b93d4c-3924-4d54-b6bd-a78ef59fad97"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:56:53 crc kubenswrapper[4821]: I1125 10:56:53.339265 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6b93d4c-3924-4d54-b6bd-a78ef59fad97-kube-api-access-n26bc" (OuterVolumeSpecName: "kube-api-access-n26bc") pod "b6b93d4c-3924-4d54-b6bd-a78ef59fad97" (UID: "b6b93d4c-3924-4d54-b6bd-a78ef59fad97"). InnerVolumeSpecName "kube-api-access-n26bc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:56:53 crc kubenswrapper[4821]: I1125 10:56:53.386875 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b6b93d4c-3924-4d54-b6bd-a78ef59fad97-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b6b93d4c-3924-4d54-b6bd-a78ef59fad97" (UID: "b6b93d4c-3924-4d54-b6bd-a78ef59fad97"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:56:53 crc kubenswrapper[4821]: I1125 10:56:53.434217 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n26bc\" (UniqueName: \"kubernetes.io/projected/b6b93d4c-3924-4d54-b6bd-a78ef59fad97-kube-api-access-n26bc\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:53 crc kubenswrapper[4821]: I1125 10:56:53.434267 4821 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6b93d4c-3924-4d54-b6bd-a78ef59fad97-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:53 crc kubenswrapper[4821]: I1125 10:56:53.434280 4821 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6b93d4c-3924-4d54-b6bd-a78ef59fad97-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:56:53 crc kubenswrapper[4821]: I1125 10:56:53.693471 4821 generic.go:334] "Generic (PLEG): container finished" podID="b6b93d4c-3924-4d54-b6bd-a78ef59fad97" containerID="c7bf8f86030c3c381dae2aeffaaff8df4c7db1ed9f73aa2946e58743cfd3e6fe" exitCode=0 Nov 25 10:56:53 crc kubenswrapper[4821]: I1125 10:56:53.693524 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6xj89" Nov 25 10:56:53 crc kubenswrapper[4821]: I1125 10:56:53.693541 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6xj89" event={"ID":"b6b93d4c-3924-4d54-b6bd-a78ef59fad97","Type":"ContainerDied","Data":"c7bf8f86030c3c381dae2aeffaaff8df4c7db1ed9f73aa2946e58743cfd3e6fe"} Nov 25 10:56:53 crc kubenswrapper[4821]: I1125 10:56:53.693725 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6xj89" event={"ID":"b6b93d4c-3924-4d54-b6bd-a78ef59fad97","Type":"ContainerDied","Data":"272c672294406fdac0da7b35fd1d7699289dfdbe71d3f73b7b4c632072f01bd3"} Nov 25 10:56:53 crc kubenswrapper[4821]: I1125 10:56:53.693826 4821 scope.go:117] "RemoveContainer" containerID="c7bf8f86030c3c381dae2aeffaaff8df4c7db1ed9f73aa2946e58743cfd3e6fe" Nov 25 10:56:53 crc kubenswrapper[4821]: I1125 10:56:53.716138 4821 scope.go:117] "RemoveContainer" containerID="a68c8f885d2fbe4ef950154567d94650454f9305bcfe2f18aebffe5d7bda05a4" Nov 25 10:56:53 crc kubenswrapper[4821]: I1125 10:56:53.738144 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6xj89"] Nov 25 10:56:53 crc kubenswrapper[4821]: I1125 10:56:53.744976 4821 scope.go:117] "RemoveContainer" containerID="c9eb1689192a7cfdcff6fa9e6dcfa26c3245efd5e918673864c45d4f7ac203f9" Nov 25 10:56:53 crc kubenswrapper[4821]: I1125 10:56:53.745488 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-6xj89"] Nov 25 10:56:53 crc kubenswrapper[4821]: I1125 10:56:53.811151 4821 scope.go:117] "RemoveContainer" containerID="c7bf8f86030c3c381dae2aeffaaff8df4c7db1ed9f73aa2946e58743cfd3e6fe" Nov 25 10:56:53 crc kubenswrapper[4821]: E1125 10:56:53.811710 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c7bf8f86030c3c381dae2aeffaaff8df4c7db1ed9f73aa2946e58743cfd3e6fe\": container with ID starting with c7bf8f86030c3c381dae2aeffaaff8df4c7db1ed9f73aa2946e58743cfd3e6fe not found: ID does not exist" containerID="c7bf8f86030c3c381dae2aeffaaff8df4c7db1ed9f73aa2946e58743cfd3e6fe" Nov 25 10:56:53 crc kubenswrapper[4821]: I1125 10:56:53.811765 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7bf8f86030c3c381dae2aeffaaff8df4c7db1ed9f73aa2946e58743cfd3e6fe"} err="failed to get container status \"c7bf8f86030c3c381dae2aeffaaff8df4c7db1ed9f73aa2946e58743cfd3e6fe\": rpc error: code = NotFound desc = could not find container \"c7bf8f86030c3c381dae2aeffaaff8df4c7db1ed9f73aa2946e58743cfd3e6fe\": container with ID starting with c7bf8f86030c3c381dae2aeffaaff8df4c7db1ed9f73aa2946e58743cfd3e6fe not found: ID does not exist" Nov 25 10:56:53 crc kubenswrapper[4821]: I1125 10:56:53.811788 4821 scope.go:117] "RemoveContainer" containerID="a68c8f885d2fbe4ef950154567d94650454f9305bcfe2f18aebffe5d7bda05a4" Nov 25 10:56:53 crc kubenswrapper[4821]: E1125 10:56:53.812200 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a68c8f885d2fbe4ef950154567d94650454f9305bcfe2f18aebffe5d7bda05a4\": container with ID starting with a68c8f885d2fbe4ef950154567d94650454f9305bcfe2f18aebffe5d7bda05a4 not found: ID does not exist" containerID="a68c8f885d2fbe4ef950154567d94650454f9305bcfe2f18aebffe5d7bda05a4" Nov 25 10:56:53 crc kubenswrapper[4821]: I1125 10:56:53.812252 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a68c8f885d2fbe4ef950154567d94650454f9305bcfe2f18aebffe5d7bda05a4"} err="failed to get container status \"a68c8f885d2fbe4ef950154567d94650454f9305bcfe2f18aebffe5d7bda05a4\": rpc error: code = NotFound desc = could not find container \"a68c8f885d2fbe4ef950154567d94650454f9305bcfe2f18aebffe5d7bda05a4\": container with ID starting with a68c8f885d2fbe4ef950154567d94650454f9305bcfe2f18aebffe5d7bda05a4 not found: ID does not exist" Nov 25 10:56:53 crc kubenswrapper[4821]: I1125 10:56:53.812281 4821 scope.go:117] "RemoveContainer" containerID="c9eb1689192a7cfdcff6fa9e6dcfa26c3245efd5e918673864c45d4f7ac203f9" Nov 25 10:56:53 crc kubenswrapper[4821]: E1125 10:56:53.812682 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c9eb1689192a7cfdcff6fa9e6dcfa26c3245efd5e918673864c45d4f7ac203f9\": container with ID starting with c9eb1689192a7cfdcff6fa9e6dcfa26c3245efd5e918673864c45d4f7ac203f9 not found: ID does not exist" containerID="c9eb1689192a7cfdcff6fa9e6dcfa26c3245efd5e918673864c45d4f7ac203f9" Nov 25 10:56:53 crc kubenswrapper[4821]: I1125 10:56:53.812736 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9eb1689192a7cfdcff6fa9e6dcfa26c3245efd5e918673864c45d4f7ac203f9"} err="failed to get container status \"c9eb1689192a7cfdcff6fa9e6dcfa26c3245efd5e918673864c45d4f7ac203f9\": rpc error: code = NotFound desc = could not find container \"c9eb1689192a7cfdcff6fa9e6dcfa26c3245efd5e918673864c45d4f7ac203f9\": container with ID starting with c9eb1689192a7cfdcff6fa9e6dcfa26c3245efd5e918673864c45d4f7ac203f9 not found: ID does not exist" Nov 25 10:56:54 crc kubenswrapper[4821]: I1125 10:56:54.127520 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6b93d4c-3924-4d54-b6bd-a78ef59fad97" path="/var/lib/kubelet/pods/b6b93d4c-3924-4d54-b6bd-a78ef59fad97/volumes" Nov 25 10:57:36 crc kubenswrapper[4821]: I1125 10:57:36.223457 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-qsbgm"] Nov 25 10:57:36 crc kubenswrapper[4821]: E1125 10:57:36.224445 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6b93d4c-3924-4d54-b6bd-a78ef59fad97" containerName="registry-server" Nov 25 10:57:36 crc kubenswrapper[4821]: I1125 10:57:36.224463 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6b93d4c-3924-4d54-b6bd-a78ef59fad97" containerName="registry-server" Nov 25 10:57:36 crc kubenswrapper[4821]: E1125 10:57:36.224486 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6b93d4c-3924-4d54-b6bd-a78ef59fad97" containerName="extract-content" Nov 25 10:57:36 crc kubenswrapper[4821]: I1125 10:57:36.224493 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6b93d4c-3924-4d54-b6bd-a78ef59fad97" containerName="extract-content" Nov 25 10:57:36 crc kubenswrapper[4821]: E1125 10:57:36.224515 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6b93d4c-3924-4d54-b6bd-a78ef59fad97" containerName="extract-utilities" Nov 25 10:57:36 crc kubenswrapper[4821]: I1125 10:57:36.224522 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6b93d4c-3924-4d54-b6bd-a78ef59fad97" containerName="extract-utilities" Nov 25 10:57:36 crc kubenswrapper[4821]: I1125 10:57:36.224700 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6b93d4c-3924-4d54-b6bd-a78ef59fad97" containerName="registry-server" Nov 25 10:57:36 crc kubenswrapper[4821]: I1125 10:57:36.226842 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qsbgm" Nov 25 10:57:36 crc kubenswrapper[4821]: I1125 10:57:36.254636 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qsbgm"] Nov 25 10:57:36 crc kubenswrapper[4821]: I1125 10:57:36.377821 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03dec67a-7050-4d6a-81ae-7a41c494572f-catalog-content\") pod \"certified-operators-qsbgm\" (UID: \"03dec67a-7050-4d6a-81ae-7a41c494572f\") " pod="openshift-marketplace/certified-operators-qsbgm" Nov 25 10:57:36 crc kubenswrapper[4821]: I1125 10:57:36.377882 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-74lhh\" (UniqueName: \"kubernetes.io/projected/03dec67a-7050-4d6a-81ae-7a41c494572f-kube-api-access-74lhh\") pod \"certified-operators-qsbgm\" (UID: \"03dec67a-7050-4d6a-81ae-7a41c494572f\") " pod="openshift-marketplace/certified-operators-qsbgm" Nov 25 10:57:36 crc kubenswrapper[4821]: I1125 10:57:36.378028 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03dec67a-7050-4d6a-81ae-7a41c494572f-utilities\") pod \"certified-operators-qsbgm\" (UID: \"03dec67a-7050-4d6a-81ae-7a41c494572f\") " pod="openshift-marketplace/certified-operators-qsbgm" Nov 25 10:57:36 crc kubenswrapper[4821]: I1125 10:57:36.479548 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03dec67a-7050-4d6a-81ae-7a41c494572f-catalog-content\") pod \"certified-operators-qsbgm\" (UID: \"03dec67a-7050-4d6a-81ae-7a41c494572f\") " pod="openshift-marketplace/certified-operators-qsbgm" Nov 25 10:57:36 crc kubenswrapper[4821]: I1125 10:57:36.479605 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-74lhh\" (UniqueName: \"kubernetes.io/projected/03dec67a-7050-4d6a-81ae-7a41c494572f-kube-api-access-74lhh\") pod \"certified-operators-qsbgm\" (UID: \"03dec67a-7050-4d6a-81ae-7a41c494572f\") " pod="openshift-marketplace/certified-operators-qsbgm" Nov 25 10:57:36 crc kubenswrapper[4821]: I1125 10:57:36.479705 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03dec67a-7050-4d6a-81ae-7a41c494572f-utilities\") pod \"certified-operators-qsbgm\" (UID: \"03dec67a-7050-4d6a-81ae-7a41c494572f\") " pod="openshift-marketplace/certified-operators-qsbgm" Nov 25 10:57:36 crc kubenswrapper[4821]: I1125 10:57:36.480261 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03dec67a-7050-4d6a-81ae-7a41c494572f-catalog-content\") pod \"certified-operators-qsbgm\" (UID: \"03dec67a-7050-4d6a-81ae-7a41c494572f\") " pod="openshift-marketplace/certified-operators-qsbgm" Nov 25 10:57:36 crc kubenswrapper[4821]: I1125 10:57:36.480261 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03dec67a-7050-4d6a-81ae-7a41c494572f-utilities\") pod \"certified-operators-qsbgm\" (UID: \"03dec67a-7050-4d6a-81ae-7a41c494572f\") " pod="openshift-marketplace/certified-operators-qsbgm" Nov 25 10:57:36 crc kubenswrapper[4821]: I1125 10:57:36.503863 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-74lhh\" (UniqueName: \"kubernetes.io/projected/03dec67a-7050-4d6a-81ae-7a41c494572f-kube-api-access-74lhh\") pod \"certified-operators-qsbgm\" (UID: \"03dec67a-7050-4d6a-81ae-7a41c494572f\") " pod="openshift-marketplace/certified-operators-qsbgm" Nov 25 10:57:36 crc kubenswrapper[4821]: I1125 10:57:36.553238 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qsbgm" Nov 25 10:57:37 crc kubenswrapper[4821]: I1125 10:57:37.077197 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qsbgm"] Nov 25 10:57:38 crc kubenswrapper[4821]: I1125 10:57:38.094915 4821 generic.go:334] "Generic (PLEG): container finished" podID="03dec67a-7050-4d6a-81ae-7a41c494572f" containerID="2332540b100a5965287eba0c79003020523bafbed7e9a80d21aea4dc7061d5d3" exitCode=0 Nov 25 10:57:38 crc kubenswrapper[4821]: I1125 10:57:38.094966 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qsbgm" event={"ID":"03dec67a-7050-4d6a-81ae-7a41c494572f","Type":"ContainerDied","Data":"2332540b100a5965287eba0c79003020523bafbed7e9a80d21aea4dc7061d5d3"} Nov 25 10:57:38 crc kubenswrapper[4821]: I1125 10:57:38.094998 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qsbgm" event={"ID":"03dec67a-7050-4d6a-81ae-7a41c494572f","Type":"ContainerStarted","Data":"a28d451a8b8beb5ca6bf39f49866e2534bcc9aea188a073c4b843cc004f5b6ad"} Nov 25 10:57:38 crc kubenswrapper[4821]: I1125 10:57:38.097444 4821 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 10:57:39 crc kubenswrapper[4821]: I1125 10:57:39.106583 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qsbgm" event={"ID":"03dec67a-7050-4d6a-81ae-7a41c494572f","Type":"ContainerStarted","Data":"2839e4443f172676494ffd5971b37772fc4562cfcff4cabb9355a718f98d4010"} Nov 25 10:57:40 crc kubenswrapper[4821]: I1125 10:57:40.119888 4821 generic.go:334] "Generic (PLEG): container finished" podID="03dec67a-7050-4d6a-81ae-7a41c494572f" containerID="2839e4443f172676494ffd5971b37772fc4562cfcff4cabb9355a718f98d4010" exitCode=0 Nov 25 10:57:40 crc kubenswrapper[4821]: I1125 10:57:40.127464 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qsbgm" event={"ID":"03dec67a-7050-4d6a-81ae-7a41c494572f","Type":"ContainerDied","Data":"2839e4443f172676494ffd5971b37772fc4562cfcff4cabb9355a718f98d4010"} Nov 25 10:57:41 crc kubenswrapper[4821]: I1125 10:57:41.135748 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qsbgm" event={"ID":"03dec67a-7050-4d6a-81ae-7a41c494572f","Type":"ContainerStarted","Data":"74c82bb37e0b0621ce242ba349e34e619fe418668b97f227a35374ccd1008a92"} Nov 25 10:57:41 crc kubenswrapper[4821]: I1125 10:57:41.173441 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-qsbgm" podStartSLOduration=2.689914267 podStartE2EDuration="5.173416582s" podCreationTimestamp="2025-11-25 10:57:36 +0000 UTC" firstStartedPulling="2025-11-25 10:57:38.097209137 +0000 UTC m=+1528.633528984" lastFinishedPulling="2025-11-25 10:57:40.580711452 +0000 UTC m=+1531.117031299" observedRunningTime="2025-11-25 10:57:41.165308068 +0000 UTC m=+1531.701627925" watchObservedRunningTime="2025-11-25 10:57:41.173416582 +0000 UTC m=+1531.709736429" Nov 25 10:57:46 crc kubenswrapper[4821]: I1125 10:57:46.554245 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-qsbgm" Nov 25 10:57:46 crc kubenswrapper[4821]: I1125 10:57:46.554778 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-qsbgm" Nov 25 10:57:46 crc kubenswrapper[4821]: I1125 10:57:46.601814 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-qsbgm" Nov 25 10:57:47 crc kubenswrapper[4821]: I1125 10:57:47.259654 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-qsbgm" Nov 25 10:57:47 crc kubenswrapper[4821]: I1125 10:57:47.315658 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qsbgm"] Nov 25 10:57:48 crc kubenswrapper[4821]: I1125 10:57:48.039212 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-kt7h7"] Nov 25 10:57:48 crc kubenswrapper[4821]: I1125 10:57:48.050083 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-7774-account-create-xq572"] Nov 25 10:57:48 crc kubenswrapper[4821]: I1125 10:57:48.058482 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-7774-account-create-xq572"] Nov 25 10:57:48 crc kubenswrapper[4821]: I1125 10:57:48.065464 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-kt7h7"] Nov 25 10:57:48 crc kubenswrapper[4821]: I1125 10:57:48.125817 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="802c0ca8-b806-4218-86a0-e88b3325a762" path="/var/lib/kubelet/pods/802c0ca8-b806-4218-86a0-e88b3325a762/volumes" Nov 25 10:57:48 crc kubenswrapper[4821]: I1125 10:57:48.126457 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f10473c7-e8cd-4534-8232-7cb6a97be557" path="/var/lib/kubelet/pods/f10473c7-e8cd-4534-8232-7cb6a97be557/volumes" Nov 25 10:57:49 crc kubenswrapper[4821]: I1125 10:57:49.029936 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-kxgp9"] Nov 25 10:57:49 crc kubenswrapper[4821]: I1125 10:57:49.039785 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-kxgp9"] Nov 25 10:57:49 crc kubenswrapper[4821]: I1125 10:57:49.202876 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-qsbgm" podUID="03dec67a-7050-4d6a-81ae-7a41c494572f" containerName="registry-server" containerID="cri-o://74c82bb37e0b0621ce242ba349e34e619fe418668b97f227a35374ccd1008a92" gracePeriod=2 Nov 25 10:57:49 crc kubenswrapper[4821]: I1125 10:57:49.655302 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qsbgm" Nov 25 10:57:49 crc kubenswrapper[4821]: I1125 10:57:49.853449 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03dec67a-7050-4d6a-81ae-7a41c494572f-utilities\") pod \"03dec67a-7050-4d6a-81ae-7a41c494572f\" (UID: \"03dec67a-7050-4d6a-81ae-7a41c494572f\") " Nov 25 10:57:49 crc kubenswrapper[4821]: I1125 10:57:49.853776 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03dec67a-7050-4d6a-81ae-7a41c494572f-catalog-content\") pod \"03dec67a-7050-4d6a-81ae-7a41c494572f\" (UID: \"03dec67a-7050-4d6a-81ae-7a41c494572f\") " Nov 25 10:57:49 crc kubenswrapper[4821]: I1125 10:57:49.853900 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-74lhh\" (UniqueName: \"kubernetes.io/projected/03dec67a-7050-4d6a-81ae-7a41c494572f-kube-api-access-74lhh\") pod \"03dec67a-7050-4d6a-81ae-7a41c494572f\" (UID: \"03dec67a-7050-4d6a-81ae-7a41c494572f\") " Nov 25 10:57:49 crc kubenswrapper[4821]: I1125 10:57:49.854674 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03dec67a-7050-4d6a-81ae-7a41c494572f-utilities" (OuterVolumeSpecName: "utilities") pod "03dec67a-7050-4d6a-81ae-7a41c494572f" (UID: "03dec67a-7050-4d6a-81ae-7a41c494572f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:57:49 crc kubenswrapper[4821]: I1125 10:57:49.860026 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03dec67a-7050-4d6a-81ae-7a41c494572f-kube-api-access-74lhh" (OuterVolumeSpecName: "kube-api-access-74lhh") pod "03dec67a-7050-4d6a-81ae-7a41c494572f" (UID: "03dec67a-7050-4d6a-81ae-7a41c494572f"). InnerVolumeSpecName "kube-api-access-74lhh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:57:49 crc kubenswrapper[4821]: I1125 10:57:49.956210 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-74lhh\" (UniqueName: \"kubernetes.io/projected/03dec67a-7050-4d6a-81ae-7a41c494572f-kube-api-access-74lhh\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:49 crc kubenswrapper[4821]: I1125 10:57:49.956255 4821 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03dec67a-7050-4d6a-81ae-7a41c494572f-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:50 crc kubenswrapper[4821]: I1125 10:57:50.127973 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9ee739b-0ed7-43aa-b6f1-a779cb718758" path="/var/lib/kubelet/pods/a9ee739b-0ed7-43aa-b6f1-a779cb718758/volumes" Nov 25 10:57:50 crc kubenswrapper[4821]: I1125 10:57:50.214589 4821 generic.go:334] "Generic (PLEG): container finished" podID="03dec67a-7050-4d6a-81ae-7a41c494572f" containerID="74c82bb37e0b0621ce242ba349e34e619fe418668b97f227a35374ccd1008a92" exitCode=0 Nov 25 10:57:50 crc kubenswrapper[4821]: I1125 10:57:50.214644 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qsbgm" event={"ID":"03dec67a-7050-4d6a-81ae-7a41c494572f","Type":"ContainerDied","Data":"74c82bb37e0b0621ce242ba349e34e619fe418668b97f227a35374ccd1008a92"} Nov 25 10:57:50 crc kubenswrapper[4821]: I1125 10:57:50.214673 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qsbgm" event={"ID":"03dec67a-7050-4d6a-81ae-7a41c494572f","Type":"ContainerDied","Data":"a28d451a8b8beb5ca6bf39f49866e2534bcc9aea188a073c4b843cc004f5b6ad"} Nov 25 10:57:50 crc kubenswrapper[4821]: I1125 10:57:50.214694 4821 scope.go:117] "RemoveContainer" containerID="74c82bb37e0b0621ce242ba349e34e619fe418668b97f227a35374ccd1008a92" Nov 25 10:57:50 crc kubenswrapper[4821]: I1125 10:57:50.214876 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qsbgm" Nov 25 10:57:50 crc kubenswrapper[4821]: I1125 10:57:50.243295 4821 scope.go:117] "RemoveContainer" containerID="2839e4443f172676494ffd5971b37772fc4562cfcff4cabb9355a718f98d4010" Nov 25 10:57:50 crc kubenswrapper[4821]: I1125 10:57:50.276867 4821 scope.go:117] "RemoveContainer" containerID="2332540b100a5965287eba0c79003020523bafbed7e9a80d21aea4dc7061d5d3" Nov 25 10:57:50 crc kubenswrapper[4821]: I1125 10:57:50.306606 4821 scope.go:117] "RemoveContainer" containerID="74c82bb37e0b0621ce242ba349e34e619fe418668b97f227a35374ccd1008a92" Nov 25 10:57:50 crc kubenswrapper[4821]: E1125 10:57:50.307121 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"74c82bb37e0b0621ce242ba349e34e619fe418668b97f227a35374ccd1008a92\": container with ID starting with 74c82bb37e0b0621ce242ba349e34e619fe418668b97f227a35374ccd1008a92 not found: ID does not exist" containerID="74c82bb37e0b0621ce242ba349e34e619fe418668b97f227a35374ccd1008a92" Nov 25 10:57:50 crc kubenswrapper[4821]: I1125 10:57:50.307173 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74c82bb37e0b0621ce242ba349e34e619fe418668b97f227a35374ccd1008a92"} err="failed to get container status \"74c82bb37e0b0621ce242ba349e34e619fe418668b97f227a35374ccd1008a92\": rpc error: code = NotFound desc = could not find container \"74c82bb37e0b0621ce242ba349e34e619fe418668b97f227a35374ccd1008a92\": container with ID starting with 74c82bb37e0b0621ce242ba349e34e619fe418668b97f227a35374ccd1008a92 not found: ID does not exist" Nov 25 10:57:50 crc kubenswrapper[4821]: I1125 10:57:50.307202 4821 scope.go:117] "RemoveContainer" containerID="2839e4443f172676494ffd5971b37772fc4562cfcff4cabb9355a718f98d4010" Nov 25 10:57:50 crc kubenswrapper[4821]: E1125 10:57:50.307527 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2839e4443f172676494ffd5971b37772fc4562cfcff4cabb9355a718f98d4010\": container with ID starting with 2839e4443f172676494ffd5971b37772fc4562cfcff4cabb9355a718f98d4010 not found: ID does not exist" containerID="2839e4443f172676494ffd5971b37772fc4562cfcff4cabb9355a718f98d4010" Nov 25 10:57:50 crc kubenswrapper[4821]: I1125 10:57:50.307584 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2839e4443f172676494ffd5971b37772fc4562cfcff4cabb9355a718f98d4010"} err="failed to get container status \"2839e4443f172676494ffd5971b37772fc4562cfcff4cabb9355a718f98d4010\": rpc error: code = NotFound desc = could not find container \"2839e4443f172676494ffd5971b37772fc4562cfcff4cabb9355a718f98d4010\": container with ID starting with 2839e4443f172676494ffd5971b37772fc4562cfcff4cabb9355a718f98d4010 not found: ID does not exist" Nov 25 10:57:50 crc kubenswrapper[4821]: I1125 10:57:50.307614 4821 scope.go:117] "RemoveContainer" containerID="2332540b100a5965287eba0c79003020523bafbed7e9a80d21aea4dc7061d5d3" Nov 25 10:57:50 crc kubenswrapper[4821]: E1125 10:57:50.307957 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2332540b100a5965287eba0c79003020523bafbed7e9a80d21aea4dc7061d5d3\": container with ID starting with 2332540b100a5965287eba0c79003020523bafbed7e9a80d21aea4dc7061d5d3 not found: ID does not exist" containerID="2332540b100a5965287eba0c79003020523bafbed7e9a80d21aea4dc7061d5d3" Nov 25 10:57:50 crc kubenswrapper[4821]: I1125 10:57:50.307993 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2332540b100a5965287eba0c79003020523bafbed7e9a80d21aea4dc7061d5d3"} err="failed to get container status \"2332540b100a5965287eba0c79003020523bafbed7e9a80d21aea4dc7061d5d3\": rpc error: code = NotFound desc = could not find container \"2332540b100a5965287eba0c79003020523bafbed7e9a80d21aea4dc7061d5d3\": container with ID starting with 2332540b100a5965287eba0c79003020523bafbed7e9a80d21aea4dc7061d5d3 not found: ID does not exist" Nov 25 10:57:51 crc kubenswrapper[4821]: I1125 10:57:51.033799 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-9c2a-account-create-r76sx"] Nov 25 10:57:51 crc kubenswrapper[4821]: I1125 10:57:51.043961 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-dzlxn"] Nov 25 10:57:51 crc kubenswrapper[4821]: I1125 10:57:51.051865 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-9c2a-account-create-r76sx"] Nov 25 10:57:51 crc kubenswrapper[4821]: I1125 10:57:51.059964 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-dzlxn"] Nov 25 10:57:51 crc kubenswrapper[4821]: I1125 10:57:51.067236 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-5ef2-account-create-mmmzg"] Nov 25 10:57:51 crc kubenswrapper[4821]: I1125 10:57:51.075549 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-5ef2-account-create-mmmzg"] Nov 25 10:57:51 crc kubenswrapper[4821]: I1125 10:57:51.309949 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03dec67a-7050-4d6a-81ae-7a41c494572f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "03dec67a-7050-4d6a-81ae-7a41c494572f" (UID: "03dec67a-7050-4d6a-81ae-7a41c494572f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:57:51 crc kubenswrapper[4821]: I1125 10:57:51.388068 4821 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03dec67a-7050-4d6a-81ae-7a41c494572f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:51 crc kubenswrapper[4821]: I1125 10:57:51.445535 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qsbgm"] Nov 25 10:57:51 crc kubenswrapper[4821]: I1125 10:57:51.452690 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-qsbgm"] Nov 25 10:57:52 crc kubenswrapper[4821]: I1125 10:57:52.125560 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03dec67a-7050-4d6a-81ae-7a41c494572f" path="/var/lib/kubelet/pods/03dec67a-7050-4d6a-81ae-7a41c494572f/volumes" Nov 25 10:57:52 crc kubenswrapper[4821]: I1125 10:57:52.126637 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0685db28-2cf1-4e13-b6d8-d37983c03d42" path="/var/lib/kubelet/pods/0685db28-2cf1-4e13-b6d8-d37983c03d42/volumes" Nov 25 10:57:52 crc kubenswrapper[4821]: I1125 10:57:52.127332 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="52803d63-d5ab-42c6-a17e-fcef8e92ea44" path="/var/lib/kubelet/pods/52803d63-d5ab-42c6-a17e-fcef8e92ea44/volumes" Nov 25 10:57:52 crc kubenswrapper[4821]: I1125 10:57:52.128621 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ed5a7f1-1e7b-438a-b083-f7f45b28443a" path="/var/lib/kubelet/pods/7ed5a7f1-1e7b-438a-b083-f7f45b28443a/volumes" Nov 25 10:58:00 crc kubenswrapper[4821]: I1125 10:58:00.322270 4821 generic.go:334] "Generic (PLEG): container finished" podID="0b93320f-a41f-4c0e-874f-b4f945f14c90" containerID="39033e10b4d59c74ae7b891bd41f3fdc0693d727c43e7840f122c90546b40d80" exitCode=0 Nov 25 10:58:00 crc kubenswrapper[4821]: I1125 10:58:00.322370 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pkx5r" event={"ID":"0b93320f-a41f-4c0e-874f-b4f945f14c90","Type":"ContainerDied","Data":"39033e10b4d59c74ae7b891bd41f3fdc0693d727c43e7840f122c90546b40d80"} Nov 25 10:58:01 crc kubenswrapper[4821]: I1125 10:58:01.729881 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pkx5r" Nov 25 10:58:01 crc kubenswrapper[4821]: I1125 10:58:01.793803 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0b93320f-a41f-4c0e-874f-b4f945f14c90-inventory\") pod \"0b93320f-a41f-4c0e-874f-b4f945f14c90\" (UID: \"0b93320f-a41f-4c0e-874f-b4f945f14c90\") " Nov 25 10:58:01 crc kubenswrapper[4821]: I1125 10:58:01.793878 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b93320f-a41f-4c0e-874f-b4f945f14c90-bootstrap-combined-ca-bundle\") pod \"0b93320f-a41f-4c0e-874f-b4f945f14c90\" (UID: \"0b93320f-a41f-4c0e-874f-b4f945f14c90\") " Nov 25 10:58:01 crc kubenswrapper[4821]: I1125 10:58:01.793931 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vdg26\" (UniqueName: \"kubernetes.io/projected/0b93320f-a41f-4c0e-874f-b4f945f14c90-kube-api-access-vdg26\") pod \"0b93320f-a41f-4c0e-874f-b4f945f14c90\" (UID: \"0b93320f-a41f-4c0e-874f-b4f945f14c90\") " Nov 25 10:58:01 crc kubenswrapper[4821]: I1125 10:58:01.793992 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0b93320f-a41f-4c0e-874f-b4f945f14c90-ssh-key\") pod \"0b93320f-a41f-4c0e-874f-b4f945f14c90\" (UID: \"0b93320f-a41f-4c0e-874f-b4f945f14c90\") " Nov 25 10:58:01 crc kubenswrapper[4821]: I1125 10:58:01.799863 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b93320f-a41f-4c0e-874f-b4f945f14c90-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "0b93320f-a41f-4c0e-874f-b4f945f14c90" (UID: "0b93320f-a41f-4c0e-874f-b4f945f14c90"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:58:01 crc kubenswrapper[4821]: I1125 10:58:01.800064 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b93320f-a41f-4c0e-874f-b4f945f14c90-kube-api-access-vdg26" (OuterVolumeSpecName: "kube-api-access-vdg26") pod "0b93320f-a41f-4c0e-874f-b4f945f14c90" (UID: "0b93320f-a41f-4c0e-874f-b4f945f14c90"). InnerVolumeSpecName "kube-api-access-vdg26". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:58:01 crc kubenswrapper[4821]: I1125 10:58:01.820743 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b93320f-a41f-4c0e-874f-b4f945f14c90-inventory" (OuterVolumeSpecName: "inventory") pod "0b93320f-a41f-4c0e-874f-b4f945f14c90" (UID: "0b93320f-a41f-4c0e-874f-b4f945f14c90"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:58:01 crc kubenswrapper[4821]: I1125 10:58:01.821203 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b93320f-a41f-4c0e-874f-b4f945f14c90-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0b93320f-a41f-4c0e-874f-b4f945f14c90" (UID: "0b93320f-a41f-4c0e-874f-b4f945f14c90"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:58:01 crc kubenswrapper[4821]: I1125 10:58:01.896038 4821 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0b93320f-a41f-4c0e-874f-b4f945f14c90-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 10:58:01 crc kubenswrapper[4821]: I1125 10:58:01.896088 4821 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b93320f-a41f-4c0e-874f-b4f945f14c90-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:58:01 crc kubenswrapper[4821]: I1125 10:58:01.896104 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vdg26\" (UniqueName: \"kubernetes.io/projected/0b93320f-a41f-4c0e-874f-b4f945f14c90-kube-api-access-vdg26\") on node \"crc\" DevicePath \"\"" Nov 25 10:58:01 crc kubenswrapper[4821]: I1125 10:58:01.896116 4821 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0b93320f-a41f-4c0e-874f-b4f945f14c90-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 10:58:02 crc kubenswrapper[4821]: I1125 10:58:02.338771 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pkx5r" event={"ID":"0b93320f-a41f-4c0e-874f-b4f945f14c90","Type":"ContainerDied","Data":"d7f4a85585c2000173582fdcd64f7070858b4170674fd53298a6c7eacd454786"} Nov 25 10:58:02 crc kubenswrapper[4821]: I1125 10:58:02.338818 4821 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d7f4a85585c2000173582fdcd64f7070858b4170674fd53298a6c7eacd454786" Nov 25 10:58:02 crc kubenswrapper[4821]: I1125 10:58:02.338816 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-pkx5r" Nov 25 10:58:02 crc kubenswrapper[4821]: I1125 10:58:02.422302 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-tfcn9"] Nov 25 10:58:02 crc kubenswrapper[4821]: E1125 10:58:02.422802 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03dec67a-7050-4d6a-81ae-7a41c494572f" containerName="registry-server" Nov 25 10:58:02 crc kubenswrapper[4821]: I1125 10:58:02.422825 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="03dec67a-7050-4d6a-81ae-7a41c494572f" containerName="registry-server" Nov 25 10:58:02 crc kubenswrapper[4821]: E1125 10:58:02.422844 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03dec67a-7050-4d6a-81ae-7a41c494572f" containerName="extract-utilities" Nov 25 10:58:02 crc kubenswrapper[4821]: I1125 10:58:02.422852 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="03dec67a-7050-4d6a-81ae-7a41c494572f" containerName="extract-utilities" Nov 25 10:58:02 crc kubenswrapper[4821]: E1125 10:58:02.422880 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b93320f-a41f-4c0e-874f-b4f945f14c90" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 25 10:58:02 crc kubenswrapper[4821]: I1125 10:58:02.422891 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b93320f-a41f-4c0e-874f-b4f945f14c90" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 25 10:58:02 crc kubenswrapper[4821]: E1125 10:58:02.422925 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03dec67a-7050-4d6a-81ae-7a41c494572f" containerName="extract-content" Nov 25 10:58:02 crc kubenswrapper[4821]: I1125 10:58:02.422933 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="03dec67a-7050-4d6a-81ae-7a41c494572f" containerName="extract-content" Nov 25 10:58:02 crc kubenswrapper[4821]: I1125 10:58:02.423180 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="03dec67a-7050-4d6a-81ae-7a41c494572f" containerName="registry-server" Nov 25 10:58:02 crc kubenswrapper[4821]: I1125 10:58:02.423214 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b93320f-a41f-4c0e-874f-b4f945f14c90" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 25 10:58:02 crc kubenswrapper[4821]: I1125 10:58:02.423966 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-tfcn9" Nov 25 10:58:02 crc kubenswrapper[4821]: I1125 10:58:02.425757 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-z764x" Nov 25 10:58:02 crc kubenswrapper[4821]: I1125 10:58:02.426910 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 10:58:02 crc kubenswrapper[4821]: I1125 10:58:02.427089 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 10:58:02 crc kubenswrapper[4821]: I1125 10:58:02.427267 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 10:58:02 crc kubenswrapper[4821]: I1125 10:58:02.439285 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-tfcn9"] Nov 25 10:58:02 crc kubenswrapper[4821]: I1125 10:58:02.508492 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qlwjr\" (UniqueName: \"kubernetes.io/projected/fcf6a791-0724-4ac2-8881-c7f8804f6540-kube-api-access-qlwjr\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-tfcn9\" (UID: \"fcf6a791-0724-4ac2-8881-c7f8804f6540\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-tfcn9" Nov 25 10:58:02 crc kubenswrapper[4821]: I1125 10:58:02.508569 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fcf6a791-0724-4ac2-8881-c7f8804f6540-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-tfcn9\" (UID: \"fcf6a791-0724-4ac2-8881-c7f8804f6540\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-tfcn9" Nov 25 10:58:02 crc kubenswrapper[4821]: I1125 10:58:02.508594 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fcf6a791-0724-4ac2-8881-c7f8804f6540-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-tfcn9\" (UID: \"fcf6a791-0724-4ac2-8881-c7f8804f6540\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-tfcn9" Nov 25 10:58:02 crc kubenswrapper[4821]: I1125 10:58:02.610450 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qlwjr\" (UniqueName: \"kubernetes.io/projected/fcf6a791-0724-4ac2-8881-c7f8804f6540-kube-api-access-qlwjr\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-tfcn9\" (UID: \"fcf6a791-0724-4ac2-8881-c7f8804f6540\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-tfcn9" Nov 25 10:58:02 crc kubenswrapper[4821]: I1125 10:58:02.610512 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fcf6a791-0724-4ac2-8881-c7f8804f6540-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-tfcn9\" (UID: \"fcf6a791-0724-4ac2-8881-c7f8804f6540\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-tfcn9" Nov 25 10:58:02 crc kubenswrapper[4821]: I1125 10:58:02.610539 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fcf6a791-0724-4ac2-8881-c7f8804f6540-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-tfcn9\" (UID: \"fcf6a791-0724-4ac2-8881-c7f8804f6540\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-tfcn9" Nov 25 10:58:02 crc kubenswrapper[4821]: I1125 10:58:02.614406 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fcf6a791-0724-4ac2-8881-c7f8804f6540-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-tfcn9\" (UID: \"fcf6a791-0724-4ac2-8881-c7f8804f6540\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-tfcn9" Nov 25 10:58:02 crc kubenswrapper[4821]: I1125 10:58:02.614747 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fcf6a791-0724-4ac2-8881-c7f8804f6540-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-tfcn9\" (UID: \"fcf6a791-0724-4ac2-8881-c7f8804f6540\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-tfcn9" Nov 25 10:58:02 crc kubenswrapper[4821]: I1125 10:58:02.628901 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qlwjr\" (UniqueName: \"kubernetes.io/projected/fcf6a791-0724-4ac2-8881-c7f8804f6540-kube-api-access-qlwjr\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-tfcn9\" (UID: \"fcf6a791-0724-4ac2-8881-c7f8804f6540\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-tfcn9" Nov 25 10:58:02 crc kubenswrapper[4821]: I1125 10:58:02.746677 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-tfcn9" Nov 25 10:58:03 crc kubenswrapper[4821]: I1125 10:58:03.267081 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-tfcn9"] Nov 25 10:58:03 crc kubenswrapper[4821]: I1125 10:58:03.349318 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-tfcn9" event={"ID":"fcf6a791-0724-4ac2-8881-c7f8804f6540","Type":"ContainerStarted","Data":"6c5b0f2e4a99dba1ac7783cc046d579e61c1d3d66ce06b357ea2ed12c6005c79"} Nov 25 10:58:04 crc kubenswrapper[4821]: I1125 10:58:04.361204 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-tfcn9" event={"ID":"fcf6a791-0724-4ac2-8881-c7f8804f6540","Type":"ContainerStarted","Data":"114a19ba1968cb83990799ae8706d8b40e30ace961a55bdb46320e7559e6e2ca"} Nov 25 10:58:04 crc kubenswrapper[4821]: I1125 10:58:04.386728 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-tfcn9" podStartSLOduration=1.98337298 podStartE2EDuration="2.386709716s" podCreationTimestamp="2025-11-25 10:58:02 +0000 UTC" firstStartedPulling="2025-11-25 10:58:03.271527605 +0000 UTC m=+1553.807847452" lastFinishedPulling="2025-11-25 10:58:03.674864341 +0000 UTC m=+1554.211184188" observedRunningTime="2025-11-25 10:58:04.376779557 +0000 UTC m=+1554.913099404" watchObservedRunningTime="2025-11-25 10:58:04.386709716 +0000 UTC m=+1554.923029553" Nov 25 10:58:04 crc kubenswrapper[4821]: I1125 10:58:04.715332 4821 patch_prober.go:28] interesting pod/machine-config-daemon-2krbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:58:04 crc kubenswrapper[4821]: I1125 10:58:04.715388 4821 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:58:22 crc kubenswrapper[4821]: I1125 10:58:22.483642 4821 scope.go:117] "RemoveContainer" containerID="8e6ed8fd87660ebcfb7743265992ee1ef1b5540b8654a15cbd85c05a88d48d39" Nov 25 10:58:22 crc kubenswrapper[4821]: I1125 10:58:22.514420 4821 scope.go:117] "RemoveContainer" containerID="5266b3949d733e0f91f36d6c1b21457c8f847bfd3498f3f8137f27d730d1b9bb" Nov 25 10:58:22 crc kubenswrapper[4821]: I1125 10:58:22.553620 4821 scope.go:117] "RemoveContainer" containerID="5fe0a53af69a3e9008fa221903a96105bc4aeebd297467181649e66df5d458d6" Nov 25 10:58:22 crc kubenswrapper[4821]: I1125 10:58:22.598621 4821 scope.go:117] "RemoveContainer" containerID="96b7eb0d2e5a8a8aa0bfb1dcf6c3663a69b3ad48e3c9efc121471d2b781c58e3" Nov 25 10:58:22 crc kubenswrapper[4821]: I1125 10:58:22.660523 4821 scope.go:117] "RemoveContainer" containerID="f939ace9603f0998d01d1c273b0f949305cc4aac5343cf21417718d0ccc3c1b0" Nov 25 10:58:22 crc kubenswrapper[4821]: I1125 10:58:22.685011 4821 scope.go:117] "RemoveContainer" containerID="9dc278880a147a8959c8dae238461baf2c4e42d54c0ed542c1d08aa498941b9a" Nov 25 10:58:22 crc kubenswrapper[4821]: I1125 10:58:22.734237 4821 scope.go:117] "RemoveContainer" containerID="b4c4327cc6c0a7cd77d3806069faf04d3f5c4c2d18f990f2d67b0909e777e4e8" Nov 25 10:58:22 crc kubenswrapper[4821]: I1125 10:58:22.755248 4821 scope.go:117] "RemoveContainer" containerID="2c49eaddef93d8519ac175e2b0c354fb919e1f6745d0a5c19c8cb6a79c6a1fc1" Nov 25 10:58:31 crc kubenswrapper[4821]: I1125 10:58:31.050867 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-7da5-account-create-8wqv6"] Nov 25 10:58:31 crc kubenswrapper[4821]: I1125 10:58:31.062584 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-55n47"] Nov 25 10:58:31 crc kubenswrapper[4821]: I1125 10:58:31.074236 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-ff23-account-create-n68t7"] Nov 25 10:58:31 crc kubenswrapper[4821]: I1125 10:58:31.082571 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-55n47"] Nov 25 10:58:31 crc kubenswrapper[4821]: I1125 10:58:31.089891 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-7da5-account-create-8wqv6"] Nov 25 10:58:31 crc kubenswrapper[4821]: I1125 10:58:31.097136 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-a9df-account-create-xlkf6"] Nov 25 10:58:31 crc kubenswrapper[4821]: I1125 10:58:31.103846 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-ff23-account-create-n68t7"] Nov 25 10:58:31 crc kubenswrapper[4821]: I1125 10:58:31.110493 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-xpctj"] Nov 25 10:58:31 crc kubenswrapper[4821]: I1125 10:58:31.117101 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-a9df-account-create-xlkf6"] Nov 25 10:58:31 crc kubenswrapper[4821]: I1125 10:58:31.123933 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-xpctj"] Nov 25 10:58:31 crc kubenswrapper[4821]: I1125 10:58:31.135313 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-sq6xx"] Nov 25 10:58:31 crc kubenswrapper[4821]: I1125 10:58:31.142840 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-sq6xx"] Nov 25 10:58:32 crc kubenswrapper[4821]: I1125 10:58:32.125448 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d5f3fd0-3830-4079-bbf2-0968a8a454f1" path="/var/lib/kubelet/pods/2d5f3fd0-3830-4079-bbf2-0968a8a454f1/volumes" Nov 25 10:58:32 crc kubenswrapper[4821]: I1125 10:58:32.126033 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3049cd6a-c38c-4da9-83ce-2195225c3ac9" path="/var/lib/kubelet/pods/3049cd6a-c38c-4da9-83ce-2195225c3ac9/volumes" Nov 25 10:58:32 crc kubenswrapper[4821]: I1125 10:58:32.126566 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65d5f171-33f0-4c2d-9cfb-6847f2b98c49" path="/var/lib/kubelet/pods/65d5f171-33f0-4c2d-9cfb-6847f2b98c49/volumes" Nov 25 10:58:32 crc kubenswrapper[4821]: I1125 10:58:32.127102 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a764b1fe-69e0-4609-814e-ab2d4e8b4438" path="/var/lib/kubelet/pods/a764b1fe-69e0-4609-814e-ab2d4e8b4438/volumes" Nov 25 10:58:32 crc kubenswrapper[4821]: I1125 10:58:32.128286 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba88f146-f158-4703-b77d-319f4f65d419" path="/var/lib/kubelet/pods/ba88f146-f158-4703-b77d-319f4f65d419/volumes" Nov 25 10:58:32 crc kubenswrapper[4821]: I1125 10:58:32.128966 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cec42c9b-c925-4248-b73c-a076d1d0a8c4" path="/var/lib/kubelet/pods/cec42c9b-c925-4248-b73c-a076d1d0a8c4/volumes" Nov 25 10:58:34 crc kubenswrapper[4821]: I1125 10:58:34.715295 4821 patch_prober.go:28] interesting pod/machine-config-daemon-2krbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:58:34 crc kubenswrapper[4821]: I1125 10:58:34.716041 4821 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:58:37 crc kubenswrapper[4821]: I1125 10:58:37.029182 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-2nhpq"] Nov 25 10:58:37 crc kubenswrapper[4821]: I1125 10:58:37.039480 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-2nhpq"] Nov 25 10:58:38 crc kubenswrapper[4821]: I1125 10:58:38.126889 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a8123055-1953-42b3-9c04-716061499d03" path="/var/lib/kubelet/pods/a8123055-1953-42b3-9c04-716061499d03/volumes" Nov 25 10:58:50 crc kubenswrapper[4821]: I1125 10:58:50.032972 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-bhnq8"] Nov 25 10:58:50 crc kubenswrapper[4821]: I1125 10:58:50.045143 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-bhnq8"] Nov 25 10:58:50 crc kubenswrapper[4821]: I1125 10:58:50.125688 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dbdf62cf-071e-4e1c-9ef6-22bc39cfa795" path="/var/lib/kubelet/pods/dbdf62cf-071e-4e1c-9ef6-22bc39cfa795/volumes" Nov 25 10:59:04 crc kubenswrapper[4821]: I1125 10:59:04.715084 4821 patch_prober.go:28] interesting pod/machine-config-daemon-2krbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:59:04 crc kubenswrapper[4821]: I1125 10:59:04.716099 4821 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:59:04 crc kubenswrapper[4821]: I1125 10:59:04.716200 4821 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" Nov 25 10:59:04 crc kubenswrapper[4821]: I1125 10:59:04.717381 4821 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fc16b2ea0cd36da4cef842d21755ec621134a4c21c5c964f8856fb8e6a102b2d"} pod="openshift-machine-config-operator/machine-config-daemon-2krbf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 10:59:04 crc kubenswrapper[4821]: I1125 10:59:04.717448 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerName="machine-config-daemon" containerID="cri-o://fc16b2ea0cd36da4cef842d21755ec621134a4c21c5c964f8856fb8e6a102b2d" gracePeriod=600 Nov 25 10:59:05 crc kubenswrapper[4821]: I1125 10:59:05.946316 4821 generic.go:334] "Generic (PLEG): container finished" podID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerID="fc16b2ea0cd36da4cef842d21755ec621134a4c21c5c964f8856fb8e6a102b2d" exitCode=0 Nov 25 10:59:05 crc kubenswrapper[4821]: I1125 10:59:05.946410 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" event={"ID":"5f948b87-ac86-4de6-ad64-c2ef947f84d4","Type":"ContainerDied","Data":"fc16b2ea0cd36da4cef842d21755ec621134a4c21c5c964f8856fb8e6a102b2d"} Nov 25 10:59:05 crc kubenswrapper[4821]: I1125 10:59:05.946861 4821 scope.go:117] "RemoveContainer" containerID="1cba28a33f40c656b6dedea9b9df6fa03370b5e6b1aff86ef7fcc571ea800429" Nov 25 10:59:06 crc kubenswrapper[4821]: E1125 10:59:06.100289 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 10:59:06 crc kubenswrapper[4821]: I1125 10:59:06.957958 4821 scope.go:117] "RemoveContainer" containerID="fc16b2ea0cd36da4cef842d21755ec621134a4c21c5c964f8856fb8e6a102b2d" Nov 25 10:59:06 crc kubenswrapper[4821]: E1125 10:59:06.960133 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 10:59:18 crc kubenswrapper[4821]: I1125 10:59:18.114566 4821 scope.go:117] "RemoveContainer" containerID="fc16b2ea0cd36da4cef842d21755ec621134a4c21c5c964f8856fb8e6a102b2d" Nov 25 10:59:18 crc kubenswrapper[4821]: E1125 10:59:18.115406 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 10:59:22 crc kubenswrapper[4821]: I1125 10:59:22.948934 4821 scope.go:117] "RemoveContainer" containerID="71f2dc7f92b4c1969d58583c104e8d6f5aae05207de27f3adf8134f2ab203e20" Nov 25 10:59:22 crc kubenswrapper[4821]: I1125 10:59:22.978054 4821 scope.go:117] "RemoveContainer" containerID="4e9bd31c0491a3c78124341f952f334b25e755f3eac5df640fcc5a9961e6359f" Nov 25 10:59:23 crc kubenswrapper[4821]: I1125 10:59:23.083764 4821 scope.go:117] "RemoveContainer" containerID="93f388ffcf02e4a8292f29e53d8423f497f064c3108c92b59c7e7f8ea4230679" Nov 25 10:59:23 crc kubenswrapper[4821]: I1125 10:59:23.113643 4821 scope.go:117] "RemoveContainer" containerID="19b20fe65f03e445fec02e7da2a869b70f0effcc221783c3f191c8e4dde87178" Nov 25 10:59:23 crc kubenswrapper[4821]: I1125 10:59:23.166864 4821 scope.go:117] "RemoveContainer" containerID="c8ccf9a6a3df8286261caebc19be09c2fc61f8c79473aded6a8b5d775e5bfc6a" Nov 25 10:59:23 crc kubenswrapper[4821]: I1125 10:59:23.238305 4821 scope.go:117] "RemoveContainer" containerID="3c1126a2eb0007e787199e1622a82fcd8e726c76a4cdb353b4ac66d5433a61dd" Nov 25 10:59:23 crc kubenswrapper[4821]: I1125 10:59:23.260505 4821 scope.go:117] "RemoveContainer" containerID="2a62662f27a8cfece54ff23a865831648780fa1fb785af45951a406accc7c6a9" Nov 25 10:59:23 crc kubenswrapper[4821]: I1125 10:59:23.281849 4821 scope.go:117] "RemoveContainer" containerID="eef603c99d063234ffafa903350973a5c52aabccededa1c1088b6e9c03b89934" Nov 25 10:59:30 crc kubenswrapper[4821]: I1125 10:59:30.077488 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-bkbhq"] Nov 25 10:59:30 crc kubenswrapper[4821]: I1125 10:59:30.192058 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-bkbhq"] Nov 25 10:59:32 crc kubenswrapper[4821]: I1125 10:59:32.115411 4821 scope.go:117] "RemoveContainer" containerID="fc16b2ea0cd36da4cef842d21755ec621134a4c21c5c964f8856fb8e6a102b2d" Nov 25 10:59:32 crc kubenswrapper[4821]: E1125 10:59:32.115965 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 10:59:32 crc kubenswrapper[4821]: I1125 10:59:32.128059 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5302114a-3896-429e-9693-3483c106730a" path="/var/lib/kubelet/pods/5302114a-3896-429e-9693-3483c106730a/volumes" Nov 25 10:59:36 crc kubenswrapper[4821]: I1125 10:59:36.027365 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-tq9mv"] Nov 25 10:59:36 crc kubenswrapper[4821]: I1125 10:59:36.038671 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-tq9mv"] Nov 25 10:59:36 crc kubenswrapper[4821]: I1125 10:59:36.129944 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="325653d2-264c-4c6c-851f-6e38fe3e61a5" path="/var/lib/kubelet/pods/325653d2-264c-4c6c-851f-6e38fe3e61a5/volumes" Nov 25 10:59:37 crc kubenswrapper[4821]: I1125 10:59:37.032396 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-xxdmc"] Nov 25 10:59:37 crc kubenswrapper[4821]: I1125 10:59:37.040023 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-xxdmc"] Nov 25 10:59:38 crc kubenswrapper[4821]: I1125 10:59:38.125890 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bee9a405-f60c-449e-beba-71cd802daa79" path="/var/lib/kubelet/pods/bee9a405-f60c-449e-beba-71cd802daa79/volumes" Nov 25 10:59:45 crc kubenswrapper[4821]: I1125 10:59:45.114965 4821 scope.go:117] "RemoveContainer" containerID="fc16b2ea0cd36da4cef842d21755ec621134a4c21c5c964f8856fb8e6a102b2d" Nov 25 10:59:45 crc kubenswrapper[4821]: E1125 10:59:45.115818 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 10:59:47 crc kubenswrapper[4821]: I1125 10:59:47.037488 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-7znwh"] Nov 25 10:59:47 crc kubenswrapper[4821]: I1125 10:59:47.047105 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-8bp9n"] Nov 25 10:59:47 crc kubenswrapper[4821]: I1125 10:59:47.056403 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-7znwh"] Nov 25 10:59:47 crc kubenswrapper[4821]: I1125 10:59:47.064078 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-8bp9n"] Nov 25 10:59:48 crc kubenswrapper[4821]: I1125 10:59:48.128258 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2eb6852b-84a1-4d4a-8402-77a7e2252ab2" path="/var/lib/kubelet/pods/2eb6852b-84a1-4d4a-8402-77a7e2252ab2/volumes" Nov 25 10:59:48 crc kubenswrapper[4821]: I1125 10:59:48.129285 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb6c9acf-d69c-4788-a85d-b72135de1ef3" path="/var/lib/kubelet/pods/eb6c9acf-d69c-4788-a85d-b72135de1ef3/volumes" Nov 25 10:59:48 crc kubenswrapper[4821]: I1125 10:59:48.356674 4821 generic.go:334] "Generic (PLEG): container finished" podID="fcf6a791-0724-4ac2-8881-c7f8804f6540" containerID="114a19ba1968cb83990799ae8706d8b40e30ace961a55bdb46320e7559e6e2ca" exitCode=0 Nov 25 10:59:48 crc kubenswrapper[4821]: I1125 10:59:48.356735 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-tfcn9" event={"ID":"fcf6a791-0724-4ac2-8881-c7f8804f6540","Type":"ContainerDied","Data":"114a19ba1968cb83990799ae8706d8b40e30ace961a55bdb46320e7559e6e2ca"} Nov 25 10:59:49 crc kubenswrapper[4821]: I1125 10:59:49.769048 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-tfcn9" Nov 25 10:59:49 crc kubenswrapper[4821]: I1125 10:59:49.946395 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fcf6a791-0724-4ac2-8881-c7f8804f6540-ssh-key\") pod \"fcf6a791-0724-4ac2-8881-c7f8804f6540\" (UID: \"fcf6a791-0724-4ac2-8881-c7f8804f6540\") " Nov 25 10:59:49 crc kubenswrapper[4821]: I1125 10:59:49.946835 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fcf6a791-0724-4ac2-8881-c7f8804f6540-inventory\") pod \"fcf6a791-0724-4ac2-8881-c7f8804f6540\" (UID: \"fcf6a791-0724-4ac2-8881-c7f8804f6540\") " Nov 25 10:59:49 crc kubenswrapper[4821]: I1125 10:59:49.947092 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qlwjr\" (UniqueName: \"kubernetes.io/projected/fcf6a791-0724-4ac2-8881-c7f8804f6540-kube-api-access-qlwjr\") pod \"fcf6a791-0724-4ac2-8881-c7f8804f6540\" (UID: \"fcf6a791-0724-4ac2-8881-c7f8804f6540\") " Nov 25 10:59:49 crc kubenswrapper[4821]: I1125 10:59:49.954042 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fcf6a791-0724-4ac2-8881-c7f8804f6540-kube-api-access-qlwjr" (OuterVolumeSpecName: "kube-api-access-qlwjr") pod "fcf6a791-0724-4ac2-8881-c7f8804f6540" (UID: "fcf6a791-0724-4ac2-8881-c7f8804f6540"). InnerVolumeSpecName "kube-api-access-qlwjr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:59:49 crc kubenswrapper[4821]: I1125 10:59:49.975063 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fcf6a791-0724-4ac2-8881-c7f8804f6540-inventory" (OuterVolumeSpecName: "inventory") pod "fcf6a791-0724-4ac2-8881-c7f8804f6540" (UID: "fcf6a791-0724-4ac2-8881-c7f8804f6540"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:59:49 crc kubenswrapper[4821]: I1125 10:59:49.976661 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fcf6a791-0724-4ac2-8881-c7f8804f6540-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "fcf6a791-0724-4ac2-8881-c7f8804f6540" (UID: "fcf6a791-0724-4ac2-8881-c7f8804f6540"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:59:50 crc kubenswrapper[4821]: I1125 10:59:50.049181 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qlwjr\" (UniqueName: \"kubernetes.io/projected/fcf6a791-0724-4ac2-8881-c7f8804f6540-kube-api-access-qlwjr\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:50 crc kubenswrapper[4821]: I1125 10:59:50.049215 4821 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fcf6a791-0724-4ac2-8881-c7f8804f6540-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:50 crc kubenswrapper[4821]: I1125 10:59:50.049224 4821 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fcf6a791-0724-4ac2-8881-c7f8804f6540-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:50 crc kubenswrapper[4821]: I1125 10:59:50.375612 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-tfcn9" Nov 25 10:59:50 crc kubenswrapper[4821]: I1125 10:59:50.375629 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-tfcn9" event={"ID":"fcf6a791-0724-4ac2-8881-c7f8804f6540","Type":"ContainerDied","Data":"6c5b0f2e4a99dba1ac7783cc046d579e61c1d3d66ce06b357ea2ed12c6005c79"} Nov 25 10:59:50 crc kubenswrapper[4821]: I1125 10:59:50.375922 4821 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6c5b0f2e4a99dba1ac7783cc046d579e61c1d3d66ce06b357ea2ed12c6005c79" Nov 25 10:59:50 crc kubenswrapper[4821]: I1125 10:59:50.454041 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-sg8cc"] Nov 25 10:59:50 crc kubenswrapper[4821]: E1125 10:59:50.454454 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcf6a791-0724-4ac2-8881-c7f8804f6540" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Nov 25 10:59:50 crc kubenswrapper[4821]: I1125 10:59:50.454473 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcf6a791-0724-4ac2-8881-c7f8804f6540" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Nov 25 10:59:50 crc kubenswrapper[4821]: I1125 10:59:50.454652 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="fcf6a791-0724-4ac2-8881-c7f8804f6540" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Nov 25 10:59:50 crc kubenswrapper[4821]: I1125 10:59:50.456365 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-sg8cc" Nov 25 10:59:50 crc kubenswrapper[4821]: I1125 10:59:50.459389 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 10:59:50 crc kubenswrapper[4821]: I1125 10:59:50.459491 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-z764x" Nov 25 10:59:50 crc kubenswrapper[4821]: I1125 10:59:50.459712 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 10:59:50 crc kubenswrapper[4821]: I1125 10:59:50.459860 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 10:59:50 crc kubenswrapper[4821]: I1125 10:59:50.475296 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-sg8cc"] Nov 25 10:59:50 crc kubenswrapper[4821]: I1125 10:59:50.560860 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7jhk\" (UniqueName: \"kubernetes.io/projected/5a15c7fc-4220-46ae-8af3-490c28f8b4ff-kube-api-access-h7jhk\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-sg8cc\" (UID: \"5a15c7fc-4220-46ae-8af3-490c28f8b4ff\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-sg8cc" Nov 25 10:59:50 crc kubenswrapper[4821]: I1125 10:59:50.561279 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5a15c7fc-4220-46ae-8af3-490c28f8b4ff-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-sg8cc\" (UID: \"5a15c7fc-4220-46ae-8af3-490c28f8b4ff\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-sg8cc" Nov 25 10:59:50 crc kubenswrapper[4821]: I1125 10:59:50.561418 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5a15c7fc-4220-46ae-8af3-490c28f8b4ff-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-sg8cc\" (UID: \"5a15c7fc-4220-46ae-8af3-490c28f8b4ff\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-sg8cc" Nov 25 10:59:50 crc kubenswrapper[4821]: I1125 10:59:50.662761 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5a15c7fc-4220-46ae-8af3-490c28f8b4ff-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-sg8cc\" (UID: \"5a15c7fc-4220-46ae-8af3-490c28f8b4ff\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-sg8cc" Nov 25 10:59:50 crc kubenswrapper[4821]: I1125 10:59:50.662865 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5a15c7fc-4220-46ae-8af3-490c28f8b4ff-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-sg8cc\" (UID: \"5a15c7fc-4220-46ae-8af3-490c28f8b4ff\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-sg8cc" Nov 25 10:59:50 crc kubenswrapper[4821]: I1125 10:59:50.662952 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7jhk\" (UniqueName: \"kubernetes.io/projected/5a15c7fc-4220-46ae-8af3-490c28f8b4ff-kube-api-access-h7jhk\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-sg8cc\" (UID: \"5a15c7fc-4220-46ae-8af3-490c28f8b4ff\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-sg8cc" Nov 25 10:59:50 crc kubenswrapper[4821]: I1125 10:59:50.668202 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5a15c7fc-4220-46ae-8af3-490c28f8b4ff-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-sg8cc\" (UID: \"5a15c7fc-4220-46ae-8af3-490c28f8b4ff\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-sg8cc" Nov 25 10:59:50 crc kubenswrapper[4821]: I1125 10:59:50.670235 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5a15c7fc-4220-46ae-8af3-490c28f8b4ff-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-sg8cc\" (UID: \"5a15c7fc-4220-46ae-8af3-490c28f8b4ff\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-sg8cc" Nov 25 10:59:50 crc kubenswrapper[4821]: I1125 10:59:50.690722 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7jhk\" (UniqueName: \"kubernetes.io/projected/5a15c7fc-4220-46ae-8af3-490c28f8b4ff-kube-api-access-h7jhk\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-sg8cc\" (UID: \"5a15c7fc-4220-46ae-8af3-490c28f8b4ff\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-sg8cc" Nov 25 10:59:50 crc kubenswrapper[4821]: I1125 10:59:50.785106 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-sg8cc" Nov 25 10:59:51 crc kubenswrapper[4821]: I1125 10:59:51.353716 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-sg8cc"] Nov 25 10:59:51 crc kubenswrapper[4821]: W1125 10:59:51.358980 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5a15c7fc_4220_46ae_8af3_490c28f8b4ff.slice/crio-c0660d7ce48a3a08a041caf378fbf5aa080a624da1cd1ef152b2d0d54f5c4cca WatchSource:0}: Error finding container c0660d7ce48a3a08a041caf378fbf5aa080a624da1cd1ef152b2d0d54f5c4cca: Status 404 returned error can't find the container with id c0660d7ce48a3a08a041caf378fbf5aa080a624da1cd1ef152b2d0d54f5c4cca Nov 25 10:59:51 crc kubenswrapper[4821]: I1125 10:59:51.387033 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-sg8cc" event={"ID":"5a15c7fc-4220-46ae-8af3-490c28f8b4ff","Type":"ContainerStarted","Data":"c0660d7ce48a3a08a041caf378fbf5aa080a624da1cd1ef152b2d0d54f5c4cca"} Nov 25 10:59:56 crc kubenswrapper[4821]: I1125 10:59:56.442194 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-sg8cc" event={"ID":"5a15c7fc-4220-46ae-8af3-490c28f8b4ff","Type":"ContainerStarted","Data":"9c1876861e12a0eb2418f2fb1f64571f41412a1eb78e6fb65c29919b3f2f4449"} Nov 25 10:59:56 crc kubenswrapper[4821]: I1125 10:59:56.472527 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-sg8cc" podStartSLOduration=2.506344205 podStartE2EDuration="6.472507029s" podCreationTimestamp="2025-11-25 10:59:50 +0000 UTC" firstStartedPulling="2025-11-25 10:59:51.361827689 +0000 UTC m=+1661.898147536" lastFinishedPulling="2025-11-25 10:59:55.327990523 +0000 UTC m=+1665.864310360" observedRunningTime="2025-11-25 10:59:56.463000762 +0000 UTC m=+1666.999320619" watchObservedRunningTime="2025-11-25 10:59:56.472507029 +0000 UTC m=+1667.008826896" Nov 25 11:00:00 crc kubenswrapper[4821]: I1125 11:00:00.129530 4821 scope.go:117] "RemoveContainer" containerID="fc16b2ea0cd36da4cef842d21755ec621134a4c21c5c964f8856fb8e6a102b2d" Nov 25 11:00:00 crc kubenswrapper[4821]: E1125 11:00:00.130375 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:00:00 crc kubenswrapper[4821]: I1125 11:00:00.161538 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401140-cqkws"] Nov 25 11:00:00 crc kubenswrapper[4821]: I1125 11:00:00.162975 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401140-cqkws" Nov 25 11:00:00 crc kubenswrapper[4821]: I1125 11:00:00.165676 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 25 11:00:00 crc kubenswrapper[4821]: I1125 11:00:00.165897 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 25 11:00:00 crc kubenswrapper[4821]: I1125 11:00:00.185047 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401140-cqkws"] Nov 25 11:00:00 crc kubenswrapper[4821]: I1125 11:00:00.269935 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2chl\" (UniqueName: \"kubernetes.io/projected/6d1e4062-40fb-4409-823a-3e1b5023dea0-kube-api-access-f2chl\") pod \"collect-profiles-29401140-cqkws\" (UID: \"6d1e4062-40fb-4409-823a-3e1b5023dea0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401140-cqkws" Nov 25 11:00:00 crc kubenswrapper[4821]: I1125 11:00:00.270047 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6d1e4062-40fb-4409-823a-3e1b5023dea0-secret-volume\") pod \"collect-profiles-29401140-cqkws\" (UID: \"6d1e4062-40fb-4409-823a-3e1b5023dea0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401140-cqkws" Nov 25 11:00:00 crc kubenswrapper[4821]: I1125 11:00:00.270274 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6d1e4062-40fb-4409-823a-3e1b5023dea0-config-volume\") pod \"collect-profiles-29401140-cqkws\" (UID: \"6d1e4062-40fb-4409-823a-3e1b5023dea0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401140-cqkws" Nov 25 11:00:00 crc kubenswrapper[4821]: I1125 11:00:00.373895 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6d1e4062-40fb-4409-823a-3e1b5023dea0-config-volume\") pod \"collect-profiles-29401140-cqkws\" (UID: \"6d1e4062-40fb-4409-823a-3e1b5023dea0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401140-cqkws" Nov 25 11:00:00 crc kubenswrapper[4821]: I1125 11:00:00.374270 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2chl\" (UniqueName: \"kubernetes.io/projected/6d1e4062-40fb-4409-823a-3e1b5023dea0-kube-api-access-f2chl\") pod \"collect-profiles-29401140-cqkws\" (UID: \"6d1e4062-40fb-4409-823a-3e1b5023dea0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401140-cqkws" Nov 25 11:00:00 crc kubenswrapper[4821]: I1125 11:00:00.374398 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6d1e4062-40fb-4409-823a-3e1b5023dea0-secret-volume\") pod \"collect-profiles-29401140-cqkws\" (UID: \"6d1e4062-40fb-4409-823a-3e1b5023dea0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401140-cqkws" Nov 25 11:00:00 crc kubenswrapper[4821]: I1125 11:00:00.374780 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6d1e4062-40fb-4409-823a-3e1b5023dea0-config-volume\") pod \"collect-profiles-29401140-cqkws\" (UID: \"6d1e4062-40fb-4409-823a-3e1b5023dea0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401140-cqkws" Nov 25 11:00:00 crc kubenswrapper[4821]: I1125 11:00:00.391286 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6d1e4062-40fb-4409-823a-3e1b5023dea0-secret-volume\") pod \"collect-profiles-29401140-cqkws\" (UID: \"6d1e4062-40fb-4409-823a-3e1b5023dea0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401140-cqkws" Nov 25 11:00:00 crc kubenswrapper[4821]: I1125 11:00:00.396090 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2chl\" (UniqueName: \"kubernetes.io/projected/6d1e4062-40fb-4409-823a-3e1b5023dea0-kube-api-access-f2chl\") pod \"collect-profiles-29401140-cqkws\" (UID: \"6d1e4062-40fb-4409-823a-3e1b5023dea0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401140-cqkws" Nov 25 11:00:00 crc kubenswrapper[4821]: I1125 11:00:00.495244 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401140-cqkws" Nov 25 11:00:00 crc kubenswrapper[4821]: I1125 11:00:00.972795 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401140-cqkws"] Nov 25 11:00:00 crc kubenswrapper[4821]: W1125 11:00:00.976506 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6d1e4062_40fb_4409_823a_3e1b5023dea0.slice/crio-b5da36b1b6b6a51abe892161f390cd08db3e9c50ed3d7d0430151d2c4c94bc7c WatchSource:0}: Error finding container b5da36b1b6b6a51abe892161f390cd08db3e9c50ed3d7d0430151d2c4c94bc7c: Status 404 returned error can't find the container with id b5da36b1b6b6a51abe892161f390cd08db3e9c50ed3d7d0430151d2c4c94bc7c Nov 25 11:00:01 crc kubenswrapper[4821]: I1125 11:00:01.487764 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401140-cqkws" event={"ID":"6d1e4062-40fb-4409-823a-3e1b5023dea0","Type":"ContainerStarted","Data":"68356fa839f71c30317b5cc58463ef2ee7d91f459aee6a8ebe563974d0d98aab"} Nov 25 11:00:01 crc kubenswrapper[4821]: I1125 11:00:01.488143 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401140-cqkws" event={"ID":"6d1e4062-40fb-4409-823a-3e1b5023dea0","Type":"ContainerStarted","Data":"b5da36b1b6b6a51abe892161f390cd08db3e9c50ed3d7d0430151d2c4c94bc7c"} Nov 25 11:00:02 crc kubenswrapper[4821]: I1125 11:00:02.498247 4821 generic.go:334] "Generic (PLEG): container finished" podID="6d1e4062-40fb-4409-823a-3e1b5023dea0" containerID="68356fa839f71c30317b5cc58463ef2ee7d91f459aee6a8ebe563974d0d98aab" exitCode=0 Nov 25 11:00:02 crc kubenswrapper[4821]: I1125 11:00:02.498330 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401140-cqkws" event={"ID":"6d1e4062-40fb-4409-823a-3e1b5023dea0","Type":"ContainerDied","Data":"68356fa839f71c30317b5cc58463ef2ee7d91f459aee6a8ebe563974d0d98aab"} Nov 25 11:00:03 crc kubenswrapper[4821]: I1125 11:00:03.877129 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401140-cqkws" Nov 25 11:00:04 crc kubenswrapper[4821]: I1125 11:00:04.043503 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6d1e4062-40fb-4409-823a-3e1b5023dea0-secret-volume\") pod \"6d1e4062-40fb-4409-823a-3e1b5023dea0\" (UID: \"6d1e4062-40fb-4409-823a-3e1b5023dea0\") " Nov 25 11:00:04 crc kubenswrapper[4821]: I1125 11:00:04.043585 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f2chl\" (UniqueName: \"kubernetes.io/projected/6d1e4062-40fb-4409-823a-3e1b5023dea0-kube-api-access-f2chl\") pod \"6d1e4062-40fb-4409-823a-3e1b5023dea0\" (UID: \"6d1e4062-40fb-4409-823a-3e1b5023dea0\") " Nov 25 11:00:04 crc kubenswrapper[4821]: I1125 11:00:04.043917 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6d1e4062-40fb-4409-823a-3e1b5023dea0-config-volume\") pod \"6d1e4062-40fb-4409-823a-3e1b5023dea0\" (UID: \"6d1e4062-40fb-4409-823a-3e1b5023dea0\") " Nov 25 11:00:04 crc kubenswrapper[4821]: I1125 11:00:04.045004 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d1e4062-40fb-4409-823a-3e1b5023dea0-config-volume" (OuterVolumeSpecName: "config-volume") pod "6d1e4062-40fb-4409-823a-3e1b5023dea0" (UID: "6d1e4062-40fb-4409-823a-3e1b5023dea0"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:00:04 crc kubenswrapper[4821]: I1125 11:00:04.050011 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d1e4062-40fb-4409-823a-3e1b5023dea0-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "6d1e4062-40fb-4409-823a-3e1b5023dea0" (UID: "6d1e4062-40fb-4409-823a-3e1b5023dea0"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:00:04 crc kubenswrapper[4821]: I1125 11:00:04.052449 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d1e4062-40fb-4409-823a-3e1b5023dea0-kube-api-access-f2chl" (OuterVolumeSpecName: "kube-api-access-f2chl") pod "6d1e4062-40fb-4409-823a-3e1b5023dea0" (UID: "6d1e4062-40fb-4409-823a-3e1b5023dea0"). InnerVolumeSpecName "kube-api-access-f2chl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:00:04 crc kubenswrapper[4821]: I1125 11:00:04.145926 4821 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6d1e4062-40fb-4409-823a-3e1b5023dea0-config-volume\") on node \"crc\" DevicePath \"\"" Nov 25 11:00:04 crc kubenswrapper[4821]: I1125 11:00:04.145959 4821 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6d1e4062-40fb-4409-823a-3e1b5023dea0-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 25 11:00:04 crc kubenswrapper[4821]: I1125 11:00:04.145971 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f2chl\" (UniqueName: \"kubernetes.io/projected/6d1e4062-40fb-4409-823a-3e1b5023dea0-kube-api-access-f2chl\") on node \"crc\" DevicePath \"\"" Nov 25 11:00:04 crc kubenswrapper[4821]: I1125 11:00:04.518742 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401140-cqkws" event={"ID":"6d1e4062-40fb-4409-823a-3e1b5023dea0","Type":"ContainerDied","Data":"b5da36b1b6b6a51abe892161f390cd08db3e9c50ed3d7d0430151d2c4c94bc7c"} Nov 25 11:00:04 crc kubenswrapper[4821]: I1125 11:00:04.519148 4821 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b5da36b1b6b6a51abe892161f390cd08db3e9c50ed3d7d0430151d2c4c94bc7c" Nov 25 11:00:04 crc kubenswrapper[4821]: I1125 11:00:04.518806 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401140-cqkws" Nov 25 11:00:13 crc kubenswrapper[4821]: I1125 11:00:13.114925 4821 scope.go:117] "RemoveContainer" containerID="fc16b2ea0cd36da4cef842d21755ec621134a4c21c5c964f8856fb8e6a102b2d" Nov 25 11:00:13 crc kubenswrapper[4821]: E1125 11:00:13.115708 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:00:23 crc kubenswrapper[4821]: I1125 11:00:23.474272 4821 scope.go:117] "RemoveContainer" containerID="068886c87f6a67c94577ca2971d3df071fe62df65a20f17589ce4f0a9d9008c5" Nov 25 11:00:23 crc kubenswrapper[4821]: I1125 11:00:23.861885 4821 scope.go:117] "RemoveContainer" containerID="c9151e75ad24f3fd80b6828468427f0883ab2f705ae36476b0fd3c2fa4619697" Nov 25 11:00:23 crc kubenswrapper[4821]: I1125 11:00:23.940439 4821 scope.go:117] "RemoveContainer" containerID="8f3e63e3b217fdcee82a9e2153ed5efa0a9ca89caff4b8c351478c6944df32c4" Nov 25 11:00:23 crc kubenswrapper[4821]: I1125 11:00:23.984391 4821 scope.go:117] "RemoveContainer" containerID="789f7caa916d8d2c62126a6a258d74071bb67e90d27c88e23ea85583c8c12ef0" Nov 25 11:00:24 crc kubenswrapper[4821]: I1125 11:00:24.033340 4821 scope.go:117] "RemoveContainer" containerID="391a9970a4c25e3326a56a91137910d12c345d3bc1f2e0e88b931ee47b675f82" Nov 25 11:00:26 crc kubenswrapper[4821]: I1125 11:00:26.115284 4821 scope.go:117] "RemoveContainer" containerID="fc16b2ea0cd36da4cef842d21755ec621134a4c21c5c964f8856fb8e6a102b2d" Nov 25 11:00:26 crc kubenswrapper[4821]: E1125 11:00:26.116778 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:00:41 crc kubenswrapper[4821]: I1125 11:00:41.114742 4821 scope.go:117] "RemoveContainer" containerID="fc16b2ea0cd36da4cef842d21755ec621134a4c21c5c964f8856fb8e6a102b2d" Nov 25 11:00:41 crc kubenswrapper[4821]: E1125 11:00:41.115671 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:00:49 crc kubenswrapper[4821]: I1125 11:00:49.043202 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-4n4fc"] Nov 25 11:00:49 crc kubenswrapper[4821]: I1125 11:00:49.051863 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-4n4fc"] Nov 25 11:00:50 crc kubenswrapper[4821]: I1125 11:00:50.028708 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-8787-account-create-grx4m"] Nov 25 11:00:50 crc kubenswrapper[4821]: I1125 11:00:50.035233 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-8787-account-create-grx4m"] Nov 25 11:00:50 crc kubenswrapper[4821]: I1125 11:00:50.129312 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3305a4a3-6c80-4595-9b1b-9adbbbd3c83a" path="/var/lib/kubelet/pods/3305a4a3-6c80-4595-9b1b-9adbbbd3c83a/volumes" Nov 25 11:00:50 crc kubenswrapper[4821]: I1125 11:00:50.130023 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63e2c033-bfd5-4f03-96de-5ab6333b179d" path="/var/lib/kubelet/pods/63e2c033-bfd5-4f03-96de-5ab6333b179d/volumes" Nov 25 11:00:51 crc kubenswrapper[4821]: I1125 11:00:51.030858 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-57l9j"] Nov 25 11:00:51 crc kubenswrapper[4821]: I1125 11:00:51.040925 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-eaac-account-create-f4xs9"] Nov 25 11:00:51 crc kubenswrapper[4821]: I1125 11:00:51.048567 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-eaac-account-create-f4xs9"] Nov 25 11:00:51 crc kubenswrapper[4821]: I1125 11:00:51.054790 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-57l9j"] Nov 25 11:00:52 crc kubenswrapper[4821]: I1125 11:00:52.029379 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-5xnrf"] Nov 25 11:00:52 crc kubenswrapper[4821]: I1125 11:00:52.038547 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-30f9-account-create-szdhp"] Nov 25 11:00:52 crc kubenswrapper[4821]: I1125 11:00:52.046494 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-5xnrf"] Nov 25 11:00:52 crc kubenswrapper[4821]: I1125 11:00:52.053011 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-30f9-account-create-szdhp"] Nov 25 11:00:52 crc kubenswrapper[4821]: I1125 11:00:52.129296 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a0917a6-232d-43e0-8c9d-44eb9485d0d8" path="/var/lib/kubelet/pods/7a0917a6-232d-43e0-8c9d-44eb9485d0d8/volumes" Nov 25 11:00:52 crc kubenswrapper[4821]: I1125 11:00:52.130297 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af27979d-3739-453b-8c9a-1fe2c311fa17" path="/var/lib/kubelet/pods/af27979d-3739-453b-8c9a-1fe2c311fa17/volumes" Nov 25 11:00:52 crc kubenswrapper[4821]: I1125 11:00:52.131111 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2a4919e-4ad6-4cdc-b1cb-5a0200dcc8cc" path="/var/lib/kubelet/pods/b2a4919e-4ad6-4cdc-b1cb-5a0200dcc8cc/volumes" Nov 25 11:00:52 crc kubenswrapper[4821]: I1125 11:00:52.131920 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec0ad62e-84ad-44d8-a8f2-264d374ed70f" path="/var/lib/kubelet/pods/ec0ad62e-84ad-44d8-a8f2-264d374ed70f/volumes" Nov 25 11:00:54 crc kubenswrapper[4821]: I1125 11:00:54.114381 4821 scope.go:117] "RemoveContainer" containerID="fc16b2ea0cd36da4cef842d21755ec621134a4c21c5c964f8856fb8e6a102b2d" Nov 25 11:00:54 crc kubenswrapper[4821]: E1125 11:00:54.115121 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:01:00 crc kubenswrapper[4821]: I1125 11:01:00.160583 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29401141-j95bv"] Nov 25 11:01:00 crc kubenswrapper[4821]: E1125 11:01:00.162917 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d1e4062-40fb-4409-823a-3e1b5023dea0" containerName="collect-profiles" Nov 25 11:01:00 crc kubenswrapper[4821]: I1125 11:01:00.162941 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d1e4062-40fb-4409-823a-3e1b5023dea0" containerName="collect-profiles" Nov 25 11:01:00 crc kubenswrapper[4821]: I1125 11:01:00.163124 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d1e4062-40fb-4409-823a-3e1b5023dea0" containerName="collect-profiles" Nov 25 11:01:00 crc kubenswrapper[4821]: I1125 11:01:00.164019 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29401141-j95bv" Nov 25 11:01:00 crc kubenswrapper[4821]: I1125 11:01:00.177027 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29401141-j95bv"] Nov 25 11:01:00 crc kubenswrapper[4821]: I1125 11:01:00.328303 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f34ab84-5b6a-423d-b622-9b90c288695e-combined-ca-bundle\") pod \"keystone-cron-29401141-j95bv\" (UID: \"0f34ab84-5b6a-423d-b622-9b90c288695e\") " pod="openstack/keystone-cron-29401141-j95bv" Nov 25 11:01:00 crc kubenswrapper[4821]: I1125 11:01:00.328523 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f34ab84-5b6a-423d-b622-9b90c288695e-config-data\") pod \"keystone-cron-29401141-j95bv\" (UID: \"0f34ab84-5b6a-423d-b622-9b90c288695e\") " pod="openstack/keystone-cron-29401141-j95bv" Nov 25 11:01:00 crc kubenswrapper[4821]: I1125 11:01:00.328676 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2f7gg\" (UniqueName: \"kubernetes.io/projected/0f34ab84-5b6a-423d-b622-9b90c288695e-kube-api-access-2f7gg\") pod \"keystone-cron-29401141-j95bv\" (UID: \"0f34ab84-5b6a-423d-b622-9b90c288695e\") " pod="openstack/keystone-cron-29401141-j95bv" Nov 25 11:01:00 crc kubenswrapper[4821]: I1125 11:01:00.328830 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0f34ab84-5b6a-423d-b622-9b90c288695e-fernet-keys\") pod \"keystone-cron-29401141-j95bv\" (UID: \"0f34ab84-5b6a-423d-b622-9b90c288695e\") " pod="openstack/keystone-cron-29401141-j95bv" Nov 25 11:01:00 crc kubenswrapper[4821]: I1125 11:01:00.432723 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f34ab84-5b6a-423d-b622-9b90c288695e-combined-ca-bundle\") pod \"keystone-cron-29401141-j95bv\" (UID: \"0f34ab84-5b6a-423d-b622-9b90c288695e\") " pod="openstack/keystone-cron-29401141-j95bv" Nov 25 11:01:00 crc kubenswrapper[4821]: I1125 11:01:00.432830 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f34ab84-5b6a-423d-b622-9b90c288695e-config-data\") pod \"keystone-cron-29401141-j95bv\" (UID: \"0f34ab84-5b6a-423d-b622-9b90c288695e\") " pod="openstack/keystone-cron-29401141-j95bv" Nov 25 11:01:00 crc kubenswrapper[4821]: I1125 11:01:00.432892 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2f7gg\" (UniqueName: \"kubernetes.io/projected/0f34ab84-5b6a-423d-b622-9b90c288695e-kube-api-access-2f7gg\") pod \"keystone-cron-29401141-j95bv\" (UID: \"0f34ab84-5b6a-423d-b622-9b90c288695e\") " pod="openstack/keystone-cron-29401141-j95bv" Nov 25 11:01:00 crc kubenswrapper[4821]: I1125 11:01:00.432984 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0f34ab84-5b6a-423d-b622-9b90c288695e-fernet-keys\") pod \"keystone-cron-29401141-j95bv\" (UID: \"0f34ab84-5b6a-423d-b622-9b90c288695e\") " pod="openstack/keystone-cron-29401141-j95bv" Nov 25 11:01:00 crc kubenswrapper[4821]: I1125 11:01:00.442327 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0f34ab84-5b6a-423d-b622-9b90c288695e-fernet-keys\") pod \"keystone-cron-29401141-j95bv\" (UID: \"0f34ab84-5b6a-423d-b622-9b90c288695e\") " pod="openstack/keystone-cron-29401141-j95bv" Nov 25 11:01:00 crc kubenswrapper[4821]: I1125 11:01:00.453060 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f34ab84-5b6a-423d-b622-9b90c288695e-combined-ca-bundle\") pod \"keystone-cron-29401141-j95bv\" (UID: \"0f34ab84-5b6a-423d-b622-9b90c288695e\") " pod="openstack/keystone-cron-29401141-j95bv" Nov 25 11:01:00 crc kubenswrapper[4821]: I1125 11:01:00.453560 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f34ab84-5b6a-423d-b622-9b90c288695e-config-data\") pod \"keystone-cron-29401141-j95bv\" (UID: \"0f34ab84-5b6a-423d-b622-9b90c288695e\") " pod="openstack/keystone-cron-29401141-j95bv" Nov 25 11:01:00 crc kubenswrapper[4821]: I1125 11:01:00.459197 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2f7gg\" (UniqueName: \"kubernetes.io/projected/0f34ab84-5b6a-423d-b622-9b90c288695e-kube-api-access-2f7gg\") pod \"keystone-cron-29401141-j95bv\" (UID: \"0f34ab84-5b6a-423d-b622-9b90c288695e\") " pod="openstack/keystone-cron-29401141-j95bv" Nov 25 11:01:00 crc kubenswrapper[4821]: I1125 11:01:00.509089 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29401141-j95bv" Nov 25 11:01:00 crc kubenswrapper[4821]: I1125 11:01:00.976553 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29401141-j95bv"] Nov 25 11:01:01 crc kubenswrapper[4821]: I1125 11:01:01.054348 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29401141-j95bv" event={"ID":"0f34ab84-5b6a-423d-b622-9b90c288695e","Type":"ContainerStarted","Data":"c4d6886c91bf7f8ff5481a63cb449718257405f336e04aad7b1bf2b06500e490"} Nov 25 11:01:02 crc kubenswrapper[4821]: I1125 11:01:02.069803 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29401141-j95bv" event={"ID":"0f34ab84-5b6a-423d-b622-9b90c288695e","Type":"ContainerStarted","Data":"944f7089d3e97fab41fb51004c5889add82d1b8e568bc3a2449ea7f0b1fcfb1e"} Nov 25 11:01:02 crc kubenswrapper[4821]: I1125 11:01:02.094296 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29401141-j95bv" podStartSLOduration=2.094275453 podStartE2EDuration="2.094275453s" podCreationTimestamp="2025-11-25 11:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:01:02.090198554 +0000 UTC m=+1732.626518431" watchObservedRunningTime="2025-11-25 11:01:02.094275453 +0000 UTC m=+1732.630595300" Nov 25 11:01:04 crc kubenswrapper[4821]: I1125 11:01:04.092618 4821 generic.go:334] "Generic (PLEG): container finished" podID="0f34ab84-5b6a-423d-b622-9b90c288695e" containerID="944f7089d3e97fab41fb51004c5889add82d1b8e568bc3a2449ea7f0b1fcfb1e" exitCode=0 Nov 25 11:01:04 crc kubenswrapper[4821]: I1125 11:01:04.092715 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29401141-j95bv" event={"ID":"0f34ab84-5b6a-423d-b622-9b90c288695e","Type":"ContainerDied","Data":"944f7089d3e97fab41fb51004c5889add82d1b8e568bc3a2449ea7f0b1fcfb1e"} Nov 25 11:01:05 crc kubenswrapper[4821]: I1125 11:01:05.521567 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29401141-j95bv" Nov 25 11:01:05 crc kubenswrapper[4821]: I1125 11:01:05.641771 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2f7gg\" (UniqueName: \"kubernetes.io/projected/0f34ab84-5b6a-423d-b622-9b90c288695e-kube-api-access-2f7gg\") pod \"0f34ab84-5b6a-423d-b622-9b90c288695e\" (UID: \"0f34ab84-5b6a-423d-b622-9b90c288695e\") " Nov 25 11:01:05 crc kubenswrapper[4821]: I1125 11:01:05.642288 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0f34ab84-5b6a-423d-b622-9b90c288695e-fernet-keys\") pod \"0f34ab84-5b6a-423d-b622-9b90c288695e\" (UID: \"0f34ab84-5b6a-423d-b622-9b90c288695e\") " Nov 25 11:01:05 crc kubenswrapper[4821]: I1125 11:01:05.642316 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f34ab84-5b6a-423d-b622-9b90c288695e-combined-ca-bundle\") pod \"0f34ab84-5b6a-423d-b622-9b90c288695e\" (UID: \"0f34ab84-5b6a-423d-b622-9b90c288695e\") " Nov 25 11:01:05 crc kubenswrapper[4821]: I1125 11:01:05.642342 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f34ab84-5b6a-423d-b622-9b90c288695e-config-data\") pod \"0f34ab84-5b6a-423d-b622-9b90c288695e\" (UID: \"0f34ab84-5b6a-423d-b622-9b90c288695e\") " Nov 25 11:01:05 crc kubenswrapper[4821]: I1125 11:01:05.654385 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f34ab84-5b6a-423d-b622-9b90c288695e-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "0f34ab84-5b6a-423d-b622-9b90c288695e" (UID: "0f34ab84-5b6a-423d-b622-9b90c288695e"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:01:05 crc kubenswrapper[4821]: I1125 11:01:05.654541 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f34ab84-5b6a-423d-b622-9b90c288695e-kube-api-access-2f7gg" (OuterVolumeSpecName: "kube-api-access-2f7gg") pod "0f34ab84-5b6a-423d-b622-9b90c288695e" (UID: "0f34ab84-5b6a-423d-b622-9b90c288695e"). InnerVolumeSpecName "kube-api-access-2f7gg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:01:05 crc kubenswrapper[4821]: I1125 11:01:05.676050 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f34ab84-5b6a-423d-b622-9b90c288695e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0f34ab84-5b6a-423d-b622-9b90c288695e" (UID: "0f34ab84-5b6a-423d-b622-9b90c288695e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:01:05 crc kubenswrapper[4821]: I1125 11:01:05.696546 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f34ab84-5b6a-423d-b622-9b90c288695e-config-data" (OuterVolumeSpecName: "config-data") pod "0f34ab84-5b6a-423d-b622-9b90c288695e" (UID: "0f34ab84-5b6a-423d-b622-9b90c288695e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:01:05 crc kubenswrapper[4821]: I1125 11:01:05.745241 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2f7gg\" (UniqueName: \"kubernetes.io/projected/0f34ab84-5b6a-423d-b622-9b90c288695e-kube-api-access-2f7gg\") on node \"crc\" DevicePath \"\"" Nov 25 11:01:05 crc kubenswrapper[4821]: I1125 11:01:05.745277 4821 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0f34ab84-5b6a-423d-b622-9b90c288695e-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 25 11:01:05 crc kubenswrapper[4821]: I1125 11:01:05.745290 4821 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f34ab84-5b6a-423d-b622-9b90c288695e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:01:05 crc kubenswrapper[4821]: I1125 11:01:05.745300 4821 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f34ab84-5b6a-423d-b622-9b90c288695e-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 11:01:06 crc kubenswrapper[4821]: I1125 11:01:06.110441 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29401141-j95bv" event={"ID":"0f34ab84-5b6a-423d-b622-9b90c288695e","Type":"ContainerDied","Data":"c4d6886c91bf7f8ff5481a63cb449718257405f336e04aad7b1bf2b06500e490"} Nov 25 11:01:06 crc kubenswrapper[4821]: I1125 11:01:06.110714 4821 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c4d6886c91bf7f8ff5481a63cb449718257405f336e04aad7b1bf2b06500e490" Nov 25 11:01:06 crc kubenswrapper[4821]: I1125 11:01:06.110536 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29401141-j95bv" Nov 25 11:01:08 crc kubenswrapper[4821]: I1125 11:01:08.114152 4821 scope.go:117] "RemoveContainer" containerID="fc16b2ea0cd36da4cef842d21755ec621134a4c21c5c964f8856fb8e6a102b2d" Nov 25 11:01:08 crc kubenswrapper[4821]: E1125 11:01:08.114772 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:01:21 crc kubenswrapper[4821]: I1125 11:01:21.039370 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-42f5v"] Nov 25 11:01:21 crc kubenswrapper[4821]: I1125 11:01:21.047072 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-42f5v"] Nov 25 11:01:22 crc kubenswrapper[4821]: I1125 11:01:22.128608 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7745217e-45ba-403f-bf4f-1cbced054784" path="/var/lib/kubelet/pods/7745217e-45ba-403f-bf4f-1cbced054784/volumes" Nov 25 11:01:23 crc kubenswrapper[4821]: I1125 11:01:23.118566 4821 scope.go:117] "RemoveContainer" containerID="fc16b2ea0cd36da4cef842d21755ec621134a4c21c5c964f8856fb8e6a102b2d" Nov 25 11:01:23 crc kubenswrapper[4821]: E1125 11:01:23.119335 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:01:24 crc kubenswrapper[4821]: I1125 11:01:24.378722 4821 scope.go:117] "RemoveContainer" containerID="56412c7ea7e93aa6cbba9324eaff183a3a64bdbb1869f833b15d246ab3c9e9ae" Nov 25 11:01:24 crc kubenswrapper[4821]: I1125 11:01:24.408836 4821 scope.go:117] "RemoveContainer" containerID="df96d8a8aa5a3a8fcdbb3300533a9784a900a827714f334a3c3c4e608c75c433" Nov 25 11:01:24 crc kubenswrapper[4821]: I1125 11:01:24.456955 4821 scope.go:117] "RemoveContainer" containerID="3da8abe15577b506083dffcb7a9f7e50bafc6e525b9d574ed741d769adcfd66c" Nov 25 11:01:24 crc kubenswrapper[4821]: I1125 11:01:24.505357 4821 scope.go:117] "RemoveContainer" containerID="a40830e437f33f17d87382d4727c377ad8dc87a7a26dc8d568de4098842642d9" Nov 25 11:01:24 crc kubenswrapper[4821]: I1125 11:01:24.552852 4821 scope.go:117] "RemoveContainer" containerID="04c32b3179869b2701526757c21af3020ccc290479543c3c6f210843407d8ba1" Nov 25 11:01:24 crc kubenswrapper[4821]: I1125 11:01:24.591396 4821 scope.go:117] "RemoveContainer" containerID="05e371437d8c7a317b3cfd78207ff0ea04163ceb02785bce5e5f1497fbf42e9a" Nov 25 11:01:24 crc kubenswrapper[4821]: I1125 11:01:24.636349 4821 scope.go:117] "RemoveContainer" containerID="4148d293fc86b2ec5132a58cee28e829308cf0fbcd9ab995a4217bc85f1c0c50" Nov 25 11:01:32 crc kubenswrapper[4821]: I1125 11:01:32.357801 4821 generic.go:334] "Generic (PLEG): container finished" podID="5a15c7fc-4220-46ae-8af3-490c28f8b4ff" containerID="9c1876861e12a0eb2418f2fb1f64571f41412a1eb78e6fb65c29919b3f2f4449" exitCode=0 Nov 25 11:01:32 crc kubenswrapper[4821]: I1125 11:01:32.357897 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-sg8cc" event={"ID":"5a15c7fc-4220-46ae-8af3-490c28f8b4ff","Type":"ContainerDied","Data":"9c1876861e12a0eb2418f2fb1f64571f41412a1eb78e6fb65c29919b3f2f4449"} Nov 25 11:01:33 crc kubenswrapper[4821]: I1125 11:01:33.788216 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-sg8cc" Nov 25 11:01:33 crc kubenswrapper[4821]: I1125 11:01:33.891990 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5a15c7fc-4220-46ae-8af3-490c28f8b4ff-ssh-key\") pod \"5a15c7fc-4220-46ae-8af3-490c28f8b4ff\" (UID: \"5a15c7fc-4220-46ae-8af3-490c28f8b4ff\") " Nov 25 11:01:33 crc kubenswrapper[4821]: I1125 11:01:33.892053 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5a15c7fc-4220-46ae-8af3-490c28f8b4ff-inventory\") pod \"5a15c7fc-4220-46ae-8af3-490c28f8b4ff\" (UID: \"5a15c7fc-4220-46ae-8af3-490c28f8b4ff\") " Nov 25 11:01:33 crc kubenswrapper[4821]: I1125 11:01:33.892175 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h7jhk\" (UniqueName: \"kubernetes.io/projected/5a15c7fc-4220-46ae-8af3-490c28f8b4ff-kube-api-access-h7jhk\") pod \"5a15c7fc-4220-46ae-8af3-490c28f8b4ff\" (UID: \"5a15c7fc-4220-46ae-8af3-490c28f8b4ff\") " Nov 25 11:01:33 crc kubenswrapper[4821]: I1125 11:01:33.897418 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a15c7fc-4220-46ae-8af3-490c28f8b4ff-kube-api-access-h7jhk" (OuterVolumeSpecName: "kube-api-access-h7jhk") pod "5a15c7fc-4220-46ae-8af3-490c28f8b4ff" (UID: "5a15c7fc-4220-46ae-8af3-490c28f8b4ff"). InnerVolumeSpecName "kube-api-access-h7jhk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:01:33 crc kubenswrapper[4821]: I1125 11:01:33.919106 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a15c7fc-4220-46ae-8af3-490c28f8b4ff-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5a15c7fc-4220-46ae-8af3-490c28f8b4ff" (UID: "5a15c7fc-4220-46ae-8af3-490c28f8b4ff"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:01:33 crc kubenswrapper[4821]: I1125 11:01:33.925633 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a15c7fc-4220-46ae-8af3-490c28f8b4ff-inventory" (OuterVolumeSpecName: "inventory") pod "5a15c7fc-4220-46ae-8af3-490c28f8b4ff" (UID: "5a15c7fc-4220-46ae-8af3-490c28f8b4ff"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:01:33 crc kubenswrapper[4821]: I1125 11:01:33.994778 4821 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5a15c7fc-4220-46ae-8af3-490c28f8b4ff-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 11:01:33 crc kubenswrapper[4821]: I1125 11:01:33.994844 4821 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5a15c7fc-4220-46ae-8af3-490c28f8b4ff-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 11:01:33 crc kubenswrapper[4821]: I1125 11:01:33.994856 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h7jhk\" (UniqueName: \"kubernetes.io/projected/5a15c7fc-4220-46ae-8af3-490c28f8b4ff-kube-api-access-h7jhk\") on node \"crc\" DevicePath \"\"" Nov 25 11:01:34 crc kubenswrapper[4821]: I1125 11:01:34.382351 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-sg8cc" event={"ID":"5a15c7fc-4220-46ae-8af3-490c28f8b4ff","Type":"ContainerDied","Data":"c0660d7ce48a3a08a041caf378fbf5aa080a624da1cd1ef152b2d0d54f5c4cca"} Nov 25 11:01:34 crc kubenswrapper[4821]: I1125 11:01:34.382858 4821 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c0660d7ce48a3a08a041caf378fbf5aa080a624da1cd1ef152b2d0d54f5c4cca" Nov 25 11:01:34 crc kubenswrapper[4821]: I1125 11:01:34.382432 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-sg8cc" Nov 25 11:01:34 crc kubenswrapper[4821]: I1125 11:01:34.472126 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x5tw6"] Nov 25 11:01:34 crc kubenswrapper[4821]: E1125 11:01:34.472691 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f34ab84-5b6a-423d-b622-9b90c288695e" containerName="keystone-cron" Nov 25 11:01:34 crc kubenswrapper[4821]: I1125 11:01:34.472717 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f34ab84-5b6a-423d-b622-9b90c288695e" containerName="keystone-cron" Nov 25 11:01:34 crc kubenswrapper[4821]: E1125 11:01:34.472734 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a15c7fc-4220-46ae-8af3-490c28f8b4ff" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 25 11:01:34 crc kubenswrapper[4821]: I1125 11:01:34.472743 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a15c7fc-4220-46ae-8af3-490c28f8b4ff" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 25 11:01:34 crc kubenswrapper[4821]: I1125 11:01:34.472965 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a15c7fc-4220-46ae-8af3-490c28f8b4ff" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 25 11:01:34 crc kubenswrapper[4821]: I1125 11:01:34.472991 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f34ab84-5b6a-423d-b622-9b90c288695e" containerName="keystone-cron" Nov 25 11:01:34 crc kubenswrapper[4821]: I1125 11:01:34.473794 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x5tw6" Nov 25 11:01:34 crc kubenswrapper[4821]: I1125 11:01:34.485047 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x5tw6"] Nov 25 11:01:34 crc kubenswrapper[4821]: I1125 11:01:34.486101 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 11:01:34 crc kubenswrapper[4821]: I1125 11:01:34.486451 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 11:01:34 crc kubenswrapper[4821]: I1125 11:01:34.486567 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-z764x" Nov 25 11:01:34 crc kubenswrapper[4821]: I1125 11:01:34.487529 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 11:01:34 crc kubenswrapper[4821]: I1125 11:01:34.609772 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d77aa250-b203-4596-913c-7878a4d8b517-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-x5tw6\" (UID: \"d77aa250-b203-4596-913c-7878a4d8b517\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x5tw6" Nov 25 11:01:34 crc kubenswrapper[4821]: I1125 11:01:34.610921 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d77aa250-b203-4596-913c-7878a4d8b517-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-x5tw6\" (UID: \"d77aa250-b203-4596-913c-7878a4d8b517\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x5tw6" Nov 25 11:01:34 crc kubenswrapper[4821]: I1125 11:01:34.611153 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8k47\" (UniqueName: \"kubernetes.io/projected/d77aa250-b203-4596-913c-7878a4d8b517-kube-api-access-d8k47\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-x5tw6\" (UID: \"d77aa250-b203-4596-913c-7878a4d8b517\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x5tw6" Nov 25 11:01:34 crc kubenswrapper[4821]: I1125 11:01:34.712869 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d77aa250-b203-4596-913c-7878a4d8b517-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-x5tw6\" (UID: \"d77aa250-b203-4596-913c-7878a4d8b517\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x5tw6" Nov 25 11:01:34 crc kubenswrapper[4821]: I1125 11:01:34.712923 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8k47\" (UniqueName: \"kubernetes.io/projected/d77aa250-b203-4596-913c-7878a4d8b517-kube-api-access-d8k47\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-x5tw6\" (UID: \"d77aa250-b203-4596-913c-7878a4d8b517\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x5tw6" Nov 25 11:01:34 crc kubenswrapper[4821]: I1125 11:01:34.712979 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d77aa250-b203-4596-913c-7878a4d8b517-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-x5tw6\" (UID: \"d77aa250-b203-4596-913c-7878a4d8b517\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x5tw6" Nov 25 11:01:34 crc kubenswrapper[4821]: I1125 11:01:34.718043 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d77aa250-b203-4596-913c-7878a4d8b517-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-x5tw6\" (UID: \"d77aa250-b203-4596-913c-7878a4d8b517\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x5tw6" Nov 25 11:01:34 crc kubenswrapper[4821]: I1125 11:01:34.718084 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d77aa250-b203-4596-913c-7878a4d8b517-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-x5tw6\" (UID: \"d77aa250-b203-4596-913c-7878a4d8b517\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x5tw6" Nov 25 11:01:34 crc kubenswrapper[4821]: I1125 11:01:34.734935 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8k47\" (UniqueName: \"kubernetes.io/projected/d77aa250-b203-4596-913c-7878a4d8b517-kube-api-access-d8k47\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-x5tw6\" (UID: \"d77aa250-b203-4596-913c-7878a4d8b517\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x5tw6" Nov 25 11:01:34 crc kubenswrapper[4821]: I1125 11:01:34.796683 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x5tw6" Nov 25 11:01:35 crc kubenswrapper[4821]: I1125 11:01:35.114562 4821 scope.go:117] "RemoveContainer" containerID="fc16b2ea0cd36da4cef842d21755ec621134a4c21c5c964f8856fb8e6a102b2d" Nov 25 11:01:35 crc kubenswrapper[4821]: E1125 11:01:35.115185 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:01:35 crc kubenswrapper[4821]: I1125 11:01:35.409413 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x5tw6"] Nov 25 11:01:36 crc kubenswrapper[4821]: I1125 11:01:36.402040 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x5tw6" event={"ID":"d77aa250-b203-4596-913c-7878a4d8b517","Type":"ContainerStarted","Data":"b75b4ec2fc8bcac6abc5f3bf6b09edf8aeb7bfd33966f642e80b98362fa36712"} Nov 25 11:01:37 crc kubenswrapper[4821]: I1125 11:01:37.415952 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x5tw6" event={"ID":"d77aa250-b203-4596-913c-7878a4d8b517","Type":"ContainerStarted","Data":"7878d16d9de2a9034b858c94f088bf95872c21ee7c924c234ae276f0cba147d0"} Nov 25 11:01:37 crc kubenswrapper[4821]: I1125 11:01:37.445529 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x5tw6" podStartSLOduration=1.9454389810000001 podStartE2EDuration="3.445498132s" podCreationTimestamp="2025-11-25 11:01:34 +0000 UTC" firstStartedPulling="2025-11-25 11:01:35.430839915 +0000 UTC m=+1765.967159762" lastFinishedPulling="2025-11-25 11:01:36.930899066 +0000 UTC m=+1767.467218913" observedRunningTime="2025-11-25 11:01:37.435547161 +0000 UTC m=+1767.971867028" watchObservedRunningTime="2025-11-25 11:01:37.445498132 +0000 UTC m=+1767.981817979" Nov 25 11:01:42 crc kubenswrapper[4821]: I1125 11:01:42.460623 4821 generic.go:334] "Generic (PLEG): container finished" podID="d77aa250-b203-4596-913c-7878a4d8b517" containerID="7878d16d9de2a9034b858c94f088bf95872c21ee7c924c234ae276f0cba147d0" exitCode=0 Nov 25 11:01:42 crc kubenswrapper[4821]: I1125 11:01:42.460727 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x5tw6" event={"ID":"d77aa250-b203-4596-913c-7878a4d8b517","Type":"ContainerDied","Data":"7878d16d9de2a9034b858c94f088bf95872c21ee7c924c234ae276f0cba147d0"} Nov 25 11:01:43 crc kubenswrapper[4821]: I1125 11:01:43.896540 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x5tw6" Nov 25 11:01:44 crc kubenswrapper[4821]: I1125 11:01:44.015888 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d8k47\" (UniqueName: \"kubernetes.io/projected/d77aa250-b203-4596-913c-7878a4d8b517-kube-api-access-d8k47\") pod \"d77aa250-b203-4596-913c-7878a4d8b517\" (UID: \"d77aa250-b203-4596-913c-7878a4d8b517\") " Nov 25 11:01:44 crc kubenswrapper[4821]: I1125 11:01:44.015969 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d77aa250-b203-4596-913c-7878a4d8b517-inventory\") pod \"d77aa250-b203-4596-913c-7878a4d8b517\" (UID: \"d77aa250-b203-4596-913c-7878a4d8b517\") " Nov 25 11:01:44 crc kubenswrapper[4821]: I1125 11:01:44.016193 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d77aa250-b203-4596-913c-7878a4d8b517-ssh-key\") pod \"d77aa250-b203-4596-913c-7878a4d8b517\" (UID: \"d77aa250-b203-4596-913c-7878a4d8b517\") " Nov 25 11:01:44 crc kubenswrapper[4821]: I1125 11:01:44.022018 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d77aa250-b203-4596-913c-7878a4d8b517-kube-api-access-d8k47" (OuterVolumeSpecName: "kube-api-access-d8k47") pod "d77aa250-b203-4596-913c-7878a4d8b517" (UID: "d77aa250-b203-4596-913c-7878a4d8b517"). InnerVolumeSpecName "kube-api-access-d8k47". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:01:44 crc kubenswrapper[4821]: I1125 11:01:44.045354 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d77aa250-b203-4596-913c-7878a4d8b517-inventory" (OuterVolumeSpecName: "inventory") pod "d77aa250-b203-4596-913c-7878a4d8b517" (UID: "d77aa250-b203-4596-913c-7878a4d8b517"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:01:44 crc kubenswrapper[4821]: I1125 11:01:44.046109 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d77aa250-b203-4596-913c-7878a4d8b517-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d77aa250-b203-4596-913c-7878a4d8b517" (UID: "d77aa250-b203-4596-913c-7878a4d8b517"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:01:44 crc kubenswrapper[4821]: I1125 11:01:44.119110 4821 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d77aa250-b203-4596-913c-7878a4d8b517-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 11:01:44 crc kubenswrapper[4821]: I1125 11:01:44.119150 4821 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d77aa250-b203-4596-913c-7878a4d8b517-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 11:01:44 crc kubenswrapper[4821]: I1125 11:01:44.119174 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d8k47\" (UniqueName: \"kubernetes.io/projected/d77aa250-b203-4596-913c-7878a4d8b517-kube-api-access-d8k47\") on node \"crc\" DevicePath \"\"" Nov 25 11:01:44 crc kubenswrapper[4821]: I1125 11:01:44.482712 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x5tw6" event={"ID":"d77aa250-b203-4596-913c-7878a4d8b517","Type":"ContainerDied","Data":"b75b4ec2fc8bcac6abc5f3bf6b09edf8aeb7bfd33966f642e80b98362fa36712"} Nov 25 11:01:44 crc kubenswrapper[4821]: I1125 11:01:44.482770 4821 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b75b4ec2fc8bcac6abc5f3bf6b09edf8aeb7bfd33966f642e80b98362fa36712" Nov 25 11:01:44 crc kubenswrapper[4821]: I1125 11:01:44.482871 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x5tw6" Nov 25 11:01:44 crc kubenswrapper[4821]: I1125 11:01:44.553202 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-x5qj5"] Nov 25 11:01:44 crc kubenswrapper[4821]: E1125 11:01:44.553683 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d77aa250-b203-4596-913c-7878a4d8b517" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 25 11:01:44 crc kubenswrapper[4821]: I1125 11:01:44.553709 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="d77aa250-b203-4596-913c-7878a4d8b517" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 25 11:01:44 crc kubenswrapper[4821]: I1125 11:01:44.553899 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="d77aa250-b203-4596-913c-7878a4d8b517" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 25 11:01:44 crc kubenswrapper[4821]: I1125 11:01:44.554659 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-x5qj5" Nov 25 11:01:44 crc kubenswrapper[4821]: I1125 11:01:44.556975 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 11:01:44 crc kubenswrapper[4821]: I1125 11:01:44.557574 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-z764x" Nov 25 11:01:44 crc kubenswrapper[4821]: I1125 11:01:44.558158 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 11:01:44 crc kubenswrapper[4821]: I1125 11:01:44.559397 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 11:01:44 crc kubenswrapper[4821]: I1125 11:01:44.562746 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-x5qj5"] Nov 25 11:01:44 crc kubenswrapper[4821]: I1125 11:01:44.628066 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2kbm4\" (UniqueName: \"kubernetes.io/projected/9418fa73-031c-4c0a-bfbd-ffb206714075-kube-api-access-2kbm4\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-x5qj5\" (UID: \"9418fa73-031c-4c0a-bfbd-ffb206714075\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-x5qj5" Nov 25 11:01:44 crc kubenswrapper[4821]: I1125 11:01:44.628312 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9418fa73-031c-4c0a-bfbd-ffb206714075-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-x5qj5\" (UID: \"9418fa73-031c-4c0a-bfbd-ffb206714075\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-x5qj5" Nov 25 11:01:44 crc kubenswrapper[4821]: I1125 11:01:44.628377 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9418fa73-031c-4c0a-bfbd-ffb206714075-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-x5qj5\" (UID: \"9418fa73-031c-4c0a-bfbd-ffb206714075\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-x5qj5" Nov 25 11:01:44 crc kubenswrapper[4821]: I1125 11:01:44.730308 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9418fa73-031c-4c0a-bfbd-ffb206714075-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-x5qj5\" (UID: \"9418fa73-031c-4c0a-bfbd-ffb206714075\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-x5qj5" Nov 25 11:01:44 crc kubenswrapper[4821]: I1125 11:01:44.730376 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9418fa73-031c-4c0a-bfbd-ffb206714075-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-x5qj5\" (UID: \"9418fa73-031c-4c0a-bfbd-ffb206714075\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-x5qj5" Nov 25 11:01:44 crc kubenswrapper[4821]: I1125 11:01:44.730469 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2kbm4\" (UniqueName: \"kubernetes.io/projected/9418fa73-031c-4c0a-bfbd-ffb206714075-kube-api-access-2kbm4\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-x5qj5\" (UID: \"9418fa73-031c-4c0a-bfbd-ffb206714075\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-x5qj5" Nov 25 11:01:44 crc kubenswrapper[4821]: I1125 11:01:44.735403 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9418fa73-031c-4c0a-bfbd-ffb206714075-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-x5qj5\" (UID: \"9418fa73-031c-4c0a-bfbd-ffb206714075\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-x5qj5" Nov 25 11:01:44 crc kubenswrapper[4821]: I1125 11:01:44.735772 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9418fa73-031c-4c0a-bfbd-ffb206714075-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-x5qj5\" (UID: \"9418fa73-031c-4c0a-bfbd-ffb206714075\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-x5qj5" Nov 25 11:01:44 crc kubenswrapper[4821]: I1125 11:01:44.748960 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2kbm4\" (UniqueName: \"kubernetes.io/projected/9418fa73-031c-4c0a-bfbd-ffb206714075-kube-api-access-2kbm4\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-x5qj5\" (UID: \"9418fa73-031c-4c0a-bfbd-ffb206714075\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-x5qj5" Nov 25 11:01:44 crc kubenswrapper[4821]: I1125 11:01:44.870491 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-x5qj5" Nov 25 11:01:45 crc kubenswrapper[4821]: I1125 11:01:45.395870 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-x5qj5"] Nov 25 11:01:45 crc kubenswrapper[4821]: W1125 11:01:45.402895 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9418fa73_031c_4c0a_bfbd_ffb206714075.slice/crio-bf966e7debdebaf82c69e0c0028317ecae1814887f41183bfd8dc8b74bc8b99a WatchSource:0}: Error finding container bf966e7debdebaf82c69e0c0028317ecae1814887f41183bfd8dc8b74bc8b99a: Status 404 returned error can't find the container with id bf966e7debdebaf82c69e0c0028317ecae1814887f41183bfd8dc8b74bc8b99a Nov 25 11:01:45 crc kubenswrapper[4821]: I1125 11:01:45.492051 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-x5qj5" event={"ID":"9418fa73-031c-4c0a-bfbd-ffb206714075","Type":"ContainerStarted","Data":"bf966e7debdebaf82c69e0c0028317ecae1814887f41183bfd8dc8b74bc8b99a"} Nov 25 11:01:47 crc kubenswrapper[4821]: I1125 11:01:47.115898 4821 scope.go:117] "RemoveContainer" containerID="fc16b2ea0cd36da4cef842d21755ec621134a4c21c5c964f8856fb8e6a102b2d" Nov 25 11:01:47 crc kubenswrapper[4821]: E1125 11:01:47.117442 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:01:47 crc kubenswrapper[4821]: I1125 11:01:47.507807 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-x5qj5" event={"ID":"9418fa73-031c-4c0a-bfbd-ffb206714075","Type":"ContainerStarted","Data":"ee9d5bb4888fc654c5f5d506db9f659d75f553e58e4d6c5cbe7ea2cb35e91de7"} Nov 25 11:01:47 crc kubenswrapper[4821]: I1125 11:01:47.530055 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-x5qj5" podStartSLOduration=2.7246246530000002 podStartE2EDuration="3.530012697s" podCreationTimestamp="2025-11-25 11:01:44 +0000 UTC" firstStartedPulling="2025-11-25 11:01:45.406903818 +0000 UTC m=+1775.943223665" lastFinishedPulling="2025-11-25 11:01:46.212291842 +0000 UTC m=+1776.748611709" observedRunningTime="2025-11-25 11:01:47.527605898 +0000 UTC m=+1778.063925755" watchObservedRunningTime="2025-11-25 11:01:47.530012697 +0000 UTC m=+1778.066332584" Nov 25 11:01:50 crc kubenswrapper[4821]: I1125 11:01:50.042363 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-qwwdq"] Nov 25 11:01:50 crc kubenswrapper[4821]: I1125 11:01:50.050395 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-qwwdq"] Nov 25 11:01:50 crc kubenswrapper[4821]: I1125 11:01:50.144613 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06c32eab-f3e4-4ebb-bb0e-0a887c0da0de" path="/var/lib/kubelet/pods/06c32eab-f3e4-4ebb-bb0e-0a887c0da0de/volumes" Nov 25 11:01:58 crc kubenswrapper[4821]: I1125 11:01:58.114519 4821 scope.go:117] "RemoveContainer" containerID="fc16b2ea0cd36da4cef842d21755ec621134a4c21c5c964f8856fb8e6a102b2d" Nov 25 11:01:58 crc kubenswrapper[4821]: E1125 11:01:58.115732 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:01:59 crc kubenswrapper[4821]: I1125 11:01:59.034721 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-vwzcd"] Nov 25 11:01:59 crc kubenswrapper[4821]: I1125 11:01:59.041818 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-vwzcd"] Nov 25 11:02:00 crc kubenswrapper[4821]: I1125 11:02:00.126247 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e22e60b5-5698-493b-9a3f-8cfdf289c6d9" path="/var/lib/kubelet/pods/e22e60b5-5698-493b-9a3f-8cfdf289c6d9/volumes" Nov 25 11:02:10 crc kubenswrapper[4821]: I1125 11:02:10.121085 4821 scope.go:117] "RemoveContainer" containerID="fc16b2ea0cd36da4cef842d21755ec621134a4c21c5c964f8856fb8e6a102b2d" Nov 25 11:02:10 crc kubenswrapper[4821]: E1125 11:02:10.122411 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:02:22 crc kubenswrapper[4821]: I1125 11:02:22.114772 4821 scope.go:117] "RemoveContainer" containerID="fc16b2ea0cd36da4cef842d21755ec621134a4c21c5c964f8856fb8e6a102b2d" Nov 25 11:02:22 crc kubenswrapper[4821]: E1125 11:02:22.115564 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:02:24 crc kubenswrapper[4821]: I1125 11:02:24.823086 4821 scope.go:117] "RemoveContainer" containerID="0c2fa815c3e17265e814630a7030173763f041b2d9166a413539261379e799be" Nov 25 11:02:24 crc kubenswrapper[4821]: I1125 11:02:24.877361 4821 scope.go:117] "RemoveContainer" containerID="7cf37449c9b799d7e943c3e4748fcbac868542b3396da366d9c1cfbfb5ba0a26" Nov 25 11:02:27 crc kubenswrapper[4821]: I1125 11:02:27.892224 4821 generic.go:334] "Generic (PLEG): container finished" podID="9418fa73-031c-4c0a-bfbd-ffb206714075" containerID="ee9d5bb4888fc654c5f5d506db9f659d75f553e58e4d6c5cbe7ea2cb35e91de7" exitCode=0 Nov 25 11:02:27 crc kubenswrapper[4821]: I1125 11:02:27.892340 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-x5qj5" event={"ID":"9418fa73-031c-4c0a-bfbd-ffb206714075","Type":"ContainerDied","Data":"ee9d5bb4888fc654c5f5d506db9f659d75f553e58e4d6c5cbe7ea2cb35e91de7"} Nov 25 11:02:29 crc kubenswrapper[4821]: I1125 11:02:29.309279 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-x5qj5" Nov 25 11:02:29 crc kubenswrapper[4821]: I1125 11:02:29.350356 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2kbm4\" (UniqueName: \"kubernetes.io/projected/9418fa73-031c-4c0a-bfbd-ffb206714075-kube-api-access-2kbm4\") pod \"9418fa73-031c-4c0a-bfbd-ffb206714075\" (UID: \"9418fa73-031c-4c0a-bfbd-ffb206714075\") " Nov 25 11:02:29 crc kubenswrapper[4821]: I1125 11:02:29.350484 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9418fa73-031c-4c0a-bfbd-ffb206714075-inventory\") pod \"9418fa73-031c-4c0a-bfbd-ffb206714075\" (UID: \"9418fa73-031c-4c0a-bfbd-ffb206714075\") " Nov 25 11:02:29 crc kubenswrapper[4821]: I1125 11:02:29.350558 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9418fa73-031c-4c0a-bfbd-ffb206714075-ssh-key\") pod \"9418fa73-031c-4c0a-bfbd-ffb206714075\" (UID: \"9418fa73-031c-4c0a-bfbd-ffb206714075\") " Nov 25 11:02:29 crc kubenswrapper[4821]: I1125 11:02:29.360036 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9418fa73-031c-4c0a-bfbd-ffb206714075-kube-api-access-2kbm4" (OuterVolumeSpecName: "kube-api-access-2kbm4") pod "9418fa73-031c-4c0a-bfbd-ffb206714075" (UID: "9418fa73-031c-4c0a-bfbd-ffb206714075"). InnerVolumeSpecName "kube-api-access-2kbm4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:02:29 crc kubenswrapper[4821]: I1125 11:02:29.377358 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9418fa73-031c-4c0a-bfbd-ffb206714075-inventory" (OuterVolumeSpecName: "inventory") pod "9418fa73-031c-4c0a-bfbd-ffb206714075" (UID: "9418fa73-031c-4c0a-bfbd-ffb206714075"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:02:29 crc kubenswrapper[4821]: I1125 11:02:29.378130 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9418fa73-031c-4c0a-bfbd-ffb206714075-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9418fa73-031c-4c0a-bfbd-ffb206714075" (UID: "9418fa73-031c-4c0a-bfbd-ffb206714075"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:02:29 crc kubenswrapper[4821]: I1125 11:02:29.451911 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2kbm4\" (UniqueName: \"kubernetes.io/projected/9418fa73-031c-4c0a-bfbd-ffb206714075-kube-api-access-2kbm4\") on node \"crc\" DevicePath \"\"" Nov 25 11:02:29 crc kubenswrapper[4821]: I1125 11:02:29.451940 4821 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9418fa73-031c-4c0a-bfbd-ffb206714075-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 11:02:29 crc kubenswrapper[4821]: I1125 11:02:29.451949 4821 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9418fa73-031c-4c0a-bfbd-ffb206714075-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 11:02:29 crc kubenswrapper[4821]: I1125 11:02:29.911835 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-x5qj5" event={"ID":"9418fa73-031c-4c0a-bfbd-ffb206714075","Type":"ContainerDied","Data":"bf966e7debdebaf82c69e0c0028317ecae1814887f41183bfd8dc8b74bc8b99a"} Nov 25 11:02:29 crc kubenswrapper[4821]: I1125 11:02:29.911909 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-x5qj5" Nov 25 11:02:29 crc kubenswrapper[4821]: I1125 11:02:29.911929 4821 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bf966e7debdebaf82c69e0c0028317ecae1814887f41183bfd8dc8b74bc8b99a" Nov 25 11:02:30 crc kubenswrapper[4821]: I1125 11:02:30.019744 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-d8lr8"] Nov 25 11:02:30 crc kubenswrapper[4821]: E1125 11:02:30.020313 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9418fa73-031c-4c0a-bfbd-ffb206714075" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 25 11:02:30 crc kubenswrapper[4821]: I1125 11:02:30.020346 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="9418fa73-031c-4c0a-bfbd-ffb206714075" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 25 11:02:30 crc kubenswrapper[4821]: I1125 11:02:30.020719 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="9418fa73-031c-4c0a-bfbd-ffb206714075" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 25 11:02:30 crc kubenswrapper[4821]: I1125 11:02:30.021733 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-d8lr8" Nov 25 11:02:30 crc kubenswrapper[4821]: I1125 11:02:30.030596 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-d8lr8"] Nov 25 11:02:30 crc kubenswrapper[4821]: I1125 11:02:30.032114 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 11:02:30 crc kubenswrapper[4821]: I1125 11:02:30.032393 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 11:02:30 crc kubenswrapper[4821]: I1125 11:02:30.032517 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 11:02:30 crc kubenswrapper[4821]: I1125 11:02:30.032716 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-z764x" Nov 25 11:02:30 crc kubenswrapper[4821]: I1125 11:02:30.166038 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c20e6196-b767-4fbb-9d59-a88080a3aa11-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-d8lr8\" (UID: \"c20e6196-b767-4fbb-9d59-a88080a3aa11\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-d8lr8" Nov 25 11:02:30 crc kubenswrapper[4821]: I1125 11:02:30.166114 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c20e6196-b767-4fbb-9d59-a88080a3aa11-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-d8lr8\" (UID: \"c20e6196-b767-4fbb-9d59-a88080a3aa11\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-d8lr8" Nov 25 11:02:30 crc kubenswrapper[4821]: I1125 11:02:30.166371 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cstgs\" (UniqueName: \"kubernetes.io/projected/c20e6196-b767-4fbb-9d59-a88080a3aa11-kube-api-access-cstgs\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-d8lr8\" (UID: \"c20e6196-b767-4fbb-9d59-a88080a3aa11\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-d8lr8" Nov 25 11:02:30 crc kubenswrapper[4821]: I1125 11:02:30.267628 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c20e6196-b767-4fbb-9d59-a88080a3aa11-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-d8lr8\" (UID: \"c20e6196-b767-4fbb-9d59-a88080a3aa11\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-d8lr8" Nov 25 11:02:30 crc kubenswrapper[4821]: I1125 11:02:30.267738 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cstgs\" (UniqueName: \"kubernetes.io/projected/c20e6196-b767-4fbb-9d59-a88080a3aa11-kube-api-access-cstgs\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-d8lr8\" (UID: \"c20e6196-b767-4fbb-9d59-a88080a3aa11\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-d8lr8" Nov 25 11:02:30 crc kubenswrapper[4821]: I1125 11:02:30.267859 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c20e6196-b767-4fbb-9d59-a88080a3aa11-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-d8lr8\" (UID: \"c20e6196-b767-4fbb-9d59-a88080a3aa11\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-d8lr8" Nov 25 11:02:30 crc kubenswrapper[4821]: I1125 11:02:30.271958 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c20e6196-b767-4fbb-9d59-a88080a3aa11-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-d8lr8\" (UID: \"c20e6196-b767-4fbb-9d59-a88080a3aa11\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-d8lr8" Nov 25 11:02:30 crc kubenswrapper[4821]: I1125 11:02:30.274112 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c20e6196-b767-4fbb-9d59-a88080a3aa11-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-d8lr8\" (UID: \"c20e6196-b767-4fbb-9d59-a88080a3aa11\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-d8lr8" Nov 25 11:02:30 crc kubenswrapper[4821]: I1125 11:02:30.287088 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cstgs\" (UniqueName: \"kubernetes.io/projected/c20e6196-b767-4fbb-9d59-a88080a3aa11-kube-api-access-cstgs\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-d8lr8\" (UID: \"c20e6196-b767-4fbb-9d59-a88080a3aa11\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-d8lr8" Nov 25 11:02:30 crc kubenswrapper[4821]: I1125 11:02:30.345528 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-d8lr8" Nov 25 11:02:30 crc kubenswrapper[4821]: I1125 11:02:30.872883 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-d8lr8"] Nov 25 11:02:30 crc kubenswrapper[4821]: I1125 11:02:30.922950 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-d8lr8" event={"ID":"c20e6196-b767-4fbb-9d59-a88080a3aa11","Type":"ContainerStarted","Data":"e4e878551f908d7dac6d0e7ab5b25d75cbbc8922259fc98420bf82a5f5b41b09"} Nov 25 11:02:32 crc kubenswrapper[4821]: I1125 11:02:32.938146 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-d8lr8" event={"ID":"c20e6196-b767-4fbb-9d59-a88080a3aa11","Type":"ContainerStarted","Data":"7051502d6644af9dcb49c113d1611a3af0a57dc22f79509de1ade351963df9d6"} Nov 25 11:02:32 crc kubenswrapper[4821]: I1125 11:02:32.966840 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-d8lr8" podStartSLOduration=3.074101357 podStartE2EDuration="3.966805834s" podCreationTimestamp="2025-11-25 11:02:29 +0000 UTC" firstStartedPulling="2025-11-25 11:02:30.884106821 +0000 UTC m=+1821.420426668" lastFinishedPulling="2025-11-25 11:02:31.776811308 +0000 UTC m=+1822.313131145" observedRunningTime="2025-11-25 11:02:32.951789173 +0000 UTC m=+1823.488109040" watchObservedRunningTime="2025-11-25 11:02:32.966805834 +0000 UTC m=+1823.503125711" Nov 25 11:02:35 crc kubenswrapper[4821]: I1125 11:02:35.052477 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-p42p6"] Nov 25 11:02:35 crc kubenswrapper[4821]: I1125 11:02:35.062607 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-p42p6"] Nov 25 11:02:36 crc kubenswrapper[4821]: I1125 11:02:36.126829 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d463843-fed0-4215-a85c-aac27aa1de50" path="/var/lib/kubelet/pods/0d463843-fed0-4215-a85c-aac27aa1de50/volumes" Nov 25 11:02:37 crc kubenswrapper[4821]: I1125 11:02:37.114627 4821 scope.go:117] "RemoveContainer" containerID="fc16b2ea0cd36da4cef842d21755ec621134a4c21c5c964f8856fb8e6a102b2d" Nov 25 11:02:37 crc kubenswrapper[4821]: E1125 11:02:37.115140 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:02:50 crc kubenswrapper[4821]: I1125 11:02:50.121491 4821 scope.go:117] "RemoveContainer" containerID="fc16b2ea0cd36da4cef842d21755ec621134a4c21c5c964f8856fb8e6a102b2d" Nov 25 11:02:50 crc kubenswrapper[4821]: E1125 11:02:50.122417 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:03:03 crc kubenswrapper[4821]: I1125 11:03:03.115722 4821 scope.go:117] "RemoveContainer" containerID="fc16b2ea0cd36da4cef842d21755ec621134a4c21c5c964f8856fb8e6a102b2d" Nov 25 11:03:03 crc kubenswrapper[4821]: E1125 11:03:03.116783 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:03:16 crc kubenswrapper[4821]: I1125 11:03:16.114930 4821 scope.go:117] "RemoveContainer" containerID="fc16b2ea0cd36da4cef842d21755ec621134a4c21c5c964f8856fb8e6a102b2d" Nov 25 11:03:16 crc kubenswrapper[4821]: E1125 11:03:16.115621 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:03:22 crc kubenswrapper[4821]: I1125 11:03:22.348073 4821 generic.go:334] "Generic (PLEG): container finished" podID="c20e6196-b767-4fbb-9d59-a88080a3aa11" containerID="7051502d6644af9dcb49c113d1611a3af0a57dc22f79509de1ade351963df9d6" exitCode=0 Nov 25 11:03:22 crc kubenswrapper[4821]: I1125 11:03:22.348214 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-d8lr8" event={"ID":"c20e6196-b767-4fbb-9d59-a88080a3aa11","Type":"ContainerDied","Data":"7051502d6644af9dcb49c113d1611a3af0a57dc22f79509de1ade351963df9d6"} Nov 25 11:03:23 crc kubenswrapper[4821]: I1125 11:03:23.699348 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-d8lr8" Nov 25 11:03:23 crc kubenswrapper[4821]: I1125 11:03:23.816473 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cstgs\" (UniqueName: \"kubernetes.io/projected/c20e6196-b767-4fbb-9d59-a88080a3aa11-kube-api-access-cstgs\") pod \"c20e6196-b767-4fbb-9d59-a88080a3aa11\" (UID: \"c20e6196-b767-4fbb-9d59-a88080a3aa11\") " Nov 25 11:03:23 crc kubenswrapper[4821]: I1125 11:03:23.816689 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c20e6196-b767-4fbb-9d59-a88080a3aa11-ssh-key\") pod \"c20e6196-b767-4fbb-9d59-a88080a3aa11\" (UID: \"c20e6196-b767-4fbb-9d59-a88080a3aa11\") " Nov 25 11:03:23 crc kubenswrapper[4821]: I1125 11:03:23.816719 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c20e6196-b767-4fbb-9d59-a88080a3aa11-inventory\") pod \"c20e6196-b767-4fbb-9d59-a88080a3aa11\" (UID: \"c20e6196-b767-4fbb-9d59-a88080a3aa11\") " Nov 25 11:03:23 crc kubenswrapper[4821]: I1125 11:03:23.822565 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c20e6196-b767-4fbb-9d59-a88080a3aa11-kube-api-access-cstgs" (OuterVolumeSpecName: "kube-api-access-cstgs") pod "c20e6196-b767-4fbb-9d59-a88080a3aa11" (UID: "c20e6196-b767-4fbb-9d59-a88080a3aa11"). InnerVolumeSpecName "kube-api-access-cstgs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:03:23 crc kubenswrapper[4821]: I1125 11:03:23.847472 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c20e6196-b767-4fbb-9d59-a88080a3aa11-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c20e6196-b767-4fbb-9d59-a88080a3aa11" (UID: "c20e6196-b767-4fbb-9d59-a88080a3aa11"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:03:23 crc kubenswrapper[4821]: I1125 11:03:23.850635 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c20e6196-b767-4fbb-9d59-a88080a3aa11-inventory" (OuterVolumeSpecName: "inventory") pod "c20e6196-b767-4fbb-9d59-a88080a3aa11" (UID: "c20e6196-b767-4fbb-9d59-a88080a3aa11"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:03:23 crc kubenswrapper[4821]: I1125 11:03:23.919364 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cstgs\" (UniqueName: \"kubernetes.io/projected/c20e6196-b767-4fbb-9d59-a88080a3aa11-kube-api-access-cstgs\") on node \"crc\" DevicePath \"\"" Nov 25 11:03:23 crc kubenswrapper[4821]: I1125 11:03:23.919401 4821 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c20e6196-b767-4fbb-9d59-a88080a3aa11-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 11:03:23 crc kubenswrapper[4821]: I1125 11:03:23.919418 4821 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c20e6196-b767-4fbb-9d59-a88080a3aa11-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 11:03:24 crc kubenswrapper[4821]: I1125 11:03:24.365776 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-d8lr8" event={"ID":"c20e6196-b767-4fbb-9d59-a88080a3aa11","Type":"ContainerDied","Data":"e4e878551f908d7dac6d0e7ab5b25d75cbbc8922259fc98420bf82a5f5b41b09"} Nov 25 11:03:24 crc kubenswrapper[4821]: I1125 11:03:24.365820 4821 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e4e878551f908d7dac6d0e7ab5b25d75cbbc8922259fc98420bf82a5f5b41b09" Nov 25 11:03:24 crc kubenswrapper[4821]: I1125 11:03:24.365856 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-d8lr8" Nov 25 11:03:24 crc kubenswrapper[4821]: I1125 11:03:24.442492 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-dxdpz"] Nov 25 11:03:24 crc kubenswrapper[4821]: E1125 11:03:24.442868 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c20e6196-b767-4fbb-9d59-a88080a3aa11" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 25 11:03:24 crc kubenswrapper[4821]: I1125 11:03:24.442888 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="c20e6196-b767-4fbb-9d59-a88080a3aa11" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 25 11:03:24 crc kubenswrapper[4821]: I1125 11:03:24.443094 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="c20e6196-b767-4fbb-9d59-a88080a3aa11" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 25 11:03:24 crc kubenswrapper[4821]: I1125 11:03:24.443820 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-dxdpz" Nov 25 11:03:24 crc kubenswrapper[4821]: I1125 11:03:24.446100 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-z764x" Nov 25 11:03:24 crc kubenswrapper[4821]: I1125 11:03:24.446101 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 11:03:24 crc kubenswrapper[4821]: I1125 11:03:24.448198 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 11:03:24 crc kubenswrapper[4821]: I1125 11:03:24.450936 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 11:03:24 crc kubenswrapper[4821]: I1125 11:03:24.454547 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-dxdpz"] Nov 25 11:03:24 crc kubenswrapper[4821]: I1125 11:03:24.631514 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/e1247bbd-54fc-45ce-8a90-0b8a5df02c47-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-dxdpz\" (UID: \"e1247bbd-54fc-45ce-8a90-0b8a5df02c47\") " pod="openstack/ssh-known-hosts-edpm-deployment-dxdpz" Nov 25 11:03:24 crc kubenswrapper[4821]: I1125 11:03:24.631600 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e1247bbd-54fc-45ce-8a90-0b8a5df02c47-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-dxdpz\" (UID: \"e1247bbd-54fc-45ce-8a90-0b8a5df02c47\") " pod="openstack/ssh-known-hosts-edpm-deployment-dxdpz" Nov 25 11:03:24 crc kubenswrapper[4821]: I1125 11:03:24.631644 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8sbl\" (UniqueName: \"kubernetes.io/projected/e1247bbd-54fc-45ce-8a90-0b8a5df02c47-kube-api-access-n8sbl\") pod \"ssh-known-hosts-edpm-deployment-dxdpz\" (UID: \"e1247bbd-54fc-45ce-8a90-0b8a5df02c47\") " pod="openstack/ssh-known-hosts-edpm-deployment-dxdpz" Nov 25 11:03:24 crc kubenswrapper[4821]: I1125 11:03:24.733822 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/e1247bbd-54fc-45ce-8a90-0b8a5df02c47-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-dxdpz\" (UID: \"e1247bbd-54fc-45ce-8a90-0b8a5df02c47\") " pod="openstack/ssh-known-hosts-edpm-deployment-dxdpz" Nov 25 11:03:24 crc kubenswrapper[4821]: I1125 11:03:24.733885 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e1247bbd-54fc-45ce-8a90-0b8a5df02c47-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-dxdpz\" (UID: \"e1247bbd-54fc-45ce-8a90-0b8a5df02c47\") " pod="openstack/ssh-known-hosts-edpm-deployment-dxdpz" Nov 25 11:03:24 crc kubenswrapper[4821]: I1125 11:03:24.733924 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8sbl\" (UniqueName: \"kubernetes.io/projected/e1247bbd-54fc-45ce-8a90-0b8a5df02c47-kube-api-access-n8sbl\") pod \"ssh-known-hosts-edpm-deployment-dxdpz\" (UID: \"e1247bbd-54fc-45ce-8a90-0b8a5df02c47\") " pod="openstack/ssh-known-hosts-edpm-deployment-dxdpz" Nov 25 11:03:24 crc kubenswrapper[4821]: I1125 11:03:24.738274 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e1247bbd-54fc-45ce-8a90-0b8a5df02c47-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-dxdpz\" (UID: \"e1247bbd-54fc-45ce-8a90-0b8a5df02c47\") " pod="openstack/ssh-known-hosts-edpm-deployment-dxdpz" Nov 25 11:03:24 crc kubenswrapper[4821]: I1125 11:03:24.738849 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/e1247bbd-54fc-45ce-8a90-0b8a5df02c47-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-dxdpz\" (UID: \"e1247bbd-54fc-45ce-8a90-0b8a5df02c47\") " pod="openstack/ssh-known-hosts-edpm-deployment-dxdpz" Nov 25 11:03:24 crc kubenswrapper[4821]: I1125 11:03:24.749360 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8sbl\" (UniqueName: \"kubernetes.io/projected/e1247bbd-54fc-45ce-8a90-0b8a5df02c47-kube-api-access-n8sbl\") pod \"ssh-known-hosts-edpm-deployment-dxdpz\" (UID: \"e1247bbd-54fc-45ce-8a90-0b8a5df02c47\") " pod="openstack/ssh-known-hosts-edpm-deployment-dxdpz" Nov 25 11:03:24 crc kubenswrapper[4821]: I1125 11:03:24.764302 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-dxdpz" Nov 25 11:03:25 crc kubenswrapper[4821]: I1125 11:03:25.011149 4821 scope.go:117] "RemoveContainer" containerID="93034c9c531c3037c12f30ef6bbf8418d7c92ca6f75ffe9429cb9b35c22c56c2" Nov 25 11:03:25 crc kubenswrapper[4821]: I1125 11:03:25.268364 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-dxdpz"] Nov 25 11:03:25 crc kubenswrapper[4821]: I1125 11:03:25.275381 4821 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 11:03:25 crc kubenswrapper[4821]: I1125 11:03:25.374443 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-dxdpz" event={"ID":"e1247bbd-54fc-45ce-8a90-0b8a5df02c47","Type":"ContainerStarted","Data":"f06a4c35ca05ffa4fb7aba6ddeb4a73a70846aed96a6179d594ca0eb1a8b4c0b"} Nov 25 11:03:27 crc kubenswrapper[4821]: I1125 11:03:27.114770 4821 scope.go:117] "RemoveContainer" containerID="fc16b2ea0cd36da4cef842d21755ec621134a4c21c5c964f8856fb8e6a102b2d" Nov 25 11:03:27 crc kubenswrapper[4821]: E1125 11:03:27.115789 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:03:27 crc kubenswrapper[4821]: I1125 11:03:27.393595 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-dxdpz" event={"ID":"e1247bbd-54fc-45ce-8a90-0b8a5df02c47","Type":"ContainerStarted","Data":"93caff6d8b42eb426bd3a41bb4b6b386273027141d20b4a192ce97697e4cea9a"} Nov 25 11:03:27 crc kubenswrapper[4821]: I1125 11:03:27.418237 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-dxdpz" podStartSLOduration=1.8676646049999999 podStartE2EDuration="3.418218825s" podCreationTimestamp="2025-11-25 11:03:24 +0000 UTC" firstStartedPulling="2025-11-25 11:03:25.275035686 +0000 UTC m=+1875.811355533" lastFinishedPulling="2025-11-25 11:03:26.825589906 +0000 UTC m=+1877.361909753" observedRunningTime="2025-11-25 11:03:27.415708874 +0000 UTC m=+1877.952028741" watchObservedRunningTime="2025-11-25 11:03:27.418218825 +0000 UTC m=+1877.954538682" Nov 25 11:03:34 crc kubenswrapper[4821]: I1125 11:03:34.457442 4821 generic.go:334] "Generic (PLEG): container finished" podID="e1247bbd-54fc-45ce-8a90-0b8a5df02c47" containerID="93caff6d8b42eb426bd3a41bb4b6b386273027141d20b4a192ce97697e4cea9a" exitCode=0 Nov 25 11:03:34 crc kubenswrapper[4821]: I1125 11:03:34.457554 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-dxdpz" event={"ID":"e1247bbd-54fc-45ce-8a90-0b8a5df02c47","Type":"ContainerDied","Data":"93caff6d8b42eb426bd3a41bb4b6b386273027141d20b4a192ce97697e4cea9a"} Nov 25 11:03:35 crc kubenswrapper[4821]: I1125 11:03:35.914834 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-dxdpz" Nov 25 11:03:36 crc kubenswrapper[4821]: I1125 11:03:36.057852 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/e1247bbd-54fc-45ce-8a90-0b8a5df02c47-inventory-0\") pod \"e1247bbd-54fc-45ce-8a90-0b8a5df02c47\" (UID: \"e1247bbd-54fc-45ce-8a90-0b8a5df02c47\") " Nov 25 11:03:36 crc kubenswrapper[4821]: I1125 11:03:36.058329 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n8sbl\" (UniqueName: \"kubernetes.io/projected/e1247bbd-54fc-45ce-8a90-0b8a5df02c47-kube-api-access-n8sbl\") pod \"e1247bbd-54fc-45ce-8a90-0b8a5df02c47\" (UID: \"e1247bbd-54fc-45ce-8a90-0b8a5df02c47\") " Nov 25 11:03:36 crc kubenswrapper[4821]: I1125 11:03:36.058351 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e1247bbd-54fc-45ce-8a90-0b8a5df02c47-ssh-key-openstack-edpm-ipam\") pod \"e1247bbd-54fc-45ce-8a90-0b8a5df02c47\" (UID: \"e1247bbd-54fc-45ce-8a90-0b8a5df02c47\") " Nov 25 11:03:36 crc kubenswrapper[4821]: I1125 11:03:36.063716 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1247bbd-54fc-45ce-8a90-0b8a5df02c47-kube-api-access-n8sbl" (OuterVolumeSpecName: "kube-api-access-n8sbl") pod "e1247bbd-54fc-45ce-8a90-0b8a5df02c47" (UID: "e1247bbd-54fc-45ce-8a90-0b8a5df02c47"). InnerVolumeSpecName "kube-api-access-n8sbl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:03:36 crc kubenswrapper[4821]: I1125 11:03:36.086531 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1247bbd-54fc-45ce-8a90-0b8a5df02c47-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "e1247bbd-54fc-45ce-8a90-0b8a5df02c47" (UID: "e1247bbd-54fc-45ce-8a90-0b8a5df02c47"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:03:36 crc kubenswrapper[4821]: I1125 11:03:36.091061 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1247bbd-54fc-45ce-8a90-0b8a5df02c47-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "e1247bbd-54fc-45ce-8a90-0b8a5df02c47" (UID: "e1247bbd-54fc-45ce-8a90-0b8a5df02c47"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:03:36 crc kubenswrapper[4821]: I1125 11:03:36.160410 4821 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/e1247bbd-54fc-45ce-8a90-0b8a5df02c47-inventory-0\") on node \"crc\" DevicePath \"\"" Nov 25 11:03:36 crc kubenswrapper[4821]: I1125 11:03:36.160453 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n8sbl\" (UniqueName: \"kubernetes.io/projected/e1247bbd-54fc-45ce-8a90-0b8a5df02c47-kube-api-access-n8sbl\") on node \"crc\" DevicePath \"\"" Nov 25 11:03:36 crc kubenswrapper[4821]: I1125 11:03:36.160463 4821 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e1247bbd-54fc-45ce-8a90-0b8a5df02c47-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Nov 25 11:03:36 crc kubenswrapper[4821]: E1125 11:03:36.285316 4821 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode1247bbd_54fc_45ce_8a90_0b8a5df02c47.slice\": RecentStats: unable to find data in memory cache]" Nov 25 11:03:36 crc kubenswrapper[4821]: I1125 11:03:36.489986 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-dxdpz" event={"ID":"e1247bbd-54fc-45ce-8a90-0b8a5df02c47","Type":"ContainerDied","Data":"f06a4c35ca05ffa4fb7aba6ddeb4a73a70846aed96a6179d594ca0eb1a8b4c0b"} Nov 25 11:03:36 crc kubenswrapper[4821]: I1125 11:03:36.490032 4821 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f06a4c35ca05ffa4fb7aba6ddeb4a73a70846aed96a6179d594ca0eb1a8b4c0b" Nov 25 11:03:36 crc kubenswrapper[4821]: I1125 11:03:36.490092 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-dxdpz" Nov 25 11:03:36 crc kubenswrapper[4821]: I1125 11:03:36.561846 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-j8sp8"] Nov 25 11:03:36 crc kubenswrapper[4821]: E1125 11:03:36.563005 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1247bbd-54fc-45ce-8a90-0b8a5df02c47" containerName="ssh-known-hosts-edpm-deployment" Nov 25 11:03:36 crc kubenswrapper[4821]: I1125 11:03:36.563129 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1247bbd-54fc-45ce-8a90-0b8a5df02c47" containerName="ssh-known-hosts-edpm-deployment" Nov 25 11:03:36 crc kubenswrapper[4821]: I1125 11:03:36.563501 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1247bbd-54fc-45ce-8a90-0b8a5df02c47" containerName="ssh-known-hosts-edpm-deployment" Nov 25 11:03:36 crc kubenswrapper[4821]: I1125 11:03:36.564467 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-j8sp8" Nov 25 11:03:36 crc kubenswrapper[4821]: I1125 11:03:36.567178 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 11:03:36 crc kubenswrapper[4821]: I1125 11:03:36.567959 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 11:03:36 crc kubenswrapper[4821]: I1125 11:03:36.568675 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 11:03:36 crc kubenswrapper[4821]: I1125 11:03:36.568897 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-z764x" Nov 25 11:03:36 crc kubenswrapper[4821]: I1125 11:03:36.576632 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-j8sp8"] Nov 25 11:03:36 crc kubenswrapper[4821]: I1125 11:03:36.669391 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8l8p\" (UniqueName: \"kubernetes.io/projected/3ad8bc38-c088-4632-8864-00d472c78fbf-kube-api-access-r8l8p\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-j8sp8\" (UID: \"3ad8bc38-c088-4632-8864-00d472c78fbf\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-j8sp8" Nov 25 11:03:36 crc kubenswrapper[4821]: I1125 11:03:36.669483 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3ad8bc38-c088-4632-8864-00d472c78fbf-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-j8sp8\" (UID: \"3ad8bc38-c088-4632-8864-00d472c78fbf\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-j8sp8" Nov 25 11:03:36 crc kubenswrapper[4821]: I1125 11:03:36.669725 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3ad8bc38-c088-4632-8864-00d472c78fbf-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-j8sp8\" (UID: \"3ad8bc38-c088-4632-8864-00d472c78fbf\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-j8sp8" Nov 25 11:03:36 crc kubenswrapper[4821]: I1125 11:03:36.772257 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3ad8bc38-c088-4632-8864-00d472c78fbf-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-j8sp8\" (UID: \"3ad8bc38-c088-4632-8864-00d472c78fbf\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-j8sp8" Nov 25 11:03:36 crc kubenswrapper[4821]: I1125 11:03:36.772393 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3ad8bc38-c088-4632-8864-00d472c78fbf-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-j8sp8\" (UID: \"3ad8bc38-c088-4632-8864-00d472c78fbf\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-j8sp8" Nov 25 11:03:36 crc kubenswrapper[4821]: I1125 11:03:36.772446 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8l8p\" (UniqueName: \"kubernetes.io/projected/3ad8bc38-c088-4632-8864-00d472c78fbf-kube-api-access-r8l8p\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-j8sp8\" (UID: \"3ad8bc38-c088-4632-8864-00d472c78fbf\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-j8sp8" Nov 25 11:03:36 crc kubenswrapper[4821]: I1125 11:03:36.776635 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3ad8bc38-c088-4632-8864-00d472c78fbf-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-j8sp8\" (UID: \"3ad8bc38-c088-4632-8864-00d472c78fbf\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-j8sp8" Nov 25 11:03:36 crc kubenswrapper[4821]: I1125 11:03:36.776667 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3ad8bc38-c088-4632-8864-00d472c78fbf-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-j8sp8\" (UID: \"3ad8bc38-c088-4632-8864-00d472c78fbf\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-j8sp8" Nov 25 11:03:36 crc kubenswrapper[4821]: I1125 11:03:36.800792 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8l8p\" (UniqueName: \"kubernetes.io/projected/3ad8bc38-c088-4632-8864-00d472c78fbf-kube-api-access-r8l8p\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-j8sp8\" (UID: \"3ad8bc38-c088-4632-8864-00d472c78fbf\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-j8sp8" Nov 25 11:03:36 crc kubenswrapper[4821]: I1125 11:03:36.885344 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-j8sp8" Nov 25 11:03:38 crc kubenswrapper[4821]: I1125 11:03:38.123958 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-j8sp8"] Nov 25 11:03:38 crc kubenswrapper[4821]: I1125 11:03:38.508352 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-j8sp8" event={"ID":"3ad8bc38-c088-4632-8864-00d472c78fbf","Type":"ContainerStarted","Data":"96cec5583209707e515ca84fde9f1b1897ebd0023813d3aa47c8bbf7cff3dfec"} Nov 25 11:03:39 crc kubenswrapper[4821]: I1125 11:03:39.516363 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-j8sp8" event={"ID":"3ad8bc38-c088-4632-8864-00d472c78fbf","Type":"ContainerStarted","Data":"62aa2bb459f1bc94e4789a0e26247f7d46304806f9f27c0f49a9c4f48fc11520"} Nov 25 11:03:39 crc kubenswrapper[4821]: I1125 11:03:39.535887 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-j8sp8" podStartSLOduration=2.776539169 podStartE2EDuration="3.535862609s" podCreationTimestamp="2025-11-25 11:03:36 +0000 UTC" firstStartedPulling="2025-11-25 11:03:38.126117431 +0000 UTC m=+1888.662437278" lastFinishedPulling="2025-11-25 11:03:38.885440871 +0000 UTC m=+1889.421760718" observedRunningTime="2025-11-25 11:03:39.529778909 +0000 UTC m=+1890.066098756" watchObservedRunningTime="2025-11-25 11:03:39.535862609 +0000 UTC m=+1890.072182456" Nov 25 11:03:40 crc kubenswrapper[4821]: I1125 11:03:40.120823 4821 scope.go:117] "RemoveContainer" containerID="fc16b2ea0cd36da4cef842d21755ec621134a4c21c5c964f8856fb8e6a102b2d" Nov 25 11:03:40 crc kubenswrapper[4821]: E1125 11:03:40.121366 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:03:47 crc kubenswrapper[4821]: I1125 11:03:47.578763 4821 generic.go:334] "Generic (PLEG): container finished" podID="3ad8bc38-c088-4632-8864-00d472c78fbf" containerID="62aa2bb459f1bc94e4789a0e26247f7d46304806f9f27c0f49a9c4f48fc11520" exitCode=0 Nov 25 11:03:47 crc kubenswrapper[4821]: I1125 11:03:47.578875 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-j8sp8" event={"ID":"3ad8bc38-c088-4632-8864-00d472c78fbf","Type":"ContainerDied","Data":"62aa2bb459f1bc94e4789a0e26247f7d46304806f9f27c0f49a9c4f48fc11520"} Nov 25 11:03:49 crc kubenswrapper[4821]: I1125 11:03:49.019045 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-j8sp8" Nov 25 11:03:49 crc kubenswrapper[4821]: I1125 11:03:49.099085 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3ad8bc38-c088-4632-8864-00d472c78fbf-inventory\") pod \"3ad8bc38-c088-4632-8864-00d472c78fbf\" (UID: \"3ad8bc38-c088-4632-8864-00d472c78fbf\") " Nov 25 11:03:49 crc kubenswrapper[4821]: I1125 11:03:49.099225 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3ad8bc38-c088-4632-8864-00d472c78fbf-ssh-key\") pod \"3ad8bc38-c088-4632-8864-00d472c78fbf\" (UID: \"3ad8bc38-c088-4632-8864-00d472c78fbf\") " Nov 25 11:03:49 crc kubenswrapper[4821]: I1125 11:03:49.099353 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r8l8p\" (UniqueName: \"kubernetes.io/projected/3ad8bc38-c088-4632-8864-00d472c78fbf-kube-api-access-r8l8p\") pod \"3ad8bc38-c088-4632-8864-00d472c78fbf\" (UID: \"3ad8bc38-c088-4632-8864-00d472c78fbf\") " Nov 25 11:03:49 crc kubenswrapper[4821]: I1125 11:03:49.116382 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ad8bc38-c088-4632-8864-00d472c78fbf-kube-api-access-r8l8p" (OuterVolumeSpecName: "kube-api-access-r8l8p") pod "3ad8bc38-c088-4632-8864-00d472c78fbf" (UID: "3ad8bc38-c088-4632-8864-00d472c78fbf"). InnerVolumeSpecName "kube-api-access-r8l8p". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:03:49 crc kubenswrapper[4821]: I1125 11:03:49.132696 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ad8bc38-c088-4632-8864-00d472c78fbf-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3ad8bc38-c088-4632-8864-00d472c78fbf" (UID: "3ad8bc38-c088-4632-8864-00d472c78fbf"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:03:49 crc kubenswrapper[4821]: I1125 11:03:49.137559 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ad8bc38-c088-4632-8864-00d472c78fbf-inventory" (OuterVolumeSpecName: "inventory") pod "3ad8bc38-c088-4632-8864-00d472c78fbf" (UID: "3ad8bc38-c088-4632-8864-00d472c78fbf"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:03:49 crc kubenswrapper[4821]: I1125 11:03:49.202007 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r8l8p\" (UniqueName: \"kubernetes.io/projected/3ad8bc38-c088-4632-8864-00d472c78fbf-kube-api-access-r8l8p\") on node \"crc\" DevicePath \"\"" Nov 25 11:03:49 crc kubenswrapper[4821]: I1125 11:03:49.202050 4821 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3ad8bc38-c088-4632-8864-00d472c78fbf-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 11:03:49 crc kubenswrapper[4821]: I1125 11:03:49.202059 4821 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3ad8bc38-c088-4632-8864-00d472c78fbf-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 11:03:49 crc kubenswrapper[4821]: I1125 11:03:49.597120 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-j8sp8" event={"ID":"3ad8bc38-c088-4632-8864-00d472c78fbf","Type":"ContainerDied","Data":"96cec5583209707e515ca84fde9f1b1897ebd0023813d3aa47c8bbf7cff3dfec"} Nov 25 11:03:49 crc kubenswrapper[4821]: I1125 11:03:49.597437 4821 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="96cec5583209707e515ca84fde9f1b1897ebd0023813d3aa47c8bbf7cff3dfec" Nov 25 11:03:49 crc kubenswrapper[4821]: I1125 11:03:49.597176 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-j8sp8" Nov 25 11:03:49 crc kubenswrapper[4821]: I1125 11:03:49.670095 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-98mzz"] Nov 25 11:03:49 crc kubenswrapper[4821]: E1125 11:03:49.670464 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ad8bc38-c088-4632-8864-00d472c78fbf" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 25 11:03:49 crc kubenswrapper[4821]: I1125 11:03:49.670481 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ad8bc38-c088-4632-8864-00d472c78fbf" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 25 11:03:49 crc kubenswrapper[4821]: I1125 11:03:49.670703 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ad8bc38-c088-4632-8864-00d472c78fbf" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 25 11:03:49 crc kubenswrapper[4821]: I1125 11:03:49.671325 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-98mzz" Nov 25 11:03:49 crc kubenswrapper[4821]: I1125 11:03:49.675615 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 11:03:49 crc kubenswrapper[4821]: I1125 11:03:49.675665 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 11:03:49 crc kubenswrapper[4821]: I1125 11:03:49.675833 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 11:03:49 crc kubenswrapper[4821]: I1125 11:03:49.675854 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-z764x" Nov 25 11:03:49 crc kubenswrapper[4821]: I1125 11:03:49.683877 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-98mzz"] Nov 25 11:03:49 crc kubenswrapper[4821]: I1125 11:03:49.813359 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0d9e3bea-dfb1-479d-a1b2-a17b2334cd86-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-98mzz\" (UID: \"0d9e3bea-dfb1-479d-a1b2-a17b2334cd86\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-98mzz" Nov 25 11:03:49 crc kubenswrapper[4821]: I1125 11:03:49.813529 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0d9e3bea-dfb1-479d-a1b2-a17b2334cd86-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-98mzz\" (UID: \"0d9e3bea-dfb1-479d-a1b2-a17b2334cd86\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-98mzz" Nov 25 11:03:49 crc kubenswrapper[4821]: I1125 11:03:49.813807 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-48r79\" (UniqueName: \"kubernetes.io/projected/0d9e3bea-dfb1-479d-a1b2-a17b2334cd86-kube-api-access-48r79\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-98mzz\" (UID: \"0d9e3bea-dfb1-479d-a1b2-a17b2334cd86\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-98mzz" Nov 25 11:03:49 crc kubenswrapper[4821]: I1125 11:03:49.916647 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-48r79\" (UniqueName: \"kubernetes.io/projected/0d9e3bea-dfb1-479d-a1b2-a17b2334cd86-kube-api-access-48r79\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-98mzz\" (UID: \"0d9e3bea-dfb1-479d-a1b2-a17b2334cd86\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-98mzz" Nov 25 11:03:49 crc kubenswrapper[4821]: I1125 11:03:49.919223 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0d9e3bea-dfb1-479d-a1b2-a17b2334cd86-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-98mzz\" (UID: \"0d9e3bea-dfb1-479d-a1b2-a17b2334cd86\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-98mzz" Nov 25 11:03:49 crc kubenswrapper[4821]: I1125 11:03:49.919343 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0d9e3bea-dfb1-479d-a1b2-a17b2334cd86-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-98mzz\" (UID: \"0d9e3bea-dfb1-479d-a1b2-a17b2334cd86\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-98mzz" Nov 25 11:03:49 crc kubenswrapper[4821]: I1125 11:03:49.923359 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0d9e3bea-dfb1-479d-a1b2-a17b2334cd86-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-98mzz\" (UID: \"0d9e3bea-dfb1-479d-a1b2-a17b2334cd86\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-98mzz" Nov 25 11:03:49 crc kubenswrapper[4821]: I1125 11:03:49.923667 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0d9e3bea-dfb1-479d-a1b2-a17b2334cd86-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-98mzz\" (UID: \"0d9e3bea-dfb1-479d-a1b2-a17b2334cd86\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-98mzz" Nov 25 11:03:49 crc kubenswrapper[4821]: I1125 11:03:49.933768 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-48r79\" (UniqueName: \"kubernetes.io/projected/0d9e3bea-dfb1-479d-a1b2-a17b2334cd86-kube-api-access-48r79\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-98mzz\" (UID: \"0d9e3bea-dfb1-479d-a1b2-a17b2334cd86\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-98mzz" Nov 25 11:03:49 crc kubenswrapper[4821]: I1125 11:03:49.992515 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-98mzz" Nov 25 11:03:50 crc kubenswrapper[4821]: I1125 11:03:50.575696 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-98mzz"] Nov 25 11:03:50 crc kubenswrapper[4821]: I1125 11:03:50.605492 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-98mzz" event={"ID":"0d9e3bea-dfb1-479d-a1b2-a17b2334cd86","Type":"ContainerStarted","Data":"bccc751e2fa6d32b735a4e557e93488f1bd08c0b3f206ab8444e48c1730ea63b"} Nov 25 11:03:51 crc kubenswrapper[4821]: I1125 11:03:51.117249 4821 scope.go:117] "RemoveContainer" containerID="fc16b2ea0cd36da4cef842d21755ec621134a4c21c5c964f8856fb8e6a102b2d" Nov 25 11:03:51 crc kubenswrapper[4821]: E1125 11:03:51.117728 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:03:52 crc kubenswrapper[4821]: I1125 11:03:52.624184 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-98mzz" event={"ID":"0d9e3bea-dfb1-479d-a1b2-a17b2334cd86","Type":"ContainerStarted","Data":"35356198aab5e835743637e4260c98a249a401f849419a0808e55f79190ecf37"} Nov 25 11:03:52 crc kubenswrapper[4821]: I1125 11:03:52.644238 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-98mzz" podStartSLOduration=2.300827063 podStartE2EDuration="3.644212252s" podCreationTimestamp="2025-11-25 11:03:49 +0000 UTC" firstStartedPulling="2025-11-25 11:03:50.584888406 +0000 UTC m=+1901.121208253" lastFinishedPulling="2025-11-25 11:03:51.928273585 +0000 UTC m=+1902.464593442" observedRunningTime="2025-11-25 11:03:52.638149942 +0000 UTC m=+1903.174469889" watchObservedRunningTime="2025-11-25 11:03:52.644212252 +0000 UTC m=+1903.180532099" Nov 25 11:04:02 crc kubenswrapper[4821]: I1125 11:04:02.715451 4821 generic.go:334] "Generic (PLEG): container finished" podID="0d9e3bea-dfb1-479d-a1b2-a17b2334cd86" containerID="35356198aab5e835743637e4260c98a249a401f849419a0808e55f79190ecf37" exitCode=0 Nov 25 11:04:02 crc kubenswrapper[4821]: I1125 11:04:02.715531 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-98mzz" event={"ID":"0d9e3bea-dfb1-479d-a1b2-a17b2334cd86","Type":"ContainerDied","Data":"35356198aab5e835743637e4260c98a249a401f849419a0808e55f79190ecf37"} Nov 25 11:04:03 crc kubenswrapper[4821]: I1125 11:04:03.114690 4821 scope.go:117] "RemoveContainer" containerID="fc16b2ea0cd36da4cef842d21755ec621134a4c21c5c964f8856fb8e6a102b2d" Nov 25 11:04:03 crc kubenswrapper[4821]: E1125 11:04:03.115240 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:04:04 crc kubenswrapper[4821]: I1125 11:04:04.100552 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-98mzz" Nov 25 11:04:04 crc kubenswrapper[4821]: I1125 11:04:04.213421 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0d9e3bea-dfb1-479d-a1b2-a17b2334cd86-ssh-key\") pod \"0d9e3bea-dfb1-479d-a1b2-a17b2334cd86\" (UID: \"0d9e3bea-dfb1-479d-a1b2-a17b2334cd86\") " Nov 25 11:04:04 crc kubenswrapper[4821]: I1125 11:04:04.213545 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0d9e3bea-dfb1-479d-a1b2-a17b2334cd86-inventory\") pod \"0d9e3bea-dfb1-479d-a1b2-a17b2334cd86\" (UID: \"0d9e3bea-dfb1-479d-a1b2-a17b2334cd86\") " Nov 25 11:04:04 crc kubenswrapper[4821]: I1125 11:04:04.213819 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-48r79\" (UniqueName: \"kubernetes.io/projected/0d9e3bea-dfb1-479d-a1b2-a17b2334cd86-kube-api-access-48r79\") pod \"0d9e3bea-dfb1-479d-a1b2-a17b2334cd86\" (UID: \"0d9e3bea-dfb1-479d-a1b2-a17b2334cd86\") " Nov 25 11:04:04 crc kubenswrapper[4821]: I1125 11:04:04.221084 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d9e3bea-dfb1-479d-a1b2-a17b2334cd86-kube-api-access-48r79" (OuterVolumeSpecName: "kube-api-access-48r79") pod "0d9e3bea-dfb1-479d-a1b2-a17b2334cd86" (UID: "0d9e3bea-dfb1-479d-a1b2-a17b2334cd86"). InnerVolumeSpecName "kube-api-access-48r79". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:04:04 crc kubenswrapper[4821]: I1125 11:04:04.240750 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d9e3bea-dfb1-479d-a1b2-a17b2334cd86-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0d9e3bea-dfb1-479d-a1b2-a17b2334cd86" (UID: "0d9e3bea-dfb1-479d-a1b2-a17b2334cd86"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:04:04 crc kubenswrapper[4821]: I1125 11:04:04.242781 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d9e3bea-dfb1-479d-a1b2-a17b2334cd86-inventory" (OuterVolumeSpecName: "inventory") pod "0d9e3bea-dfb1-479d-a1b2-a17b2334cd86" (UID: "0d9e3bea-dfb1-479d-a1b2-a17b2334cd86"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:04:04 crc kubenswrapper[4821]: I1125 11:04:04.316874 4821 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0d9e3bea-dfb1-479d-a1b2-a17b2334cd86-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 11:04:04 crc kubenswrapper[4821]: I1125 11:04:04.316909 4821 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0d9e3bea-dfb1-479d-a1b2-a17b2334cd86-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 11:04:04 crc kubenswrapper[4821]: I1125 11:04:04.316938 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-48r79\" (UniqueName: \"kubernetes.io/projected/0d9e3bea-dfb1-479d-a1b2-a17b2334cd86-kube-api-access-48r79\") on node \"crc\" DevicePath \"\"" Nov 25 11:04:04 crc kubenswrapper[4821]: I1125 11:04:04.732444 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-98mzz" event={"ID":"0d9e3bea-dfb1-479d-a1b2-a17b2334cd86","Type":"ContainerDied","Data":"bccc751e2fa6d32b735a4e557e93488f1bd08c0b3f206ab8444e48c1730ea63b"} Nov 25 11:04:04 crc kubenswrapper[4821]: I1125 11:04:04.732484 4821 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bccc751e2fa6d32b735a4e557e93488f1bd08c0b3f206ab8444e48c1730ea63b" Nov 25 11:04:04 crc kubenswrapper[4821]: I1125 11:04:04.732531 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-98mzz" Nov 25 11:04:04 crc kubenswrapper[4821]: I1125 11:04:04.893297 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-48x26"] Nov 25 11:04:04 crc kubenswrapper[4821]: E1125 11:04:04.893727 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d9e3bea-dfb1-479d-a1b2-a17b2334cd86" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 25 11:04:04 crc kubenswrapper[4821]: I1125 11:04:04.893750 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d9e3bea-dfb1-479d-a1b2-a17b2334cd86" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 25 11:04:04 crc kubenswrapper[4821]: I1125 11:04:04.893990 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d9e3bea-dfb1-479d-a1b2-a17b2334cd86" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 25 11:04:04 crc kubenswrapper[4821]: I1125 11:04:04.894743 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-48x26" Nov 25 11:04:04 crc kubenswrapper[4821]: I1125 11:04:04.897781 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Nov 25 11:04:04 crc kubenswrapper[4821]: I1125 11:04:04.897817 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 11:04:04 crc kubenswrapper[4821]: I1125 11:04:04.898074 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-z764x" Nov 25 11:04:04 crc kubenswrapper[4821]: I1125 11:04:04.898741 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 11:04:04 crc kubenswrapper[4821]: I1125 11:04:04.899015 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 11:04:04 crc kubenswrapper[4821]: I1125 11:04:04.899043 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Nov 25 11:04:04 crc kubenswrapper[4821]: I1125 11:04:04.903779 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Nov 25 11:04:04 crc kubenswrapper[4821]: I1125 11:04:04.904772 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Nov 25 11:04:04 crc kubenswrapper[4821]: I1125 11:04:04.914734 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-48x26"] Nov 25 11:04:05 crc kubenswrapper[4821]: I1125 11:04:05.028949 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/98d0e490-2704-4aa6-baa5-32ec701a5b68-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-48x26\" (UID: \"98d0e490-2704-4aa6-baa5-32ec701a5b68\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-48x26" Nov 25 11:04:05 crc kubenswrapper[4821]: I1125 11:04:05.029082 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98d0e490-2704-4aa6-baa5-32ec701a5b68-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-48x26\" (UID: \"98d0e490-2704-4aa6-baa5-32ec701a5b68\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-48x26" Nov 25 11:04:05 crc kubenswrapper[4821]: I1125 11:04:05.029146 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/98d0e490-2704-4aa6-baa5-32ec701a5b68-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-48x26\" (UID: \"98d0e490-2704-4aa6-baa5-32ec701a5b68\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-48x26" Nov 25 11:04:05 crc kubenswrapper[4821]: I1125 11:04:05.029208 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4x9dl\" (UniqueName: \"kubernetes.io/projected/98d0e490-2704-4aa6-baa5-32ec701a5b68-kube-api-access-4x9dl\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-48x26\" (UID: \"98d0e490-2704-4aa6-baa5-32ec701a5b68\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-48x26" Nov 25 11:04:05 crc kubenswrapper[4821]: I1125 11:04:05.029252 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98d0e490-2704-4aa6-baa5-32ec701a5b68-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-48x26\" (UID: \"98d0e490-2704-4aa6-baa5-32ec701a5b68\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-48x26" Nov 25 11:04:05 crc kubenswrapper[4821]: I1125 11:04:05.029275 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98d0e490-2704-4aa6-baa5-32ec701a5b68-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-48x26\" (UID: \"98d0e490-2704-4aa6-baa5-32ec701a5b68\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-48x26" Nov 25 11:04:05 crc kubenswrapper[4821]: I1125 11:04:05.029299 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/98d0e490-2704-4aa6-baa5-32ec701a5b68-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-48x26\" (UID: \"98d0e490-2704-4aa6-baa5-32ec701a5b68\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-48x26" Nov 25 11:04:05 crc kubenswrapper[4821]: I1125 11:04:05.029339 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/98d0e490-2704-4aa6-baa5-32ec701a5b68-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-48x26\" (UID: \"98d0e490-2704-4aa6-baa5-32ec701a5b68\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-48x26" Nov 25 11:04:05 crc kubenswrapper[4821]: I1125 11:04:05.029370 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98d0e490-2704-4aa6-baa5-32ec701a5b68-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-48x26\" (UID: \"98d0e490-2704-4aa6-baa5-32ec701a5b68\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-48x26" Nov 25 11:04:05 crc kubenswrapper[4821]: I1125 11:04:05.029421 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98d0e490-2704-4aa6-baa5-32ec701a5b68-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-48x26\" (UID: \"98d0e490-2704-4aa6-baa5-32ec701a5b68\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-48x26" Nov 25 11:04:05 crc kubenswrapper[4821]: I1125 11:04:05.029450 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98d0e490-2704-4aa6-baa5-32ec701a5b68-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-48x26\" (UID: \"98d0e490-2704-4aa6-baa5-32ec701a5b68\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-48x26" Nov 25 11:04:05 crc kubenswrapper[4821]: I1125 11:04:05.029568 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98d0e490-2704-4aa6-baa5-32ec701a5b68-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-48x26\" (UID: \"98d0e490-2704-4aa6-baa5-32ec701a5b68\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-48x26" Nov 25 11:04:05 crc kubenswrapper[4821]: I1125 11:04:05.029629 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/98d0e490-2704-4aa6-baa5-32ec701a5b68-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-48x26\" (UID: \"98d0e490-2704-4aa6-baa5-32ec701a5b68\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-48x26" Nov 25 11:04:05 crc kubenswrapper[4821]: I1125 11:04:05.029714 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/98d0e490-2704-4aa6-baa5-32ec701a5b68-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-48x26\" (UID: \"98d0e490-2704-4aa6-baa5-32ec701a5b68\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-48x26" Nov 25 11:04:05 crc kubenswrapper[4821]: I1125 11:04:05.132005 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98d0e490-2704-4aa6-baa5-32ec701a5b68-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-48x26\" (UID: \"98d0e490-2704-4aa6-baa5-32ec701a5b68\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-48x26" Nov 25 11:04:05 crc kubenswrapper[4821]: I1125 11:04:05.132115 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/98d0e490-2704-4aa6-baa5-32ec701a5b68-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-48x26\" (UID: \"98d0e490-2704-4aa6-baa5-32ec701a5b68\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-48x26" Nov 25 11:04:05 crc kubenswrapper[4821]: I1125 11:04:05.132230 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4x9dl\" (UniqueName: \"kubernetes.io/projected/98d0e490-2704-4aa6-baa5-32ec701a5b68-kube-api-access-4x9dl\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-48x26\" (UID: \"98d0e490-2704-4aa6-baa5-32ec701a5b68\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-48x26" Nov 25 11:04:05 crc kubenswrapper[4821]: I1125 11:04:05.132294 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98d0e490-2704-4aa6-baa5-32ec701a5b68-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-48x26\" (UID: \"98d0e490-2704-4aa6-baa5-32ec701a5b68\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-48x26" Nov 25 11:04:05 crc kubenswrapper[4821]: I1125 11:04:05.132331 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98d0e490-2704-4aa6-baa5-32ec701a5b68-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-48x26\" (UID: \"98d0e490-2704-4aa6-baa5-32ec701a5b68\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-48x26" Nov 25 11:04:05 crc kubenswrapper[4821]: I1125 11:04:05.132368 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/98d0e490-2704-4aa6-baa5-32ec701a5b68-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-48x26\" (UID: \"98d0e490-2704-4aa6-baa5-32ec701a5b68\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-48x26" Nov 25 11:04:05 crc kubenswrapper[4821]: I1125 11:04:05.132420 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/98d0e490-2704-4aa6-baa5-32ec701a5b68-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-48x26\" (UID: \"98d0e490-2704-4aa6-baa5-32ec701a5b68\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-48x26" Nov 25 11:04:05 crc kubenswrapper[4821]: I1125 11:04:05.132461 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98d0e490-2704-4aa6-baa5-32ec701a5b68-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-48x26\" (UID: \"98d0e490-2704-4aa6-baa5-32ec701a5b68\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-48x26" Nov 25 11:04:05 crc kubenswrapper[4821]: I1125 11:04:05.132500 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98d0e490-2704-4aa6-baa5-32ec701a5b68-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-48x26\" (UID: \"98d0e490-2704-4aa6-baa5-32ec701a5b68\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-48x26" Nov 25 11:04:05 crc kubenswrapper[4821]: I1125 11:04:05.132535 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98d0e490-2704-4aa6-baa5-32ec701a5b68-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-48x26\" (UID: \"98d0e490-2704-4aa6-baa5-32ec701a5b68\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-48x26" Nov 25 11:04:05 crc kubenswrapper[4821]: I1125 11:04:05.132574 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98d0e490-2704-4aa6-baa5-32ec701a5b68-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-48x26\" (UID: \"98d0e490-2704-4aa6-baa5-32ec701a5b68\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-48x26" Nov 25 11:04:05 crc kubenswrapper[4821]: I1125 11:04:05.132614 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/98d0e490-2704-4aa6-baa5-32ec701a5b68-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-48x26\" (UID: \"98d0e490-2704-4aa6-baa5-32ec701a5b68\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-48x26" Nov 25 11:04:05 crc kubenswrapper[4821]: I1125 11:04:05.132689 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/98d0e490-2704-4aa6-baa5-32ec701a5b68-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-48x26\" (UID: \"98d0e490-2704-4aa6-baa5-32ec701a5b68\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-48x26" Nov 25 11:04:05 crc kubenswrapper[4821]: I1125 11:04:05.132775 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/98d0e490-2704-4aa6-baa5-32ec701a5b68-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-48x26\" (UID: \"98d0e490-2704-4aa6-baa5-32ec701a5b68\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-48x26" Nov 25 11:04:05 crc kubenswrapper[4821]: I1125 11:04:05.135968 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/98d0e490-2704-4aa6-baa5-32ec701a5b68-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-48x26\" (UID: \"98d0e490-2704-4aa6-baa5-32ec701a5b68\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-48x26" Nov 25 11:04:05 crc kubenswrapper[4821]: I1125 11:04:05.136118 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98d0e490-2704-4aa6-baa5-32ec701a5b68-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-48x26\" (UID: \"98d0e490-2704-4aa6-baa5-32ec701a5b68\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-48x26" Nov 25 11:04:05 crc kubenswrapper[4821]: I1125 11:04:05.136465 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98d0e490-2704-4aa6-baa5-32ec701a5b68-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-48x26\" (UID: \"98d0e490-2704-4aa6-baa5-32ec701a5b68\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-48x26" Nov 25 11:04:05 crc kubenswrapper[4821]: I1125 11:04:05.138990 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98d0e490-2704-4aa6-baa5-32ec701a5b68-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-48x26\" (UID: \"98d0e490-2704-4aa6-baa5-32ec701a5b68\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-48x26" Nov 25 11:04:05 crc kubenswrapper[4821]: I1125 11:04:05.139879 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98d0e490-2704-4aa6-baa5-32ec701a5b68-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-48x26\" (UID: \"98d0e490-2704-4aa6-baa5-32ec701a5b68\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-48x26" Nov 25 11:04:05 crc kubenswrapper[4821]: I1125 11:04:05.140687 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98d0e490-2704-4aa6-baa5-32ec701a5b68-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-48x26\" (UID: \"98d0e490-2704-4aa6-baa5-32ec701a5b68\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-48x26" Nov 25 11:04:05 crc kubenswrapper[4821]: I1125 11:04:05.140712 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98d0e490-2704-4aa6-baa5-32ec701a5b68-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-48x26\" (UID: \"98d0e490-2704-4aa6-baa5-32ec701a5b68\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-48x26" Nov 25 11:04:05 crc kubenswrapper[4821]: I1125 11:04:05.141311 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/98d0e490-2704-4aa6-baa5-32ec701a5b68-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-48x26\" (UID: \"98d0e490-2704-4aa6-baa5-32ec701a5b68\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-48x26" Nov 25 11:04:05 crc kubenswrapper[4821]: I1125 11:04:05.141524 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/98d0e490-2704-4aa6-baa5-32ec701a5b68-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-48x26\" (UID: \"98d0e490-2704-4aa6-baa5-32ec701a5b68\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-48x26" Nov 25 11:04:05 crc kubenswrapper[4821]: I1125 11:04:05.143468 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/98d0e490-2704-4aa6-baa5-32ec701a5b68-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-48x26\" (UID: \"98d0e490-2704-4aa6-baa5-32ec701a5b68\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-48x26" Nov 25 11:04:05 crc kubenswrapper[4821]: I1125 11:04:05.143503 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/98d0e490-2704-4aa6-baa5-32ec701a5b68-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-48x26\" (UID: \"98d0e490-2704-4aa6-baa5-32ec701a5b68\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-48x26" Nov 25 11:04:05 crc kubenswrapper[4821]: I1125 11:04:05.143571 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/98d0e490-2704-4aa6-baa5-32ec701a5b68-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-48x26\" (UID: \"98d0e490-2704-4aa6-baa5-32ec701a5b68\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-48x26" Nov 25 11:04:05 crc kubenswrapper[4821]: I1125 11:04:05.150046 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4x9dl\" (UniqueName: \"kubernetes.io/projected/98d0e490-2704-4aa6-baa5-32ec701a5b68-kube-api-access-4x9dl\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-48x26\" (UID: \"98d0e490-2704-4aa6-baa5-32ec701a5b68\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-48x26" Nov 25 11:04:05 crc kubenswrapper[4821]: I1125 11:04:05.151024 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98d0e490-2704-4aa6-baa5-32ec701a5b68-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-48x26\" (UID: \"98d0e490-2704-4aa6-baa5-32ec701a5b68\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-48x26" Nov 25 11:04:05 crc kubenswrapper[4821]: I1125 11:04:05.214765 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-48x26" Nov 25 11:04:05 crc kubenswrapper[4821]: I1125 11:04:05.757235 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-48x26"] Nov 25 11:04:06 crc kubenswrapper[4821]: I1125 11:04:06.751291 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-48x26" event={"ID":"98d0e490-2704-4aa6-baa5-32ec701a5b68","Type":"ContainerStarted","Data":"6c9cbed16ee27a1d66974234738007e63d57d5a119d9d3ea19fc2db9313407d7"} Nov 25 11:04:07 crc kubenswrapper[4821]: I1125 11:04:07.763000 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-48x26" event={"ID":"98d0e490-2704-4aa6-baa5-32ec701a5b68","Type":"ContainerStarted","Data":"f67b3494e84c24692140d418eec23d5fb51a7783d53b71cf7dc152efe9ea3fa9"} Nov 25 11:04:07 crc kubenswrapper[4821]: I1125 11:04:07.797823 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-48x26" podStartSLOduration=2.450741516 podStartE2EDuration="3.797805494s" podCreationTimestamp="2025-11-25 11:04:04 +0000 UTC" firstStartedPulling="2025-11-25 11:04:05.772702845 +0000 UTC m=+1916.309022702" lastFinishedPulling="2025-11-25 11:04:07.119766793 +0000 UTC m=+1917.656086680" observedRunningTime="2025-11-25 11:04:07.786209168 +0000 UTC m=+1918.322529055" watchObservedRunningTime="2025-11-25 11:04:07.797805494 +0000 UTC m=+1918.334125341" Nov 25 11:04:16 crc kubenswrapper[4821]: I1125 11:04:16.115072 4821 scope.go:117] "RemoveContainer" containerID="fc16b2ea0cd36da4cef842d21755ec621134a4c21c5c964f8856fb8e6a102b2d" Nov 25 11:04:16 crc kubenswrapper[4821]: I1125 11:04:16.842876 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" event={"ID":"5f948b87-ac86-4de6-ad64-c2ef947f84d4","Type":"ContainerStarted","Data":"67625e9eb479a7295b3be108a852f5678838cf58bdadd2502f30ab844bb33b89"} Nov 25 11:04:47 crc kubenswrapper[4821]: I1125 11:04:47.095587 4821 generic.go:334] "Generic (PLEG): container finished" podID="98d0e490-2704-4aa6-baa5-32ec701a5b68" containerID="f67b3494e84c24692140d418eec23d5fb51a7783d53b71cf7dc152efe9ea3fa9" exitCode=0 Nov 25 11:04:47 crc kubenswrapper[4821]: I1125 11:04:47.095757 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-48x26" event={"ID":"98d0e490-2704-4aa6-baa5-32ec701a5b68","Type":"ContainerDied","Data":"f67b3494e84c24692140d418eec23d5fb51a7783d53b71cf7dc152efe9ea3fa9"} Nov 25 11:04:48 crc kubenswrapper[4821]: I1125 11:04:48.484459 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-48x26" Nov 25 11:04:48 crc kubenswrapper[4821]: I1125 11:04:48.650863 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/98d0e490-2704-4aa6-baa5-32ec701a5b68-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"98d0e490-2704-4aa6-baa5-32ec701a5b68\" (UID: \"98d0e490-2704-4aa6-baa5-32ec701a5b68\") " Nov 25 11:04:48 crc kubenswrapper[4821]: I1125 11:04:48.651374 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/98d0e490-2704-4aa6-baa5-32ec701a5b68-openstack-edpm-ipam-ovn-default-certs-0\") pod \"98d0e490-2704-4aa6-baa5-32ec701a5b68\" (UID: \"98d0e490-2704-4aa6-baa5-32ec701a5b68\") " Nov 25 11:04:48 crc kubenswrapper[4821]: I1125 11:04:48.651435 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98d0e490-2704-4aa6-baa5-32ec701a5b68-repo-setup-combined-ca-bundle\") pod \"98d0e490-2704-4aa6-baa5-32ec701a5b68\" (UID: \"98d0e490-2704-4aa6-baa5-32ec701a5b68\") " Nov 25 11:04:48 crc kubenswrapper[4821]: I1125 11:04:48.651893 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/98d0e490-2704-4aa6-baa5-32ec701a5b68-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"98d0e490-2704-4aa6-baa5-32ec701a5b68\" (UID: \"98d0e490-2704-4aa6-baa5-32ec701a5b68\") " Nov 25 11:04:48 crc kubenswrapper[4821]: I1125 11:04:48.651942 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98d0e490-2704-4aa6-baa5-32ec701a5b68-telemetry-combined-ca-bundle\") pod \"98d0e490-2704-4aa6-baa5-32ec701a5b68\" (UID: \"98d0e490-2704-4aa6-baa5-32ec701a5b68\") " Nov 25 11:04:48 crc kubenswrapper[4821]: I1125 11:04:48.651974 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4x9dl\" (UniqueName: \"kubernetes.io/projected/98d0e490-2704-4aa6-baa5-32ec701a5b68-kube-api-access-4x9dl\") pod \"98d0e490-2704-4aa6-baa5-32ec701a5b68\" (UID: \"98d0e490-2704-4aa6-baa5-32ec701a5b68\") " Nov 25 11:04:48 crc kubenswrapper[4821]: I1125 11:04:48.652008 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98d0e490-2704-4aa6-baa5-32ec701a5b68-nova-combined-ca-bundle\") pod \"98d0e490-2704-4aa6-baa5-32ec701a5b68\" (UID: \"98d0e490-2704-4aa6-baa5-32ec701a5b68\") " Nov 25 11:04:48 crc kubenswrapper[4821]: I1125 11:04:48.652046 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98d0e490-2704-4aa6-baa5-32ec701a5b68-neutron-metadata-combined-ca-bundle\") pod \"98d0e490-2704-4aa6-baa5-32ec701a5b68\" (UID: \"98d0e490-2704-4aa6-baa5-32ec701a5b68\") " Nov 25 11:04:48 crc kubenswrapper[4821]: I1125 11:04:48.652083 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/98d0e490-2704-4aa6-baa5-32ec701a5b68-inventory\") pod \"98d0e490-2704-4aa6-baa5-32ec701a5b68\" (UID: \"98d0e490-2704-4aa6-baa5-32ec701a5b68\") " Nov 25 11:04:48 crc kubenswrapper[4821]: I1125 11:04:48.652105 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98d0e490-2704-4aa6-baa5-32ec701a5b68-ovn-combined-ca-bundle\") pod \"98d0e490-2704-4aa6-baa5-32ec701a5b68\" (UID: \"98d0e490-2704-4aa6-baa5-32ec701a5b68\") " Nov 25 11:04:48 crc kubenswrapper[4821]: I1125 11:04:48.652134 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/98d0e490-2704-4aa6-baa5-32ec701a5b68-ssh-key\") pod \"98d0e490-2704-4aa6-baa5-32ec701a5b68\" (UID: \"98d0e490-2704-4aa6-baa5-32ec701a5b68\") " Nov 25 11:04:48 crc kubenswrapper[4821]: I1125 11:04:48.652178 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98d0e490-2704-4aa6-baa5-32ec701a5b68-bootstrap-combined-ca-bundle\") pod \"98d0e490-2704-4aa6-baa5-32ec701a5b68\" (UID: \"98d0e490-2704-4aa6-baa5-32ec701a5b68\") " Nov 25 11:04:48 crc kubenswrapper[4821]: I1125 11:04:48.652200 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98d0e490-2704-4aa6-baa5-32ec701a5b68-libvirt-combined-ca-bundle\") pod \"98d0e490-2704-4aa6-baa5-32ec701a5b68\" (UID: \"98d0e490-2704-4aa6-baa5-32ec701a5b68\") " Nov 25 11:04:48 crc kubenswrapper[4821]: I1125 11:04:48.652225 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/98d0e490-2704-4aa6-baa5-32ec701a5b68-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"98d0e490-2704-4aa6-baa5-32ec701a5b68\" (UID: \"98d0e490-2704-4aa6-baa5-32ec701a5b68\") " Nov 25 11:04:48 crc kubenswrapper[4821]: I1125 11:04:48.657416 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98d0e490-2704-4aa6-baa5-32ec701a5b68-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "98d0e490-2704-4aa6-baa5-32ec701a5b68" (UID: "98d0e490-2704-4aa6-baa5-32ec701a5b68"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:04:48 crc kubenswrapper[4821]: I1125 11:04:48.657986 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98d0e490-2704-4aa6-baa5-32ec701a5b68-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "98d0e490-2704-4aa6-baa5-32ec701a5b68" (UID: "98d0e490-2704-4aa6-baa5-32ec701a5b68"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:04:48 crc kubenswrapper[4821]: I1125 11:04:48.658031 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98d0e490-2704-4aa6-baa5-32ec701a5b68-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "98d0e490-2704-4aa6-baa5-32ec701a5b68" (UID: "98d0e490-2704-4aa6-baa5-32ec701a5b68"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:04:48 crc kubenswrapper[4821]: I1125 11:04:48.659243 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98d0e490-2704-4aa6-baa5-32ec701a5b68-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "98d0e490-2704-4aa6-baa5-32ec701a5b68" (UID: "98d0e490-2704-4aa6-baa5-32ec701a5b68"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:04:48 crc kubenswrapper[4821]: I1125 11:04:48.660340 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98d0e490-2704-4aa6-baa5-32ec701a5b68-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "98d0e490-2704-4aa6-baa5-32ec701a5b68" (UID: "98d0e490-2704-4aa6-baa5-32ec701a5b68"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:04:48 crc kubenswrapper[4821]: I1125 11:04:48.660712 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98d0e490-2704-4aa6-baa5-32ec701a5b68-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "98d0e490-2704-4aa6-baa5-32ec701a5b68" (UID: "98d0e490-2704-4aa6-baa5-32ec701a5b68"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:04:48 crc kubenswrapper[4821]: I1125 11:04:48.661110 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98d0e490-2704-4aa6-baa5-32ec701a5b68-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "98d0e490-2704-4aa6-baa5-32ec701a5b68" (UID: "98d0e490-2704-4aa6-baa5-32ec701a5b68"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:04:48 crc kubenswrapper[4821]: I1125 11:04:48.661387 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98d0e490-2704-4aa6-baa5-32ec701a5b68-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "98d0e490-2704-4aa6-baa5-32ec701a5b68" (UID: "98d0e490-2704-4aa6-baa5-32ec701a5b68"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:04:48 crc kubenswrapper[4821]: I1125 11:04:48.662593 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98d0e490-2704-4aa6-baa5-32ec701a5b68-kube-api-access-4x9dl" (OuterVolumeSpecName: "kube-api-access-4x9dl") pod "98d0e490-2704-4aa6-baa5-32ec701a5b68" (UID: "98d0e490-2704-4aa6-baa5-32ec701a5b68"). InnerVolumeSpecName "kube-api-access-4x9dl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:04:48 crc kubenswrapper[4821]: I1125 11:04:48.663326 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98d0e490-2704-4aa6-baa5-32ec701a5b68-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "98d0e490-2704-4aa6-baa5-32ec701a5b68" (UID: "98d0e490-2704-4aa6-baa5-32ec701a5b68"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:04:48 crc kubenswrapper[4821]: I1125 11:04:48.664234 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98d0e490-2704-4aa6-baa5-32ec701a5b68-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "98d0e490-2704-4aa6-baa5-32ec701a5b68" (UID: "98d0e490-2704-4aa6-baa5-32ec701a5b68"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:04:48 crc kubenswrapper[4821]: I1125 11:04:48.664920 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98d0e490-2704-4aa6-baa5-32ec701a5b68-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "98d0e490-2704-4aa6-baa5-32ec701a5b68" (UID: "98d0e490-2704-4aa6-baa5-32ec701a5b68"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:04:48 crc kubenswrapper[4821]: I1125 11:04:48.686848 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98d0e490-2704-4aa6-baa5-32ec701a5b68-inventory" (OuterVolumeSpecName: "inventory") pod "98d0e490-2704-4aa6-baa5-32ec701a5b68" (UID: "98d0e490-2704-4aa6-baa5-32ec701a5b68"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:04:48 crc kubenswrapper[4821]: I1125 11:04:48.692898 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98d0e490-2704-4aa6-baa5-32ec701a5b68-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "98d0e490-2704-4aa6-baa5-32ec701a5b68" (UID: "98d0e490-2704-4aa6-baa5-32ec701a5b68"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:04:48 crc kubenswrapper[4821]: I1125 11:04:48.754790 4821 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/98d0e490-2704-4aa6-baa5-32ec701a5b68-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 11:04:48 crc kubenswrapper[4821]: I1125 11:04:48.755077 4821 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98d0e490-2704-4aa6-baa5-32ec701a5b68-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:04:48 crc kubenswrapper[4821]: I1125 11:04:48.755188 4821 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/98d0e490-2704-4aa6-baa5-32ec701a5b68-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 11:04:48 crc kubenswrapper[4821]: I1125 11:04:48.755254 4821 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98d0e490-2704-4aa6-baa5-32ec701a5b68-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:04:48 crc kubenswrapper[4821]: I1125 11:04:48.755310 4821 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98d0e490-2704-4aa6-baa5-32ec701a5b68-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:04:48 crc kubenswrapper[4821]: I1125 11:04:48.755360 4821 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/98d0e490-2704-4aa6-baa5-32ec701a5b68-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 25 11:04:48 crc kubenswrapper[4821]: I1125 11:04:48.755412 4821 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/98d0e490-2704-4aa6-baa5-32ec701a5b68-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 25 11:04:48 crc kubenswrapper[4821]: I1125 11:04:48.755467 4821 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/98d0e490-2704-4aa6-baa5-32ec701a5b68-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 25 11:04:48 crc kubenswrapper[4821]: I1125 11:04:48.755530 4821 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98d0e490-2704-4aa6-baa5-32ec701a5b68-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:04:48 crc kubenswrapper[4821]: I1125 11:04:48.755582 4821 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/98d0e490-2704-4aa6-baa5-32ec701a5b68-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 25 11:04:48 crc kubenswrapper[4821]: I1125 11:04:48.755632 4821 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98d0e490-2704-4aa6-baa5-32ec701a5b68-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:04:48 crc kubenswrapper[4821]: I1125 11:04:48.755680 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4x9dl\" (UniqueName: \"kubernetes.io/projected/98d0e490-2704-4aa6-baa5-32ec701a5b68-kube-api-access-4x9dl\") on node \"crc\" DevicePath \"\"" Nov 25 11:04:48 crc kubenswrapper[4821]: I1125 11:04:48.755727 4821 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98d0e490-2704-4aa6-baa5-32ec701a5b68-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:04:48 crc kubenswrapper[4821]: I1125 11:04:48.755779 4821 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98d0e490-2704-4aa6-baa5-32ec701a5b68-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:04:49 crc kubenswrapper[4821]: I1125 11:04:49.116217 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-48x26" event={"ID":"98d0e490-2704-4aa6-baa5-32ec701a5b68","Type":"ContainerDied","Data":"6c9cbed16ee27a1d66974234738007e63d57d5a119d9d3ea19fc2db9313407d7"} Nov 25 11:04:49 crc kubenswrapper[4821]: I1125 11:04:49.116569 4821 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6c9cbed16ee27a1d66974234738007e63d57d5a119d9d3ea19fc2db9313407d7" Nov 25 11:04:49 crc kubenswrapper[4821]: I1125 11:04:49.116262 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-48x26" Nov 25 11:04:49 crc kubenswrapper[4821]: I1125 11:04:49.209872 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-9vb7k"] Nov 25 11:04:49 crc kubenswrapper[4821]: E1125 11:04:49.210455 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98d0e490-2704-4aa6-baa5-32ec701a5b68" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 25 11:04:49 crc kubenswrapper[4821]: I1125 11:04:49.210477 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="98d0e490-2704-4aa6-baa5-32ec701a5b68" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 25 11:04:49 crc kubenswrapper[4821]: I1125 11:04:49.210715 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="98d0e490-2704-4aa6-baa5-32ec701a5b68" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 25 11:04:49 crc kubenswrapper[4821]: I1125 11:04:49.211730 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9vb7k" Nov 25 11:04:49 crc kubenswrapper[4821]: I1125 11:04:49.217814 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-9vb7k"] Nov 25 11:04:49 crc kubenswrapper[4821]: I1125 11:04:49.217938 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 11:04:49 crc kubenswrapper[4821]: I1125 11:04:49.218002 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 11:04:49 crc kubenswrapper[4821]: I1125 11:04:49.218455 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 11:04:49 crc kubenswrapper[4821]: I1125 11:04:49.218933 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-z764x" Nov 25 11:04:49 crc kubenswrapper[4821]: I1125 11:04:49.219192 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Nov 25 11:04:49 crc kubenswrapper[4821]: I1125 11:04:49.263323 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/df834005-0241-4897-9112-05186899309e-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-9vb7k\" (UID: \"df834005-0241-4897-9112-05186899309e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9vb7k" Nov 25 11:04:49 crc kubenswrapper[4821]: I1125 11:04:49.263377 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/df834005-0241-4897-9112-05186899309e-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-9vb7k\" (UID: \"df834005-0241-4897-9112-05186899309e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9vb7k" Nov 25 11:04:49 crc kubenswrapper[4821]: I1125 11:04:49.263532 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/df834005-0241-4897-9112-05186899309e-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-9vb7k\" (UID: \"df834005-0241-4897-9112-05186899309e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9vb7k" Nov 25 11:04:49 crc kubenswrapper[4821]: I1125 11:04:49.263580 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hnxsr\" (UniqueName: \"kubernetes.io/projected/df834005-0241-4897-9112-05186899309e-kube-api-access-hnxsr\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-9vb7k\" (UID: \"df834005-0241-4897-9112-05186899309e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9vb7k" Nov 25 11:04:49 crc kubenswrapper[4821]: I1125 11:04:49.263661 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df834005-0241-4897-9112-05186899309e-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-9vb7k\" (UID: \"df834005-0241-4897-9112-05186899309e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9vb7k" Nov 25 11:04:49 crc kubenswrapper[4821]: I1125 11:04:49.366695 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/df834005-0241-4897-9112-05186899309e-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-9vb7k\" (UID: \"df834005-0241-4897-9112-05186899309e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9vb7k" Nov 25 11:04:49 crc kubenswrapper[4821]: I1125 11:04:49.366756 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hnxsr\" (UniqueName: \"kubernetes.io/projected/df834005-0241-4897-9112-05186899309e-kube-api-access-hnxsr\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-9vb7k\" (UID: \"df834005-0241-4897-9112-05186899309e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9vb7k" Nov 25 11:04:49 crc kubenswrapper[4821]: I1125 11:04:49.366829 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df834005-0241-4897-9112-05186899309e-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-9vb7k\" (UID: \"df834005-0241-4897-9112-05186899309e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9vb7k" Nov 25 11:04:49 crc kubenswrapper[4821]: I1125 11:04:49.366941 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/df834005-0241-4897-9112-05186899309e-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-9vb7k\" (UID: \"df834005-0241-4897-9112-05186899309e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9vb7k" Nov 25 11:04:49 crc kubenswrapper[4821]: I1125 11:04:49.366972 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/df834005-0241-4897-9112-05186899309e-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-9vb7k\" (UID: \"df834005-0241-4897-9112-05186899309e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9vb7k" Nov 25 11:04:49 crc kubenswrapper[4821]: I1125 11:04:49.369212 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/df834005-0241-4897-9112-05186899309e-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-9vb7k\" (UID: \"df834005-0241-4897-9112-05186899309e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9vb7k" Nov 25 11:04:49 crc kubenswrapper[4821]: I1125 11:04:49.371876 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/df834005-0241-4897-9112-05186899309e-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-9vb7k\" (UID: \"df834005-0241-4897-9112-05186899309e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9vb7k" Nov 25 11:04:49 crc kubenswrapper[4821]: I1125 11:04:49.372982 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df834005-0241-4897-9112-05186899309e-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-9vb7k\" (UID: \"df834005-0241-4897-9112-05186899309e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9vb7k" Nov 25 11:04:49 crc kubenswrapper[4821]: I1125 11:04:49.375975 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/df834005-0241-4897-9112-05186899309e-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-9vb7k\" (UID: \"df834005-0241-4897-9112-05186899309e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9vb7k" Nov 25 11:04:49 crc kubenswrapper[4821]: I1125 11:04:49.385150 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hnxsr\" (UniqueName: \"kubernetes.io/projected/df834005-0241-4897-9112-05186899309e-kube-api-access-hnxsr\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-9vb7k\" (UID: \"df834005-0241-4897-9112-05186899309e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9vb7k" Nov 25 11:04:49 crc kubenswrapper[4821]: I1125 11:04:49.534534 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9vb7k" Nov 25 11:04:50 crc kubenswrapper[4821]: I1125 11:04:50.064694 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-9vb7k"] Nov 25 11:04:50 crc kubenswrapper[4821]: W1125 11:04:50.066988 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddf834005_0241_4897_9112_05186899309e.slice/crio-d3f18019bc030b0767390ffa22c4233327d885a5aa89e85c607911a17ca78619 WatchSource:0}: Error finding container d3f18019bc030b0767390ffa22c4233327d885a5aa89e85c607911a17ca78619: Status 404 returned error can't find the container with id d3f18019bc030b0767390ffa22c4233327d885a5aa89e85c607911a17ca78619 Nov 25 11:04:50 crc kubenswrapper[4821]: I1125 11:04:50.127648 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9vb7k" event={"ID":"df834005-0241-4897-9112-05186899309e","Type":"ContainerStarted","Data":"d3f18019bc030b0767390ffa22c4233327d885a5aa89e85c607911a17ca78619"} Nov 25 11:04:51 crc kubenswrapper[4821]: I1125 11:04:51.140605 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9vb7k" event={"ID":"df834005-0241-4897-9112-05186899309e","Type":"ContainerStarted","Data":"a16630be2fab5e4606378fab72b3125ee5ed3034e6b7cb3f3aa5200d5f301578"} Nov 25 11:04:51 crc kubenswrapper[4821]: I1125 11:04:51.158443 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9vb7k" podStartSLOduration=1.375524373 podStartE2EDuration="2.158422045s" podCreationTimestamp="2025-11-25 11:04:49 +0000 UTC" firstStartedPulling="2025-11-25 11:04:50.069516902 +0000 UTC m=+1960.605836749" lastFinishedPulling="2025-11-25 11:04:50.852414574 +0000 UTC m=+1961.388734421" observedRunningTime="2025-11-25 11:04:51.155530594 +0000 UTC m=+1961.691850461" watchObservedRunningTime="2025-11-25 11:04:51.158422045 +0000 UTC m=+1961.694741892" Nov 25 11:05:56 crc kubenswrapper[4821]: I1125 11:05:56.686030 4821 generic.go:334] "Generic (PLEG): container finished" podID="df834005-0241-4897-9112-05186899309e" containerID="a16630be2fab5e4606378fab72b3125ee5ed3034e6b7cb3f3aa5200d5f301578" exitCode=0 Nov 25 11:05:56 crc kubenswrapper[4821]: I1125 11:05:56.686129 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9vb7k" event={"ID":"df834005-0241-4897-9112-05186899309e","Type":"ContainerDied","Data":"a16630be2fab5e4606378fab72b3125ee5ed3034e6b7cb3f3aa5200d5f301578"} Nov 25 11:05:58 crc kubenswrapper[4821]: I1125 11:05:58.122135 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9vb7k" Nov 25 11:05:58 crc kubenswrapper[4821]: I1125 11:05:58.212784 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hnxsr\" (UniqueName: \"kubernetes.io/projected/df834005-0241-4897-9112-05186899309e-kube-api-access-hnxsr\") pod \"df834005-0241-4897-9112-05186899309e\" (UID: \"df834005-0241-4897-9112-05186899309e\") " Nov 25 11:05:58 crc kubenswrapper[4821]: I1125 11:05:58.212911 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df834005-0241-4897-9112-05186899309e-ovn-combined-ca-bundle\") pod \"df834005-0241-4897-9112-05186899309e\" (UID: \"df834005-0241-4897-9112-05186899309e\") " Nov 25 11:05:58 crc kubenswrapper[4821]: I1125 11:05:58.213063 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/df834005-0241-4897-9112-05186899309e-ssh-key\") pod \"df834005-0241-4897-9112-05186899309e\" (UID: \"df834005-0241-4897-9112-05186899309e\") " Nov 25 11:05:58 crc kubenswrapper[4821]: I1125 11:05:58.213102 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/df834005-0241-4897-9112-05186899309e-ovncontroller-config-0\") pod \"df834005-0241-4897-9112-05186899309e\" (UID: \"df834005-0241-4897-9112-05186899309e\") " Nov 25 11:05:58 crc kubenswrapper[4821]: I1125 11:05:58.213154 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/df834005-0241-4897-9112-05186899309e-inventory\") pod \"df834005-0241-4897-9112-05186899309e\" (UID: \"df834005-0241-4897-9112-05186899309e\") " Nov 25 11:05:58 crc kubenswrapper[4821]: I1125 11:05:58.222236 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df834005-0241-4897-9112-05186899309e-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "df834005-0241-4897-9112-05186899309e" (UID: "df834005-0241-4897-9112-05186899309e"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:05:58 crc kubenswrapper[4821]: I1125 11:05:58.223088 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df834005-0241-4897-9112-05186899309e-kube-api-access-hnxsr" (OuterVolumeSpecName: "kube-api-access-hnxsr") pod "df834005-0241-4897-9112-05186899309e" (UID: "df834005-0241-4897-9112-05186899309e"). InnerVolumeSpecName "kube-api-access-hnxsr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:05:58 crc kubenswrapper[4821]: I1125 11:05:58.245747 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df834005-0241-4897-9112-05186899309e-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "df834005-0241-4897-9112-05186899309e" (UID: "df834005-0241-4897-9112-05186899309e"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:05:58 crc kubenswrapper[4821]: I1125 11:05:58.248586 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df834005-0241-4897-9112-05186899309e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "df834005-0241-4897-9112-05186899309e" (UID: "df834005-0241-4897-9112-05186899309e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:05:58 crc kubenswrapper[4821]: I1125 11:05:58.250041 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df834005-0241-4897-9112-05186899309e-inventory" (OuterVolumeSpecName: "inventory") pod "df834005-0241-4897-9112-05186899309e" (UID: "df834005-0241-4897-9112-05186899309e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:05:58 crc kubenswrapper[4821]: I1125 11:05:58.316095 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hnxsr\" (UniqueName: \"kubernetes.io/projected/df834005-0241-4897-9112-05186899309e-kube-api-access-hnxsr\") on node \"crc\" DevicePath \"\"" Nov 25 11:05:58 crc kubenswrapper[4821]: I1125 11:05:58.316148 4821 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df834005-0241-4897-9112-05186899309e-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:05:58 crc kubenswrapper[4821]: I1125 11:05:58.316184 4821 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/df834005-0241-4897-9112-05186899309e-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 11:05:58 crc kubenswrapper[4821]: I1125 11:05:58.316203 4821 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/df834005-0241-4897-9112-05186899309e-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Nov 25 11:05:58 crc kubenswrapper[4821]: I1125 11:05:58.316215 4821 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/df834005-0241-4897-9112-05186899309e-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 11:05:58 crc kubenswrapper[4821]: I1125 11:05:58.705891 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9vb7k" event={"ID":"df834005-0241-4897-9112-05186899309e","Type":"ContainerDied","Data":"d3f18019bc030b0767390ffa22c4233327d885a5aa89e85c607911a17ca78619"} Nov 25 11:05:58 crc kubenswrapper[4821]: I1125 11:05:58.705958 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-9vb7k" Nov 25 11:05:58 crc kubenswrapper[4821]: I1125 11:05:58.706690 4821 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d3f18019bc030b0767390ffa22c4233327d885a5aa89e85c607911a17ca78619" Nov 25 11:05:58 crc kubenswrapper[4821]: I1125 11:05:58.827939 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-225z5"] Nov 25 11:05:58 crc kubenswrapper[4821]: E1125 11:05:58.828558 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df834005-0241-4897-9112-05186899309e" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 25 11:05:58 crc kubenswrapper[4821]: I1125 11:05:58.828622 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="df834005-0241-4897-9112-05186899309e" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 25 11:05:58 crc kubenswrapper[4821]: I1125 11:05:58.828859 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="df834005-0241-4897-9112-05186899309e" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 25 11:05:58 crc kubenswrapper[4821]: I1125 11:05:58.829650 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-225z5" Nov 25 11:05:58 crc kubenswrapper[4821]: I1125 11:05:58.837701 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Nov 25 11:05:58 crc kubenswrapper[4821]: I1125 11:05:58.838886 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 11:05:58 crc kubenswrapper[4821]: I1125 11:05:58.839779 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 11:05:58 crc kubenswrapper[4821]: I1125 11:05:58.839867 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-z764x" Nov 25 11:05:58 crc kubenswrapper[4821]: I1125 11:05:58.839779 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 11:05:58 crc kubenswrapper[4821]: I1125 11:05:58.839924 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Nov 25 11:05:58 crc kubenswrapper[4821]: I1125 11:05:58.844940 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-225z5"] Nov 25 11:05:58 crc kubenswrapper[4821]: I1125 11:05:58.925248 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0f26a3b0-ba05-4484-8281-215ff2f09d25-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-225z5\" (UID: \"0f26a3b0-ba05-4484-8281-215ff2f09d25\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-225z5" Nov 25 11:05:58 crc kubenswrapper[4821]: I1125 11:05:58.925301 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0f26a3b0-ba05-4484-8281-215ff2f09d25-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-225z5\" (UID: \"0f26a3b0-ba05-4484-8281-215ff2f09d25\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-225z5" Nov 25 11:05:58 crc kubenswrapper[4821]: I1125 11:05:58.925326 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f26a3b0-ba05-4484-8281-215ff2f09d25-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-225z5\" (UID: \"0f26a3b0-ba05-4484-8281-215ff2f09d25\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-225z5" Nov 25 11:05:58 crc kubenswrapper[4821]: I1125 11:05:58.925403 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0f26a3b0-ba05-4484-8281-215ff2f09d25-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-225z5\" (UID: \"0f26a3b0-ba05-4484-8281-215ff2f09d25\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-225z5" Nov 25 11:05:58 crc kubenswrapper[4821]: I1125 11:05:58.925441 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dkbzv\" (UniqueName: \"kubernetes.io/projected/0f26a3b0-ba05-4484-8281-215ff2f09d25-kube-api-access-dkbzv\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-225z5\" (UID: \"0f26a3b0-ba05-4484-8281-215ff2f09d25\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-225z5" Nov 25 11:05:58 crc kubenswrapper[4821]: I1125 11:05:58.925477 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0f26a3b0-ba05-4484-8281-215ff2f09d25-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-225z5\" (UID: \"0f26a3b0-ba05-4484-8281-215ff2f09d25\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-225z5" Nov 25 11:05:59 crc kubenswrapper[4821]: I1125 11:05:59.039659 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0f26a3b0-ba05-4484-8281-215ff2f09d25-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-225z5\" (UID: \"0f26a3b0-ba05-4484-8281-215ff2f09d25\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-225z5" Nov 25 11:05:59 crc kubenswrapper[4821]: I1125 11:05:59.039782 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dkbzv\" (UniqueName: \"kubernetes.io/projected/0f26a3b0-ba05-4484-8281-215ff2f09d25-kube-api-access-dkbzv\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-225z5\" (UID: \"0f26a3b0-ba05-4484-8281-215ff2f09d25\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-225z5" Nov 25 11:05:59 crc kubenswrapper[4821]: I1125 11:05:59.039871 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0f26a3b0-ba05-4484-8281-215ff2f09d25-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-225z5\" (UID: \"0f26a3b0-ba05-4484-8281-215ff2f09d25\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-225z5" Nov 25 11:05:59 crc kubenswrapper[4821]: I1125 11:05:59.040152 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0f26a3b0-ba05-4484-8281-215ff2f09d25-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-225z5\" (UID: \"0f26a3b0-ba05-4484-8281-215ff2f09d25\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-225z5" Nov 25 11:05:59 crc kubenswrapper[4821]: I1125 11:05:59.040206 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0f26a3b0-ba05-4484-8281-215ff2f09d25-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-225z5\" (UID: \"0f26a3b0-ba05-4484-8281-215ff2f09d25\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-225z5" Nov 25 11:05:59 crc kubenswrapper[4821]: I1125 11:05:59.040248 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f26a3b0-ba05-4484-8281-215ff2f09d25-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-225z5\" (UID: \"0f26a3b0-ba05-4484-8281-215ff2f09d25\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-225z5" Nov 25 11:05:59 crc kubenswrapper[4821]: I1125 11:05:59.076868 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0f26a3b0-ba05-4484-8281-215ff2f09d25-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-225z5\" (UID: \"0f26a3b0-ba05-4484-8281-215ff2f09d25\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-225z5" Nov 25 11:05:59 crc kubenswrapper[4821]: I1125 11:05:59.077021 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0f26a3b0-ba05-4484-8281-215ff2f09d25-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-225z5\" (UID: \"0f26a3b0-ba05-4484-8281-215ff2f09d25\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-225z5" Nov 25 11:05:59 crc kubenswrapper[4821]: I1125 11:05:59.077376 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f26a3b0-ba05-4484-8281-215ff2f09d25-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-225z5\" (UID: \"0f26a3b0-ba05-4484-8281-215ff2f09d25\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-225z5" Nov 25 11:05:59 crc kubenswrapper[4821]: I1125 11:05:59.082771 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0f26a3b0-ba05-4484-8281-215ff2f09d25-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-225z5\" (UID: \"0f26a3b0-ba05-4484-8281-215ff2f09d25\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-225z5" Nov 25 11:05:59 crc kubenswrapper[4821]: I1125 11:05:59.086031 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0f26a3b0-ba05-4484-8281-215ff2f09d25-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-225z5\" (UID: \"0f26a3b0-ba05-4484-8281-215ff2f09d25\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-225z5" Nov 25 11:05:59 crc kubenswrapper[4821]: I1125 11:05:59.086094 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dkbzv\" (UniqueName: \"kubernetes.io/projected/0f26a3b0-ba05-4484-8281-215ff2f09d25-kube-api-access-dkbzv\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-225z5\" (UID: \"0f26a3b0-ba05-4484-8281-215ff2f09d25\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-225z5" Nov 25 11:05:59 crc kubenswrapper[4821]: I1125 11:05:59.147102 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-225z5" Nov 25 11:05:59 crc kubenswrapper[4821]: I1125 11:05:59.646590 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-225z5"] Nov 25 11:05:59 crc kubenswrapper[4821]: I1125 11:05:59.714922 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-225z5" event={"ID":"0f26a3b0-ba05-4484-8281-215ff2f09d25","Type":"ContainerStarted","Data":"0485185ee68a2c205b995023a38085f1183bff73fdfcaad9900cdbc3bb75aa27"} Nov 25 11:06:02 crc kubenswrapper[4821]: I1125 11:06:02.747303 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-225z5" event={"ID":"0f26a3b0-ba05-4484-8281-215ff2f09d25","Type":"ContainerStarted","Data":"5fcef16cebfc0b72caf417d5d6ebf112efb353a539e4159b33889916e2ff8ee8"} Nov 25 11:06:02 crc kubenswrapper[4821]: I1125 11:06:02.778295 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-225z5" podStartSLOduration=2.6301288449999998 podStartE2EDuration="4.778278481s" podCreationTimestamp="2025-11-25 11:05:58 +0000 UTC" firstStartedPulling="2025-11-25 11:05:59.650545835 +0000 UTC m=+2030.186865682" lastFinishedPulling="2025-11-25 11:06:01.798695471 +0000 UTC m=+2032.335015318" observedRunningTime="2025-11-25 11:06:02.777978372 +0000 UTC m=+2033.314298229" watchObservedRunningTime="2025-11-25 11:06:02.778278481 +0000 UTC m=+2033.314598328" Nov 25 11:06:10 crc kubenswrapper[4821]: I1125 11:06:10.573398 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-6fh7d"] Nov 25 11:06:10 crc kubenswrapper[4821]: I1125 11:06:10.575783 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6fh7d" Nov 25 11:06:10 crc kubenswrapper[4821]: I1125 11:06:10.585530 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6fh7d"] Nov 25 11:06:10 crc kubenswrapper[4821]: I1125 11:06:10.665306 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2a415f6-1bf6-48e8-9f5a-2f29431b2274-catalog-content\") pod \"redhat-operators-6fh7d\" (UID: \"e2a415f6-1bf6-48e8-9f5a-2f29431b2274\") " pod="openshift-marketplace/redhat-operators-6fh7d" Nov 25 11:06:10 crc kubenswrapper[4821]: I1125 11:06:10.665380 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hntgl\" (UniqueName: \"kubernetes.io/projected/e2a415f6-1bf6-48e8-9f5a-2f29431b2274-kube-api-access-hntgl\") pod \"redhat-operators-6fh7d\" (UID: \"e2a415f6-1bf6-48e8-9f5a-2f29431b2274\") " pod="openshift-marketplace/redhat-operators-6fh7d" Nov 25 11:06:10 crc kubenswrapper[4821]: I1125 11:06:10.665435 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2a415f6-1bf6-48e8-9f5a-2f29431b2274-utilities\") pod \"redhat-operators-6fh7d\" (UID: \"e2a415f6-1bf6-48e8-9f5a-2f29431b2274\") " pod="openshift-marketplace/redhat-operators-6fh7d" Nov 25 11:06:10 crc kubenswrapper[4821]: I1125 11:06:10.767195 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hntgl\" (UniqueName: \"kubernetes.io/projected/e2a415f6-1bf6-48e8-9f5a-2f29431b2274-kube-api-access-hntgl\") pod \"redhat-operators-6fh7d\" (UID: \"e2a415f6-1bf6-48e8-9f5a-2f29431b2274\") " pod="openshift-marketplace/redhat-operators-6fh7d" Nov 25 11:06:10 crc kubenswrapper[4821]: I1125 11:06:10.767290 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2a415f6-1bf6-48e8-9f5a-2f29431b2274-utilities\") pod \"redhat-operators-6fh7d\" (UID: \"e2a415f6-1bf6-48e8-9f5a-2f29431b2274\") " pod="openshift-marketplace/redhat-operators-6fh7d" Nov 25 11:06:10 crc kubenswrapper[4821]: I1125 11:06:10.767507 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2a415f6-1bf6-48e8-9f5a-2f29431b2274-catalog-content\") pod \"redhat-operators-6fh7d\" (UID: \"e2a415f6-1bf6-48e8-9f5a-2f29431b2274\") " pod="openshift-marketplace/redhat-operators-6fh7d" Nov 25 11:06:10 crc kubenswrapper[4821]: I1125 11:06:10.767958 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2a415f6-1bf6-48e8-9f5a-2f29431b2274-utilities\") pod \"redhat-operators-6fh7d\" (UID: \"e2a415f6-1bf6-48e8-9f5a-2f29431b2274\") " pod="openshift-marketplace/redhat-operators-6fh7d" Nov 25 11:06:10 crc kubenswrapper[4821]: I1125 11:06:10.768121 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2a415f6-1bf6-48e8-9f5a-2f29431b2274-catalog-content\") pod \"redhat-operators-6fh7d\" (UID: \"e2a415f6-1bf6-48e8-9f5a-2f29431b2274\") " pod="openshift-marketplace/redhat-operators-6fh7d" Nov 25 11:06:10 crc kubenswrapper[4821]: I1125 11:06:10.786755 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hntgl\" (UniqueName: \"kubernetes.io/projected/e2a415f6-1bf6-48e8-9f5a-2f29431b2274-kube-api-access-hntgl\") pod \"redhat-operators-6fh7d\" (UID: \"e2a415f6-1bf6-48e8-9f5a-2f29431b2274\") " pod="openshift-marketplace/redhat-operators-6fh7d" Nov 25 11:06:10 crc kubenswrapper[4821]: I1125 11:06:10.897927 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6fh7d" Nov 25 11:06:11 crc kubenswrapper[4821]: I1125 11:06:11.367748 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6fh7d"] Nov 25 11:06:11 crc kubenswrapper[4821]: I1125 11:06:11.824521 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6fh7d" event={"ID":"e2a415f6-1bf6-48e8-9f5a-2f29431b2274","Type":"ContainerStarted","Data":"db16d46ec1b084955c174837ebbd4d41d388f451e6a5fea6e4eeca2f54831235"} Nov 25 11:06:12 crc kubenswrapper[4821]: I1125 11:06:12.838272 4821 generic.go:334] "Generic (PLEG): container finished" podID="e2a415f6-1bf6-48e8-9f5a-2f29431b2274" containerID="f0e10f151d2616ac0bb61af4847441ce1236ff7df6e219b881cf48b1ab4e5675" exitCode=0 Nov 25 11:06:12 crc kubenswrapper[4821]: I1125 11:06:12.838579 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6fh7d" event={"ID":"e2a415f6-1bf6-48e8-9f5a-2f29431b2274","Type":"ContainerDied","Data":"f0e10f151d2616ac0bb61af4847441ce1236ff7df6e219b881cf48b1ab4e5675"} Nov 25 11:06:14 crc kubenswrapper[4821]: I1125 11:06:14.856412 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6fh7d" event={"ID":"e2a415f6-1bf6-48e8-9f5a-2f29431b2274","Type":"ContainerStarted","Data":"ac02bd48ca1bf10f5c435751b3830d72369171b62ad7918bbb3f0f9a4a5749a7"} Nov 25 11:06:15 crc kubenswrapper[4821]: I1125 11:06:15.868412 4821 generic.go:334] "Generic (PLEG): container finished" podID="e2a415f6-1bf6-48e8-9f5a-2f29431b2274" containerID="ac02bd48ca1bf10f5c435751b3830d72369171b62ad7918bbb3f0f9a4a5749a7" exitCode=0 Nov 25 11:06:15 crc kubenswrapper[4821]: I1125 11:06:15.868499 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6fh7d" event={"ID":"e2a415f6-1bf6-48e8-9f5a-2f29431b2274","Type":"ContainerDied","Data":"ac02bd48ca1bf10f5c435751b3830d72369171b62ad7918bbb3f0f9a4a5749a7"} Nov 25 11:06:16 crc kubenswrapper[4821]: I1125 11:06:16.902129 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6fh7d" event={"ID":"e2a415f6-1bf6-48e8-9f5a-2f29431b2274","Type":"ContainerStarted","Data":"f1687930e3a59df28ca041fb80c722279d8d946b1242390d55220dc50f27583f"} Nov 25 11:06:16 crc kubenswrapper[4821]: I1125 11:06:16.933730 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-6fh7d" podStartSLOduration=3.277250681 podStartE2EDuration="6.933707115s" podCreationTimestamp="2025-11-25 11:06:10 +0000 UTC" firstStartedPulling="2025-11-25 11:06:12.840534579 +0000 UTC m=+2043.376854466" lastFinishedPulling="2025-11-25 11:06:16.496991053 +0000 UTC m=+2047.033310900" observedRunningTime="2025-11-25 11:06:16.919971627 +0000 UTC m=+2047.456291494" watchObservedRunningTime="2025-11-25 11:06:16.933707115 +0000 UTC m=+2047.470026962" Nov 25 11:06:20 crc kubenswrapper[4821]: I1125 11:06:20.898547 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-6fh7d" Nov 25 11:06:20 crc kubenswrapper[4821]: I1125 11:06:20.899043 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-6fh7d" Nov 25 11:06:22 crc kubenswrapper[4821]: I1125 11:06:22.100752 4821 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-6fh7d" podUID="e2a415f6-1bf6-48e8-9f5a-2f29431b2274" containerName="registry-server" probeResult="failure" output=< Nov 25 11:06:22 crc kubenswrapper[4821]: timeout: failed to connect service ":50051" within 1s Nov 25 11:06:22 crc kubenswrapper[4821]: > Nov 25 11:06:24 crc kubenswrapper[4821]: I1125 11:06:24.885922 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-7zq8k"] Nov 25 11:06:24 crc kubenswrapper[4821]: I1125 11:06:24.893052 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7zq8k" Nov 25 11:06:24 crc kubenswrapper[4821]: I1125 11:06:24.901633 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7zq8k"] Nov 25 11:06:25 crc kubenswrapper[4821]: I1125 11:06:25.079333 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/284b03e4-ca38-42e2-92b8-11c812a2b751-utilities\") pod \"redhat-marketplace-7zq8k\" (UID: \"284b03e4-ca38-42e2-92b8-11c812a2b751\") " pod="openshift-marketplace/redhat-marketplace-7zq8k" Nov 25 11:06:25 crc kubenswrapper[4821]: I1125 11:06:25.079634 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/284b03e4-ca38-42e2-92b8-11c812a2b751-catalog-content\") pod \"redhat-marketplace-7zq8k\" (UID: \"284b03e4-ca38-42e2-92b8-11c812a2b751\") " pod="openshift-marketplace/redhat-marketplace-7zq8k" Nov 25 11:06:25 crc kubenswrapper[4821]: I1125 11:06:25.079748 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vf9z\" (UniqueName: \"kubernetes.io/projected/284b03e4-ca38-42e2-92b8-11c812a2b751-kube-api-access-7vf9z\") pod \"redhat-marketplace-7zq8k\" (UID: \"284b03e4-ca38-42e2-92b8-11c812a2b751\") " pod="openshift-marketplace/redhat-marketplace-7zq8k" Nov 25 11:06:25 crc kubenswrapper[4821]: I1125 11:06:25.181998 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/284b03e4-ca38-42e2-92b8-11c812a2b751-utilities\") pod \"redhat-marketplace-7zq8k\" (UID: \"284b03e4-ca38-42e2-92b8-11c812a2b751\") " pod="openshift-marketplace/redhat-marketplace-7zq8k" Nov 25 11:06:25 crc kubenswrapper[4821]: I1125 11:06:25.182417 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/284b03e4-ca38-42e2-92b8-11c812a2b751-catalog-content\") pod \"redhat-marketplace-7zq8k\" (UID: \"284b03e4-ca38-42e2-92b8-11c812a2b751\") " pod="openshift-marketplace/redhat-marketplace-7zq8k" Nov 25 11:06:25 crc kubenswrapper[4821]: I1125 11:06:25.182455 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vf9z\" (UniqueName: \"kubernetes.io/projected/284b03e4-ca38-42e2-92b8-11c812a2b751-kube-api-access-7vf9z\") pod \"redhat-marketplace-7zq8k\" (UID: \"284b03e4-ca38-42e2-92b8-11c812a2b751\") " pod="openshift-marketplace/redhat-marketplace-7zq8k" Nov 25 11:06:25 crc kubenswrapper[4821]: I1125 11:06:25.183913 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/284b03e4-ca38-42e2-92b8-11c812a2b751-utilities\") pod \"redhat-marketplace-7zq8k\" (UID: \"284b03e4-ca38-42e2-92b8-11c812a2b751\") " pod="openshift-marketplace/redhat-marketplace-7zq8k" Nov 25 11:06:25 crc kubenswrapper[4821]: I1125 11:06:25.183965 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/284b03e4-ca38-42e2-92b8-11c812a2b751-catalog-content\") pod \"redhat-marketplace-7zq8k\" (UID: \"284b03e4-ca38-42e2-92b8-11c812a2b751\") " pod="openshift-marketplace/redhat-marketplace-7zq8k" Nov 25 11:06:25 crc kubenswrapper[4821]: I1125 11:06:25.202685 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vf9z\" (UniqueName: \"kubernetes.io/projected/284b03e4-ca38-42e2-92b8-11c812a2b751-kube-api-access-7vf9z\") pod \"redhat-marketplace-7zq8k\" (UID: \"284b03e4-ca38-42e2-92b8-11c812a2b751\") " pod="openshift-marketplace/redhat-marketplace-7zq8k" Nov 25 11:06:25 crc kubenswrapper[4821]: I1125 11:06:25.216663 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7zq8k" Nov 25 11:06:25 crc kubenswrapper[4821]: I1125 11:06:25.696277 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7zq8k"] Nov 25 11:06:25 crc kubenswrapper[4821]: I1125 11:06:25.980490 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7zq8k" event={"ID":"284b03e4-ca38-42e2-92b8-11c812a2b751","Type":"ContainerStarted","Data":"37130fdfbab6f1026a1f81511c42fe6898da29727aed1fb1f2072bd19e16225a"} Nov 25 11:06:26 crc kubenswrapper[4821]: I1125 11:06:26.996139 4821 generic.go:334] "Generic (PLEG): container finished" podID="284b03e4-ca38-42e2-92b8-11c812a2b751" containerID="b12c4cd3bc71781956276803db63581307ff18b705c34d0834d23f7568d3ace9" exitCode=0 Nov 25 11:06:26 crc kubenswrapper[4821]: I1125 11:06:26.996434 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7zq8k" event={"ID":"284b03e4-ca38-42e2-92b8-11c812a2b751","Type":"ContainerDied","Data":"b12c4cd3bc71781956276803db63581307ff18b705c34d0834d23f7568d3ace9"} Nov 25 11:06:29 crc kubenswrapper[4821]: I1125 11:06:29.019861 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7zq8k" event={"ID":"284b03e4-ca38-42e2-92b8-11c812a2b751","Type":"ContainerStarted","Data":"7d54739a63f6374f5b63ea1112d62583b5251a5bf0be279a55fb937f73e6c3c5"} Nov 25 11:06:30 crc kubenswrapper[4821]: I1125 11:06:30.031304 4821 generic.go:334] "Generic (PLEG): container finished" podID="284b03e4-ca38-42e2-92b8-11c812a2b751" containerID="7d54739a63f6374f5b63ea1112d62583b5251a5bf0be279a55fb937f73e6c3c5" exitCode=0 Nov 25 11:06:30 crc kubenswrapper[4821]: I1125 11:06:30.031381 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7zq8k" event={"ID":"284b03e4-ca38-42e2-92b8-11c812a2b751","Type":"ContainerDied","Data":"7d54739a63f6374f5b63ea1112d62583b5251a5bf0be279a55fb937f73e6c3c5"} Nov 25 11:06:31 crc kubenswrapper[4821]: I1125 11:06:31.940542 4821 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-6fh7d" podUID="e2a415f6-1bf6-48e8-9f5a-2f29431b2274" containerName="registry-server" probeResult="failure" output=< Nov 25 11:06:31 crc kubenswrapper[4821]: timeout: failed to connect service ":50051" within 1s Nov 25 11:06:31 crc kubenswrapper[4821]: > Nov 25 11:06:33 crc kubenswrapper[4821]: I1125 11:06:33.057069 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7zq8k" event={"ID":"284b03e4-ca38-42e2-92b8-11c812a2b751","Type":"ContainerStarted","Data":"939dd19166009fadfab2d3192e559cb4aa6aed788c04d3791a84dcbce3ae1a87"} Nov 25 11:06:33 crc kubenswrapper[4821]: I1125 11:06:33.079455 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-7zq8k" podStartSLOduration=3.83780457 podStartE2EDuration="9.079441127s" podCreationTimestamp="2025-11-25 11:06:24 +0000 UTC" firstStartedPulling="2025-11-25 11:06:26.999690697 +0000 UTC m=+2057.536010574" lastFinishedPulling="2025-11-25 11:06:32.241327284 +0000 UTC m=+2062.777647131" observedRunningTime="2025-11-25 11:06:33.07550988 +0000 UTC m=+2063.611829747" watchObservedRunningTime="2025-11-25 11:06:33.079441127 +0000 UTC m=+2063.615760974" Nov 25 11:06:34 crc kubenswrapper[4821]: I1125 11:06:34.714861 4821 patch_prober.go:28] interesting pod/machine-config-daemon-2krbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:06:34 crc kubenswrapper[4821]: I1125 11:06:34.715240 4821 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:06:35 crc kubenswrapper[4821]: I1125 11:06:35.217584 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-7zq8k" Nov 25 11:06:35 crc kubenswrapper[4821]: I1125 11:06:35.217632 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-7zq8k" Nov 25 11:06:35 crc kubenswrapper[4821]: I1125 11:06:35.262277 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-7zq8k" Nov 25 11:06:41 crc kubenswrapper[4821]: I1125 11:06:41.952805 4821 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-6fh7d" podUID="e2a415f6-1bf6-48e8-9f5a-2f29431b2274" containerName="registry-server" probeResult="failure" output=< Nov 25 11:06:41 crc kubenswrapper[4821]: timeout: failed to connect service ":50051" within 1s Nov 25 11:06:41 crc kubenswrapper[4821]: > Nov 25 11:06:45 crc kubenswrapper[4821]: I1125 11:06:45.264719 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-7zq8k" Nov 25 11:06:45 crc kubenswrapper[4821]: I1125 11:06:45.313081 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7zq8k"] Nov 25 11:06:46 crc kubenswrapper[4821]: I1125 11:06:46.173504 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-7zq8k" podUID="284b03e4-ca38-42e2-92b8-11c812a2b751" containerName="registry-server" containerID="cri-o://939dd19166009fadfab2d3192e559cb4aa6aed788c04d3791a84dcbce3ae1a87" gracePeriod=2 Nov 25 11:06:47 crc kubenswrapper[4821]: I1125 11:06:47.114891 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7zq8k" Nov 25 11:06:47 crc kubenswrapper[4821]: I1125 11:06:47.183518 4821 generic.go:334] "Generic (PLEG): container finished" podID="284b03e4-ca38-42e2-92b8-11c812a2b751" containerID="939dd19166009fadfab2d3192e559cb4aa6aed788c04d3791a84dcbce3ae1a87" exitCode=0 Nov 25 11:06:47 crc kubenswrapper[4821]: I1125 11:06:47.183570 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7zq8k" Nov 25 11:06:47 crc kubenswrapper[4821]: I1125 11:06:47.183576 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7zq8k" event={"ID":"284b03e4-ca38-42e2-92b8-11c812a2b751","Type":"ContainerDied","Data":"939dd19166009fadfab2d3192e559cb4aa6aed788c04d3791a84dcbce3ae1a87"} Nov 25 11:06:47 crc kubenswrapper[4821]: I1125 11:06:47.183676 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7zq8k" event={"ID":"284b03e4-ca38-42e2-92b8-11c812a2b751","Type":"ContainerDied","Data":"37130fdfbab6f1026a1f81511c42fe6898da29727aed1fb1f2072bd19e16225a"} Nov 25 11:06:47 crc kubenswrapper[4821]: I1125 11:06:47.183698 4821 scope.go:117] "RemoveContainer" containerID="939dd19166009fadfab2d3192e559cb4aa6aed788c04d3791a84dcbce3ae1a87" Nov 25 11:06:47 crc kubenswrapper[4821]: I1125 11:06:47.202256 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7vf9z\" (UniqueName: \"kubernetes.io/projected/284b03e4-ca38-42e2-92b8-11c812a2b751-kube-api-access-7vf9z\") pod \"284b03e4-ca38-42e2-92b8-11c812a2b751\" (UID: \"284b03e4-ca38-42e2-92b8-11c812a2b751\") " Nov 25 11:06:47 crc kubenswrapper[4821]: I1125 11:06:47.202403 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/284b03e4-ca38-42e2-92b8-11c812a2b751-utilities\") pod \"284b03e4-ca38-42e2-92b8-11c812a2b751\" (UID: \"284b03e4-ca38-42e2-92b8-11c812a2b751\") " Nov 25 11:06:47 crc kubenswrapper[4821]: I1125 11:06:47.202432 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/284b03e4-ca38-42e2-92b8-11c812a2b751-catalog-content\") pod \"284b03e4-ca38-42e2-92b8-11c812a2b751\" (UID: \"284b03e4-ca38-42e2-92b8-11c812a2b751\") " Nov 25 11:06:47 crc kubenswrapper[4821]: I1125 11:06:47.203549 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/284b03e4-ca38-42e2-92b8-11c812a2b751-utilities" (OuterVolumeSpecName: "utilities") pod "284b03e4-ca38-42e2-92b8-11c812a2b751" (UID: "284b03e4-ca38-42e2-92b8-11c812a2b751"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:06:47 crc kubenswrapper[4821]: I1125 11:06:47.206584 4821 scope.go:117] "RemoveContainer" containerID="7d54739a63f6374f5b63ea1112d62583b5251a5bf0be279a55fb937f73e6c3c5" Nov 25 11:06:47 crc kubenswrapper[4821]: I1125 11:06:47.208815 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/284b03e4-ca38-42e2-92b8-11c812a2b751-kube-api-access-7vf9z" (OuterVolumeSpecName: "kube-api-access-7vf9z") pod "284b03e4-ca38-42e2-92b8-11c812a2b751" (UID: "284b03e4-ca38-42e2-92b8-11c812a2b751"). InnerVolumeSpecName "kube-api-access-7vf9z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:06:47 crc kubenswrapper[4821]: I1125 11:06:47.219150 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/284b03e4-ca38-42e2-92b8-11c812a2b751-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "284b03e4-ca38-42e2-92b8-11c812a2b751" (UID: "284b03e4-ca38-42e2-92b8-11c812a2b751"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:06:47 crc kubenswrapper[4821]: I1125 11:06:47.263720 4821 scope.go:117] "RemoveContainer" containerID="b12c4cd3bc71781956276803db63581307ff18b705c34d0834d23f7568d3ace9" Nov 25 11:06:47 crc kubenswrapper[4821]: I1125 11:06:47.304372 4821 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/284b03e4-ca38-42e2-92b8-11c812a2b751-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 11:06:47 crc kubenswrapper[4821]: I1125 11:06:47.304408 4821 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/284b03e4-ca38-42e2-92b8-11c812a2b751-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 11:06:47 crc kubenswrapper[4821]: I1125 11:06:47.304422 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7vf9z\" (UniqueName: \"kubernetes.io/projected/284b03e4-ca38-42e2-92b8-11c812a2b751-kube-api-access-7vf9z\") on node \"crc\" DevicePath \"\"" Nov 25 11:06:47 crc kubenswrapper[4821]: I1125 11:06:47.317367 4821 scope.go:117] "RemoveContainer" containerID="939dd19166009fadfab2d3192e559cb4aa6aed788c04d3791a84dcbce3ae1a87" Nov 25 11:06:47 crc kubenswrapper[4821]: E1125 11:06:47.317892 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"939dd19166009fadfab2d3192e559cb4aa6aed788c04d3791a84dcbce3ae1a87\": container with ID starting with 939dd19166009fadfab2d3192e559cb4aa6aed788c04d3791a84dcbce3ae1a87 not found: ID does not exist" containerID="939dd19166009fadfab2d3192e559cb4aa6aed788c04d3791a84dcbce3ae1a87" Nov 25 11:06:47 crc kubenswrapper[4821]: I1125 11:06:47.317950 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"939dd19166009fadfab2d3192e559cb4aa6aed788c04d3791a84dcbce3ae1a87"} err="failed to get container status \"939dd19166009fadfab2d3192e559cb4aa6aed788c04d3791a84dcbce3ae1a87\": rpc error: code = NotFound desc = could not find container \"939dd19166009fadfab2d3192e559cb4aa6aed788c04d3791a84dcbce3ae1a87\": container with ID starting with 939dd19166009fadfab2d3192e559cb4aa6aed788c04d3791a84dcbce3ae1a87 not found: ID does not exist" Nov 25 11:06:47 crc kubenswrapper[4821]: I1125 11:06:47.318005 4821 scope.go:117] "RemoveContainer" containerID="7d54739a63f6374f5b63ea1112d62583b5251a5bf0be279a55fb937f73e6c3c5" Nov 25 11:06:47 crc kubenswrapper[4821]: E1125 11:06:47.318591 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d54739a63f6374f5b63ea1112d62583b5251a5bf0be279a55fb937f73e6c3c5\": container with ID starting with 7d54739a63f6374f5b63ea1112d62583b5251a5bf0be279a55fb937f73e6c3c5 not found: ID does not exist" containerID="7d54739a63f6374f5b63ea1112d62583b5251a5bf0be279a55fb937f73e6c3c5" Nov 25 11:06:47 crc kubenswrapper[4821]: I1125 11:06:47.318626 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d54739a63f6374f5b63ea1112d62583b5251a5bf0be279a55fb937f73e6c3c5"} err="failed to get container status \"7d54739a63f6374f5b63ea1112d62583b5251a5bf0be279a55fb937f73e6c3c5\": rpc error: code = NotFound desc = could not find container \"7d54739a63f6374f5b63ea1112d62583b5251a5bf0be279a55fb937f73e6c3c5\": container with ID starting with 7d54739a63f6374f5b63ea1112d62583b5251a5bf0be279a55fb937f73e6c3c5 not found: ID does not exist" Nov 25 11:06:47 crc kubenswrapper[4821]: I1125 11:06:47.318651 4821 scope.go:117] "RemoveContainer" containerID="b12c4cd3bc71781956276803db63581307ff18b705c34d0834d23f7568d3ace9" Nov 25 11:06:47 crc kubenswrapper[4821]: E1125 11:06:47.318954 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b12c4cd3bc71781956276803db63581307ff18b705c34d0834d23f7568d3ace9\": container with ID starting with b12c4cd3bc71781956276803db63581307ff18b705c34d0834d23f7568d3ace9 not found: ID does not exist" containerID="b12c4cd3bc71781956276803db63581307ff18b705c34d0834d23f7568d3ace9" Nov 25 11:06:47 crc kubenswrapper[4821]: I1125 11:06:47.318980 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b12c4cd3bc71781956276803db63581307ff18b705c34d0834d23f7568d3ace9"} err="failed to get container status \"b12c4cd3bc71781956276803db63581307ff18b705c34d0834d23f7568d3ace9\": rpc error: code = NotFound desc = could not find container \"b12c4cd3bc71781956276803db63581307ff18b705c34d0834d23f7568d3ace9\": container with ID starting with b12c4cd3bc71781956276803db63581307ff18b705c34d0834d23f7568d3ace9 not found: ID does not exist" Nov 25 11:06:47 crc kubenswrapper[4821]: I1125 11:06:47.513021 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7zq8k"] Nov 25 11:06:47 crc kubenswrapper[4821]: I1125 11:06:47.519591 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-7zq8k"] Nov 25 11:06:48 crc kubenswrapper[4821]: I1125 11:06:48.125675 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="284b03e4-ca38-42e2-92b8-11c812a2b751" path="/var/lib/kubelet/pods/284b03e4-ca38-42e2-92b8-11c812a2b751/volumes" Nov 25 11:06:50 crc kubenswrapper[4821]: I1125 11:06:50.942610 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-6fh7d" Nov 25 11:06:51 crc kubenswrapper[4821]: I1125 11:06:51.012865 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-6fh7d" Nov 25 11:06:51 crc kubenswrapper[4821]: I1125 11:06:51.175940 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6fh7d"] Nov 25 11:06:51 crc kubenswrapper[4821]: I1125 11:06:51.234696 4821 generic.go:334] "Generic (PLEG): container finished" podID="0f26a3b0-ba05-4484-8281-215ff2f09d25" containerID="5fcef16cebfc0b72caf417d5d6ebf112efb353a539e4159b33889916e2ff8ee8" exitCode=0 Nov 25 11:06:51 crc kubenswrapper[4821]: I1125 11:06:51.234784 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-225z5" event={"ID":"0f26a3b0-ba05-4484-8281-215ff2f09d25","Type":"ContainerDied","Data":"5fcef16cebfc0b72caf417d5d6ebf112efb353a539e4159b33889916e2ff8ee8"} Nov 25 11:06:52 crc kubenswrapper[4821]: I1125 11:06:52.244819 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-6fh7d" podUID="e2a415f6-1bf6-48e8-9f5a-2f29431b2274" containerName="registry-server" containerID="cri-o://f1687930e3a59df28ca041fb80c722279d8d946b1242390d55220dc50f27583f" gracePeriod=2 Nov 25 11:06:52 crc kubenswrapper[4821]: I1125 11:06:52.763039 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-225z5" Nov 25 11:06:52 crc kubenswrapper[4821]: I1125 11:06:52.772973 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6fh7d" Nov 25 11:06:52 crc kubenswrapper[4821]: I1125 11:06:52.829508 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0f26a3b0-ba05-4484-8281-215ff2f09d25-neutron-ovn-metadata-agent-neutron-config-0\") pod \"0f26a3b0-ba05-4484-8281-215ff2f09d25\" (UID: \"0f26a3b0-ba05-4484-8281-215ff2f09d25\") " Nov 25 11:06:52 crc kubenswrapper[4821]: I1125 11:06:52.829882 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0f26a3b0-ba05-4484-8281-215ff2f09d25-ssh-key\") pod \"0f26a3b0-ba05-4484-8281-215ff2f09d25\" (UID: \"0f26a3b0-ba05-4484-8281-215ff2f09d25\") " Nov 25 11:06:52 crc kubenswrapper[4821]: I1125 11:06:52.829917 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2a415f6-1bf6-48e8-9f5a-2f29431b2274-catalog-content\") pod \"e2a415f6-1bf6-48e8-9f5a-2f29431b2274\" (UID: \"e2a415f6-1bf6-48e8-9f5a-2f29431b2274\") " Nov 25 11:06:52 crc kubenswrapper[4821]: I1125 11:06:52.829947 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dkbzv\" (UniqueName: \"kubernetes.io/projected/0f26a3b0-ba05-4484-8281-215ff2f09d25-kube-api-access-dkbzv\") pod \"0f26a3b0-ba05-4484-8281-215ff2f09d25\" (UID: \"0f26a3b0-ba05-4484-8281-215ff2f09d25\") " Nov 25 11:06:52 crc kubenswrapper[4821]: I1125 11:06:52.829998 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f26a3b0-ba05-4484-8281-215ff2f09d25-neutron-metadata-combined-ca-bundle\") pod \"0f26a3b0-ba05-4484-8281-215ff2f09d25\" (UID: \"0f26a3b0-ba05-4484-8281-215ff2f09d25\") " Nov 25 11:06:52 crc kubenswrapper[4821]: I1125 11:06:52.830049 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hntgl\" (UniqueName: \"kubernetes.io/projected/e2a415f6-1bf6-48e8-9f5a-2f29431b2274-kube-api-access-hntgl\") pod \"e2a415f6-1bf6-48e8-9f5a-2f29431b2274\" (UID: \"e2a415f6-1bf6-48e8-9f5a-2f29431b2274\") " Nov 25 11:06:52 crc kubenswrapper[4821]: I1125 11:06:52.830082 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0f26a3b0-ba05-4484-8281-215ff2f09d25-nova-metadata-neutron-config-0\") pod \"0f26a3b0-ba05-4484-8281-215ff2f09d25\" (UID: \"0f26a3b0-ba05-4484-8281-215ff2f09d25\") " Nov 25 11:06:52 crc kubenswrapper[4821]: I1125 11:06:52.830100 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0f26a3b0-ba05-4484-8281-215ff2f09d25-inventory\") pod \"0f26a3b0-ba05-4484-8281-215ff2f09d25\" (UID: \"0f26a3b0-ba05-4484-8281-215ff2f09d25\") " Nov 25 11:06:52 crc kubenswrapper[4821]: I1125 11:06:52.830125 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2a415f6-1bf6-48e8-9f5a-2f29431b2274-utilities\") pod \"e2a415f6-1bf6-48e8-9f5a-2f29431b2274\" (UID: \"e2a415f6-1bf6-48e8-9f5a-2f29431b2274\") " Nov 25 11:06:52 crc kubenswrapper[4821]: I1125 11:06:52.831486 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e2a415f6-1bf6-48e8-9f5a-2f29431b2274-utilities" (OuterVolumeSpecName: "utilities") pod "e2a415f6-1bf6-48e8-9f5a-2f29431b2274" (UID: "e2a415f6-1bf6-48e8-9f5a-2f29431b2274"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:06:52 crc kubenswrapper[4821]: I1125 11:06:52.836249 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f26a3b0-ba05-4484-8281-215ff2f09d25-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "0f26a3b0-ba05-4484-8281-215ff2f09d25" (UID: "0f26a3b0-ba05-4484-8281-215ff2f09d25"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:06:52 crc kubenswrapper[4821]: I1125 11:06:52.845575 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f26a3b0-ba05-4484-8281-215ff2f09d25-kube-api-access-dkbzv" (OuterVolumeSpecName: "kube-api-access-dkbzv") pod "0f26a3b0-ba05-4484-8281-215ff2f09d25" (UID: "0f26a3b0-ba05-4484-8281-215ff2f09d25"). InnerVolumeSpecName "kube-api-access-dkbzv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:06:52 crc kubenswrapper[4821]: I1125 11:06:52.850495 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2a415f6-1bf6-48e8-9f5a-2f29431b2274-kube-api-access-hntgl" (OuterVolumeSpecName: "kube-api-access-hntgl") pod "e2a415f6-1bf6-48e8-9f5a-2f29431b2274" (UID: "e2a415f6-1bf6-48e8-9f5a-2f29431b2274"). InnerVolumeSpecName "kube-api-access-hntgl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:06:52 crc kubenswrapper[4821]: I1125 11:06:52.861447 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f26a3b0-ba05-4484-8281-215ff2f09d25-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "0f26a3b0-ba05-4484-8281-215ff2f09d25" (UID: "0f26a3b0-ba05-4484-8281-215ff2f09d25"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:06:52 crc kubenswrapper[4821]: I1125 11:06:52.863999 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f26a3b0-ba05-4484-8281-215ff2f09d25-inventory" (OuterVolumeSpecName: "inventory") pod "0f26a3b0-ba05-4484-8281-215ff2f09d25" (UID: "0f26a3b0-ba05-4484-8281-215ff2f09d25"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:06:52 crc kubenswrapper[4821]: I1125 11:06:52.869559 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f26a3b0-ba05-4484-8281-215ff2f09d25-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "0f26a3b0-ba05-4484-8281-215ff2f09d25" (UID: "0f26a3b0-ba05-4484-8281-215ff2f09d25"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:06:52 crc kubenswrapper[4821]: I1125 11:06:52.871805 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f26a3b0-ba05-4484-8281-215ff2f09d25-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0f26a3b0-ba05-4484-8281-215ff2f09d25" (UID: "0f26a3b0-ba05-4484-8281-215ff2f09d25"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:06:52 crc kubenswrapper[4821]: I1125 11:06:52.927836 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e2a415f6-1bf6-48e8-9f5a-2f29431b2274-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e2a415f6-1bf6-48e8-9f5a-2f29431b2274" (UID: "e2a415f6-1bf6-48e8-9f5a-2f29431b2274"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:06:52 crc kubenswrapper[4821]: I1125 11:06:52.932212 4821 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2a415f6-1bf6-48e8-9f5a-2f29431b2274-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 11:06:52 crc kubenswrapper[4821]: I1125 11:06:52.932249 4821 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0f26a3b0-ba05-4484-8281-215ff2f09d25-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Nov 25 11:06:52 crc kubenswrapper[4821]: I1125 11:06:52.932266 4821 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0f26a3b0-ba05-4484-8281-215ff2f09d25-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 11:06:52 crc kubenswrapper[4821]: I1125 11:06:52.932280 4821 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2a415f6-1bf6-48e8-9f5a-2f29431b2274-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 11:06:52 crc kubenswrapper[4821]: I1125 11:06:52.932293 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dkbzv\" (UniqueName: \"kubernetes.io/projected/0f26a3b0-ba05-4484-8281-215ff2f09d25-kube-api-access-dkbzv\") on node \"crc\" DevicePath \"\"" Nov 25 11:06:52 crc kubenswrapper[4821]: I1125 11:06:52.932305 4821 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f26a3b0-ba05-4484-8281-215ff2f09d25-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:06:52 crc kubenswrapper[4821]: I1125 11:06:52.932318 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hntgl\" (UniqueName: \"kubernetes.io/projected/e2a415f6-1bf6-48e8-9f5a-2f29431b2274-kube-api-access-hntgl\") on node \"crc\" DevicePath \"\"" Nov 25 11:06:52 crc kubenswrapper[4821]: I1125 11:06:52.932331 4821 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0f26a3b0-ba05-4484-8281-215ff2f09d25-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Nov 25 11:06:52 crc kubenswrapper[4821]: I1125 11:06:52.932342 4821 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0f26a3b0-ba05-4484-8281-215ff2f09d25-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 11:06:53 crc kubenswrapper[4821]: I1125 11:06:53.253711 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-225z5" event={"ID":"0f26a3b0-ba05-4484-8281-215ff2f09d25","Type":"ContainerDied","Data":"0485185ee68a2c205b995023a38085f1183bff73fdfcaad9900cdbc3bb75aa27"} Nov 25 11:06:53 crc kubenswrapper[4821]: I1125 11:06:53.253750 4821 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0485185ee68a2c205b995023a38085f1183bff73fdfcaad9900cdbc3bb75aa27" Nov 25 11:06:53 crc kubenswrapper[4821]: I1125 11:06:53.253754 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-225z5" Nov 25 11:06:53 crc kubenswrapper[4821]: I1125 11:06:53.257849 4821 generic.go:334] "Generic (PLEG): container finished" podID="e2a415f6-1bf6-48e8-9f5a-2f29431b2274" containerID="f1687930e3a59df28ca041fb80c722279d8d946b1242390d55220dc50f27583f" exitCode=0 Nov 25 11:06:53 crc kubenswrapper[4821]: I1125 11:06:53.257892 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6fh7d" event={"ID":"e2a415f6-1bf6-48e8-9f5a-2f29431b2274","Type":"ContainerDied","Data":"f1687930e3a59df28ca041fb80c722279d8d946b1242390d55220dc50f27583f"} Nov 25 11:06:53 crc kubenswrapper[4821]: I1125 11:06:53.257919 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6fh7d" event={"ID":"e2a415f6-1bf6-48e8-9f5a-2f29431b2274","Type":"ContainerDied","Data":"db16d46ec1b084955c174837ebbd4d41d388f451e6a5fea6e4eeca2f54831235"} Nov 25 11:06:53 crc kubenswrapper[4821]: I1125 11:06:53.257936 4821 scope.go:117] "RemoveContainer" containerID="f1687930e3a59df28ca041fb80c722279d8d946b1242390d55220dc50f27583f" Nov 25 11:06:53 crc kubenswrapper[4821]: I1125 11:06:53.258078 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6fh7d" Nov 25 11:06:53 crc kubenswrapper[4821]: I1125 11:06:53.286970 4821 scope.go:117] "RemoveContainer" containerID="ac02bd48ca1bf10f5c435751b3830d72369171b62ad7918bbb3f0f9a4a5749a7" Nov 25 11:06:53 crc kubenswrapper[4821]: I1125 11:06:53.306871 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6fh7d"] Nov 25 11:06:53 crc kubenswrapper[4821]: I1125 11:06:53.317786 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-6fh7d"] Nov 25 11:06:53 crc kubenswrapper[4821]: I1125 11:06:53.328387 4821 scope.go:117] "RemoveContainer" containerID="f0e10f151d2616ac0bb61af4847441ce1236ff7df6e219b881cf48b1ab4e5675" Nov 25 11:06:53 crc kubenswrapper[4821]: I1125 11:06:53.352538 4821 scope.go:117] "RemoveContainer" containerID="f1687930e3a59df28ca041fb80c722279d8d946b1242390d55220dc50f27583f" Nov 25 11:06:53 crc kubenswrapper[4821]: E1125 11:06:53.354369 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f1687930e3a59df28ca041fb80c722279d8d946b1242390d55220dc50f27583f\": container with ID starting with f1687930e3a59df28ca041fb80c722279d8d946b1242390d55220dc50f27583f not found: ID does not exist" containerID="f1687930e3a59df28ca041fb80c722279d8d946b1242390d55220dc50f27583f" Nov 25 11:06:53 crc kubenswrapper[4821]: I1125 11:06:53.354441 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1687930e3a59df28ca041fb80c722279d8d946b1242390d55220dc50f27583f"} err="failed to get container status \"f1687930e3a59df28ca041fb80c722279d8d946b1242390d55220dc50f27583f\": rpc error: code = NotFound desc = could not find container \"f1687930e3a59df28ca041fb80c722279d8d946b1242390d55220dc50f27583f\": container with ID starting with f1687930e3a59df28ca041fb80c722279d8d946b1242390d55220dc50f27583f not found: ID does not exist" Nov 25 11:06:53 crc kubenswrapper[4821]: I1125 11:06:53.354472 4821 scope.go:117] "RemoveContainer" containerID="ac02bd48ca1bf10f5c435751b3830d72369171b62ad7918bbb3f0f9a4a5749a7" Nov 25 11:06:53 crc kubenswrapper[4821]: E1125 11:06:53.355108 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac02bd48ca1bf10f5c435751b3830d72369171b62ad7918bbb3f0f9a4a5749a7\": container with ID starting with ac02bd48ca1bf10f5c435751b3830d72369171b62ad7918bbb3f0f9a4a5749a7 not found: ID does not exist" containerID="ac02bd48ca1bf10f5c435751b3830d72369171b62ad7918bbb3f0f9a4a5749a7" Nov 25 11:06:53 crc kubenswrapper[4821]: I1125 11:06:53.355192 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac02bd48ca1bf10f5c435751b3830d72369171b62ad7918bbb3f0f9a4a5749a7"} err="failed to get container status \"ac02bd48ca1bf10f5c435751b3830d72369171b62ad7918bbb3f0f9a4a5749a7\": rpc error: code = NotFound desc = could not find container \"ac02bd48ca1bf10f5c435751b3830d72369171b62ad7918bbb3f0f9a4a5749a7\": container with ID starting with ac02bd48ca1bf10f5c435751b3830d72369171b62ad7918bbb3f0f9a4a5749a7 not found: ID does not exist" Nov 25 11:06:53 crc kubenswrapper[4821]: I1125 11:06:53.355242 4821 scope.go:117] "RemoveContainer" containerID="f0e10f151d2616ac0bb61af4847441ce1236ff7df6e219b881cf48b1ab4e5675" Nov 25 11:06:53 crc kubenswrapper[4821]: E1125 11:06:53.357292 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f0e10f151d2616ac0bb61af4847441ce1236ff7df6e219b881cf48b1ab4e5675\": container with ID starting with f0e10f151d2616ac0bb61af4847441ce1236ff7df6e219b881cf48b1ab4e5675 not found: ID does not exist" containerID="f0e10f151d2616ac0bb61af4847441ce1236ff7df6e219b881cf48b1ab4e5675" Nov 25 11:06:53 crc kubenswrapper[4821]: I1125 11:06:53.357360 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f0e10f151d2616ac0bb61af4847441ce1236ff7df6e219b881cf48b1ab4e5675"} err="failed to get container status \"f0e10f151d2616ac0bb61af4847441ce1236ff7df6e219b881cf48b1ab4e5675\": rpc error: code = NotFound desc = could not find container \"f0e10f151d2616ac0bb61af4847441ce1236ff7df6e219b881cf48b1ab4e5675\": container with ID starting with f0e10f151d2616ac0bb61af4847441ce1236ff7df6e219b881cf48b1ab4e5675 not found: ID does not exist" Nov 25 11:06:53 crc kubenswrapper[4821]: I1125 11:06:53.386683 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9w8v8"] Nov 25 11:06:53 crc kubenswrapper[4821]: E1125 11:06:53.387148 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2a415f6-1bf6-48e8-9f5a-2f29431b2274" containerName="registry-server" Nov 25 11:06:53 crc kubenswrapper[4821]: I1125 11:06:53.387191 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2a415f6-1bf6-48e8-9f5a-2f29431b2274" containerName="registry-server" Nov 25 11:06:53 crc kubenswrapper[4821]: E1125 11:06:53.387210 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2a415f6-1bf6-48e8-9f5a-2f29431b2274" containerName="extract-utilities" Nov 25 11:06:53 crc kubenswrapper[4821]: I1125 11:06:53.387218 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2a415f6-1bf6-48e8-9f5a-2f29431b2274" containerName="extract-utilities" Nov 25 11:06:53 crc kubenswrapper[4821]: E1125 11:06:53.387234 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="284b03e4-ca38-42e2-92b8-11c812a2b751" containerName="registry-server" Nov 25 11:06:53 crc kubenswrapper[4821]: I1125 11:06:53.387243 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="284b03e4-ca38-42e2-92b8-11c812a2b751" containerName="registry-server" Nov 25 11:06:53 crc kubenswrapper[4821]: E1125 11:06:53.387261 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="284b03e4-ca38-42e2-92b8-11c812a2b751" containerName="extract-utilities" Nov 25 11:06:53 crc kubenswrapper[4821]: I1125 11:06:53.387269 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="284b03e4-ca38-42e2-92b8-11c812a2b751" containerName="extract-utilities" Nov 25 11:06:53 crc kubenswrapper[4821]: E1125 11:06:53.387286 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2a415f6-1bf6-48e8-9f5a-2f29431b2274" containerName="extract-content" Nov 25 11:06:53 crc kubenswrapper[4821]: I1125 11:06:53.387293 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2a415f6-1bf6-48e8-9f5a-2f29431b2274" containerName="extract-content" Nov 25 11:06:53 crc kubenswrapper[4821]: E1125 11:06:53.387305 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f26a3b0-ba05-4484-8281-215ff2f09d25" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 25 11:06:53 crc kubenswrapper[4821]: I1125 11:06:53.387313 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f26a3b0-ba05-4484-8281-215ff2f09d25" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 25 11:06:53 crc kubenswrapper[4821]: E1125 11:06:53.387331 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="284b03e4-ca38-42e2-92b8-11c812a2b751" containerName="extract-content" Nov 25 11:06:53 crc kubenswrapper[4821]: I1125 11:06:53.387340 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="284b03e4-ca38-42e2-92b8-11c812a2b751" containerName="extract-content" Nov 25 11:06:53 crc kubenswrapper[4821]: I1125 11:06:53.387603 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2a415f6-1bf6-48e8-9f5a-2f29431b2274" containerName="registry-server" Nov 25 11:06:53 crc kubenswrapper[4821]: I1125 11:06:53.387627 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="284b03e4-ca38-42e2-92b8-11c812a2b751" containerName="registry-server" Nov 25 11:06:53 crc kubenswrapper[4821]: I1125 11:06:53.387641 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f26a3b0-ba05-4484-8281-215ff2f09d25" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 25 11:06:53 crc kubenswrapper[4821]: I1125 11:06:53.388416 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9w8v8" Nov 25 11:06:53 crc kubenswrapper[4821]: I1125 11:06:53.391176 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 11:06:53 crc kubenswrapper[4821]: I1125 11:06:53.391331 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Nov 25 11:06:53 crc kubenswrapper[4821]: I1125 11:06:53.391411 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 11:06:53 crc kubenswrapper[4821]: I1125 11:06:53.391495 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-z764x" Nov 25 11:06:53 crc kubenswrapper[4821]: I1125 11:06:53.391604 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 11:06:53 crc kubenswrapper[4821]: I1125 11:06:53.398424 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9w8v8"] Nov 25 11:06:53 crc kubenswrapper[4821]: I1125 11:06:53.441787 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0ee4d7f9-162b-4882-b189-13c53bcdc274-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9w8v8\" (UID: \"0ee4d7f9-162b-4882-b189-13c53bcdc274\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9w8v8" Nov 25 11:06:53 crc kubenswrapper[4821]: I1125 11:06:53.441832 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7j8r\" (UniqueName: \"kubernetes.io/projected/0ee4d7f9-162b-4882-b189-13c53bcdc274-kube-api-access-r7j8r\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9w8v8\" (UID: \"0ee4d7f9-162b-4882-b189-13c53bcdc274\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9w8v8" Nov 25 11:06:53 crc kubenswrapper[4821]: I1125 11:06:53.441871 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/0ee4d7f9-162b-4882-b189-13c53bcdc274-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9w8v8\" (UID: \"0ee4d7f9-162b-4882-b189-13c53bcdc274\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9w8v8" Nov 25 11:06:53 crc kubenswrapper[4821]: I1125 11:06:53.441914 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0ee4d7f9-162b-4882-b189-13c53bcdc274-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9w8v8\" (UID: \"0ee4d7f9-162b-4882-b189-13c53bcdc274\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9w8v8" Nov 25 11:06:53 crc kubenswrapper[4821]: I1125 11:06:53.441976 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ee4d7f9-162b-4882-b189-13c53bcdc274-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9w8v8\" (UID: \"0ee4d7f9-162b-4882-b189-13c53bcdc274\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9w8v8" Nov 25 11:06:53 crc kubenswrapper[4821]: I1125 11:06:53.543361 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0ee4d7f9-162b-4882-b189-13c53bcdc274-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9w8v8\" (UID: \"0ee4d7f9-162b-4882-b189-13c53bcdc274\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9w8v8" Nov 25 11:06:53 crc kubenswrapper[4821]: I1125 11:06:53.543767 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ee4d7f9-162b-4882-b189-13c53bcdc274-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9w8v8\" (UID: \"0ee4d7f9-162b-4882-b189-13c53bcdc274\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9w8v8" Nov 25 11:06:53 crc kubenswrapper[4821]: I1125 11:06:53.543839 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7j8r\" (UniqueName: \"kubernetes.io/projected/0ee4d7f9-162b-4882-b189-13c53bcdc274-kube-api-access-r7j8r\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9w8v8\" (UID: \"0ee4d7f9-162b-4882-b189-13c53bcdc274\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9w8v8" Nov 25 11:06:53 crc kubenswrapper[4821]: I1125 11:06:53.543857 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0ee4d7f9-162b-4882-b189-13c53bcdc274-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9w8v8\" (UID: \"0ee4d7f9-162b-4882-b189-13c53bcdc274\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9w8v8" Nov 25 11:06:53 crc kubenswrapper[4821]: I1125 11:06:53.543903 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/0ee4d7f9-162b-4882-b189-13c53bcdc274-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9w8v8\" (UID: \"0ee4d7f9-162b-4882-b189-13c53bcdc274\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9w8v8" Nov 25 11:06:53 crc kubenswrapper[4821]: I1125 11:06:53.547494 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ee4d7f9-162b-4882-b189-13c53bcdc274-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9w8v8\" (UID: \"0ee4d7f9-162b-4882-b189-13c53bcdc274\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9w8v8" Nov 25 11:06:53 crc kubenswrapper[4821]: I1125 11:06:53.547557 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0ee4d7f9-162b-4882-b189-13c53bcdc274-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9w8v8\" (UID: \"0ee4d7f9-162b-4882-b189-13c53bcdc274\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9w8v8" Nov 25 11:06:53 crc kubenswrapper[4821]: I1125 11:06:53.547780 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/0ee4d7f9-162b-4882-b189-13c53bcdc274-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9w8v8\" (UID: \"0ee4d7f9-162b-4882-b189-13c53bcdc274\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9w8v8" Nov 25 11:06:53 crc kubenswrapper[4821]: I1125 11:06:53.548338 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0ee4d7f9-162b-4882-b189-13c53bcdc274-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9w8v8\" (UID: \"0ee4d7f9-162b-4882-b189-13c53bcdc274\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9w8v8" Nov 25 11:06:53 crc kubenswrapper[4821]: I1125 11:06:53.560904 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7j8r\" (UniqueName: \"kubernetes.io/projected/0ee4d7f9-162b-4882-b189-13c53bcdc274-kube-api-access-r7j8r\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9w8v8\" (UID: \"0ee4d7f9-162b-4882-b189-13c53bcdc274\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9w8v8" Nov 25 11:06:53 crc kubenswrapper[4821]: I1125 11:06:53.724763 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9w8v8" Nov 25 11:06:54 crc kubenswrapper[4821]: I1125 11:06:54.129917 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e2a415f6-1bf6-48e8-9f5a-2f29431b2274" path="/var/lib/kubelet/pods/e2a415f6-1bf6-48e8-9f5a-2f29431b2274/volumes" Nov 25 11:06:54 crc kubenswrapper[4821]: I1125 11:06:54.337986 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9w8v8"] Nov 25 11:06:55 crc kubenswrapper[4821]: I1125 11:06:55.278524 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9w8v8" event={"ID":"0ee4d7f9-162b-4882-b189-13c53bcdc274","Type":"ContainerStarted","Data":"313214dcc72f79983d312f27b9a9eb73bb5a1d376206544ebba551b0902efe58"} Nov 25 11:06:56 crc kubenswrapper[4821]: I1125 11:06:56.292350 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9w8v8" event={"ID":"0ee4d7f9-162b-4882-b189-13c53bcdc274","Type":"ContainerStarted","Data":"94388636806b204fde9e3bf85b805d500b31f590fc25a3b91d46ab2294fde1cf"} Nov 25 11:06:56 crc kubenswrapper[4821]: I1125 11:06:56.310239 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9w8v8" podStartSLOduration=2.547051485 podStartE2EDuration="3.310218827s" podCreationTimestamp="2025-11-25 11:06:53 +0000 UTC" firstStartedPulling="2025-11-25 11:06:54.342542758 +0000 UTC m=+2084.878862605" lastFinishedPulling="2025-11-25 11:06:55.10571009 +0000 UTC m=+2085.642029947" observedRunningTime="2025-11-25 11:06:56.309855318 +0000 UTC m=+2086.846175175" watchObservedRunningTime="2025-11-25 11:06:56.310218827 +0000 UTC m=+2086.846538684" Nov 25 11:06:56 crc kubenswrapper[4821]: I1125 11:06:56.586867 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-pxx4s"] Nov 25 11:06:56 crc kubenswrapper[4821]: I1125 11:06:56.588680 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pxx4s" Nov 25 11:06:56 crc kubenswrapper[4821]: I1125 11:06:56.609939 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pxx4s"] Nov 25 11:06:56 crc kubenswrapper[4821]: I1125 11:06:56.707598 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mlkb4\" (UniqueName: \"kubernetes.io/projected/1e083c7a-3c7c-4dd9-9072-a32e82c2a533-kube-api-access-mlkb4\") pod \"community-operators-pxx4s\" (UID: \"1e083c7a-3c7c-4dd9-9072-a32e82c2a533\") " pod="openshift-marketplace/community-operators-pxx4s" Nov 25 11:06:56 crc kubenswrapper[4821]: I1125 11:06:56.707710 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e083c7a-3c7c-4dd9-9072-a32e82c2a533-utilities\") pod \"community-operators-pxx4s\" (UID: \"1e083c7a-3c7c-4dd9-9072-a32e82c2a533\") " pod="openshift-marketplace/community-operators-pxx4s" Nov 25 11:06:56 crc kubenswrapper[4821]: I1125 11:06:56.707807 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e083c7a-3c7c-4dd9-9072-a32e82c2a533-catalog-content\") pod \"community-operators-pxx4s\" (UID: \"1e083c7a-3c7c-4dd9-9072-a32e82c2a533\") " pod="openshift-marketplace/community-operators-pxx4s" Nov 25 11:06:56 crc kubenswrapper[4821]: I1125 11:06:56.809535 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mlkb4\" (UniqueName: \"kubernetes.io/projected/1e083c7a-3c7c-4dd9-9072-a32e82c2a533-kube-api-access-mlkb4\") pod \"community-operators-pxx4s\" (UID: \"1e083c7a-3c7c-4dd9-9072-a32e82c2a533\") " pod="openshift-marketplace/community-operators-pxx4s" Nov 25 11:06:56 crc kubenswrapper[4821]: I1125 11:06:56.809647 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e083c7a-3c7c-4dd9-9072-a32e82c2a533-utilities\") pod \"community-operators-pxx4s\" (UID: \"1e083c7a-3c7c-4dd9-9072-a32e82c2a533\") " pod="openshift-marketplace/community-operators-pxx4s" Nov 25 11:06:56 crc kubenswrapper[4821]: I1125 11:06:56.809772 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e083c7a-3c7c-4dd9-9072-a32e82c2a533-catalog-content\") pod \"community-operators-pxx4s\" (UID: \"1e083c7a-3c7c-4dd9-9072-a32e82c2a533\") " pod="openshift-marketplace/community-operators-pxx4s" Nov 25 11:06:56 crc kubenswrapper[4821]: I1125 11:06:56.810128 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e083c7a-3c7c-4dd9-9072-a32e82c2a533-utilities\") pod \"community-operators-pxx4s\" (UID: \"1e083c7a-3c7c-4dd9-9072-a32e82c2a533\") " pod="openshift-marketplace/community-operators-pxx4s" Nov 25 11:06:56 crc kubenswrapper[4821]: I1125 11:06:56.810264 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e083c7a-3c7c-4dd9-9072-a32e82c2a533-catalog-content\") pod \"community-operators-pxx4s\" (UID: \"1e083c7a-3c7c-4dd9-9072-a32e82c2a533\") " pod="openshift-marketplace/community-operators-pxx4s" Nov 25 11:06:56 crc kubenswrapper[4821]: I1125 11:06:56.833249 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mlkb4\" (UniqueName: \"kubernetes.io/projected/1e083c7a-3c7c-4dd9-9072-a32e82c2a533-kube-api-access-mlkb4\") pod \"community-operators-pxx4s\" (UID: \"1e083c7a-3c7c-4dd9-9072-a32e82c2a533\") " pod="openshift-marketplace/community-operators-pxx4s" Nov 25 11:06:56 crc kubenswrapper[4821]: I1125 11:06:56.913073 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pxx4s" Nov 25 11:06:57 crc kubenswrapper[4821]: I1125 11:06:57.393255 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pxx4s"] Nov 25 11:06:57 crc kubenswrapper[4821]: W1125 11:06:57.396109 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1e083c7a_3c7c_4dd9_9072_a32e82c2a533.slice/crio-c35903d5e4e451a4fa3f552f9af99113eecc727d52d9b5b327872c07de25bb76 WatchSource:0}: Error finding container c35903d5e4e451a4fa3f552f9af99113eecc727d52d9b5b327872c07de25bb76: Status 404 returned error can't find the container with id c35903d5e4e451a4fa3f552f9af99113eecc727d52d9b5b327872c07de25bb76 Nov 25 11:06:58 crc kubenswrapper[4821]: I1125 11:06:58.312088 4821 generic.go:334] "Generic (PLEG): container finished" podID="1e083c7a-3c7c-4dd9-9072-a32e82c2a533" containerID="81c714fe43b28018ef0e9c81f1aa133c52cea6a88d4fee3c99754da30c3d251f" exitCode=0 Nov 25 11:06:58 crc kubenswrapper[4821]: I1125 11:06:58.312399 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pxx4s" event={"ID":"1e083c7a-3c7c-4dd9-9072-a32e82c2a533","Type":"ContainerDied","Data":"81c714fe43b28018ef0e9c81f1aa133c52cea6a88d4fee3c99754da30c3d251f"} Nov 25 11:06:58 crc kubenswrapper[4821]: I1125 11:06:58.312472 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pxx4s" event={"ID":"1e083c7a-3c7c-4dd9-9072-a32e82c2a533","Type":"ContainerStarted","Data":"c35903d5e4e451a4fa3f552f9af99113eecc727d52d9b5b327872c07de25bb76"} Nov 25 11:07:00 crc kubenswrapper[4821]: I1125 11:07:00.334955 4821 generic.go:334] "Generic (PLEG): container finished" podID="1e083c7a-3c7c-4dd9-9072-a32e82c2a533" containerID="62391c14eaccd54e29ed0c19408ed259a2d64436522ea260112df8e1493d9654" exitCode=0 Nov 25 11:07:00 crc kubenswrapper[4821]: I1125 11:07:00.335083 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pxx4s" event={"ID":"1e083c7a-3c7c-4dd9-9072-a32e82c2a533","Type":"ContainerDied","Data":"62391c14eaccd54e29ed0c19408ed259a2d64436522ea260112df8e1493d9654"} Nov 25 11:07:01 crc kubenswrapper[4821]: I1125 11:07:01.346267 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pxx4s" event={"ID":"1e083c7a-3c7c-4dd9-9072-a32e82c2a533","Type":"ContainerStarted","Data":"85f58d91063dfb06dbf7c2a3e3d2a366919e14563fed7b93f6d538232af80de5"} Nov 25 11:07:04 crc kubenswrapper[4821]: I1125 11:07:04.715057 4821 patch_prober.go:28] interesting pod/machine-config-daemon-2krbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:07:04 crc kubenswrapper[4821]: I1125 11:07:04.715478 4821 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:07:06 crc kubenswrapper[4821]: I1125 11:07:06.914009 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-pxx4s" Nov 25 11:07:06 crc kubenswrapper[4821]: I1125 11:07:06.914375 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-pxx4s" Nov 25 11:07:06 crc kubenswrapper[4821]: I1125 11:07:06.958825 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-pxx4s" Nov 25 11:07:06 crc kubenswrapper[4821]: I1125 11:07:06.981387 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-pxx4s" podStartSLOduration=8.557231275 podStartE2EDuration="10.981368883s" podCreationTimestamp="2025-11-25 11:06:56 +0000 UTC" firstStartedPulling="2025-11-25 11:06:58.313786629 +0000 UTC m=+2088.850106476" lastFinishedPulling="2025-11-25 11:07:00.737924237 +0000 UTC m=+2091.274244084" observedRunningTime="2025-11-25 11:07:01.371307152 +0000 UTC m=+2091.907627009" watchObservedRunningTime="2025-11-25 11:07:06.981368883 +0000 UTC m=+2097.517688730" Nov 25 11:07:07 crc kubenswrapper[4821]: I1125 11:07:07.440561 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-pxx4s" Nov 25 11:07:07 crc kubenswrapper[4821]: I1125 11:07:07.499519 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pxx4s"] Nov 25 11:07:09 crc kubenswrapper[4821]: I1125 11:07:09.411733 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-pxx4s" podUID="1e083c7a-3c7c-4dd9-9072-a32e82c2a533" containerName="registry-server" containerID="cri-o://85f58d91063dfb06dbf7c2a3e3d2a366919e14563fed7b93f6d538232af80de5" gracePeriod=2 Nov 25 11:07:09 crc kubenswrapper[4821]: I1125 11:07:09.957847 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pxx4s" Nov 25 11:07:10 crc kubenswrapper[4821]: I1125 11:07:10.071839 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e083c7a-3c7c-4dd9-9072-a32e82c2a533-catalog-content\") pod \"1e083c7a-3c7c-4dd9-9072-a32e82c2a533\" (UID: \"1e083c7a-3c7c-4dd9-9072-a32e82c2a533\") " Nov 25 11:07:10 crc kubenswrapper[4821]: I1125 11:07:10.071967 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mlkb4\" (UniqueName: \"kubernetes.io/projected/1e083c7a-3c7c-4dd9-9072-a32e82c2a533-kube-api-access-mlkb4\") pod \"1e083c7a-3c7c-4dd9-9072-a32e82c2a533\" (UID: \"1e083c7a-3c7c-4dd9-9072-a32e82c2a533\") " Nov 25 11:07:10 crc kubenswrapper[4821]: I1125 11:07:10.072085 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e083c7a-3c7c-4dd9-9072-a32e82c2a533-utilities\") pod \"1e083c7a-3c7c-4dd9-9072-a32e82c2a533\" (UID: \"1e083c7a-3c7c-4dd9-9072-a32e82c2a533\") " Nov 25 11:07:10 crc kubenswrapper[4821]: I1125 11:07:10.072859 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1e083c7a-3c7c-4dd9-9072-a32e82c2a533-utilities" (OuterVolumeSpecName: "utilities") pod "1e083c7a-3c7c-4dd9-9072-a32e82c2a533" (UID: "1e083c7a-3c7c-4dd9-9072-a32e82c2a533"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:07:10 crc kubenswrapper[4821]: I1125 11:07:10.077593 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e083c7a-3c7c-4dd9-9072-a32e82c2a533-kube-api-access-mlkb4" (OuterVolumeSpecName: "kube-api-access-mlkb4") pod "1e083c7a-3c7c-4dd9-9072-a32e82c2a533" (UID: "1e083c7a-3c7c-4dd9-9072-a32e82c2a533"). InnerVolumeSpecName "kube-api-access-mlkb4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:07:10 crc kubenswrapper[4821]: I1125 11:07:10.137720 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1e083c7a-3c7c-4dd9-9072-a32e82c2a533-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1e083c7a-3c7c-4dd9-9072-a32e82c2a533" (UID: "1e083c7a-3c7c-4dd9-9072-a32e82c2a533"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:07:10 crc kubenswrapper[4821]: I1125 11:07:10.174738 4821 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e083c7a-3c7c-4dd9-9072-a32e82c2a533-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 11:07:10 crc kubenswrapper[4821]: I1125 11:07:10.174773 4821 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e083c7a-3c7c-4dd9-9072-a32e82c2a533-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 11:07:10 crc kubenswrapper[4821]: I1125 11:07:10.174784 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mlkb4\" (UniqueName: \"kubernetes.io/projected/1e083c7a-3c7c-4dd9-9072-a32e82c2a533-kube-api-access-mlkb4\") on node \"crc\" DevicePath \"\"" Nov 25 11:07:10 crc kubenswrapper[4821]: I1125 11:07:10.439096 4821 generic.go:334] "Generic (PLEG): container finished" podID="1e083c7a-3c7c-4dd9-9072-a32e82c2a533" containerID="85f58d91063dfb06dbf7c2a3e3d2a366919e14563fed7b93f6d538232af80de5" exitCode=0 Nov 25 11:07:10 crc kubenswrapper[4821]: I1125 11:07:10.439187 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pxx4s" Nov 25 11:07:10 crc kubenswrapper[4821]: I1125 11:07:10.439223 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pxx4s" event={"ID":"1e083c7a-3c7c-4dd9-9072-a32e82c2a533","Type":"ContainerDied","Data":"85f58d91063dfb06dbf7c2a3e3d2a366919e14563fed7b93f6d538232af80de5"} Nov 25 11:07:10 crc kubenswrapper[4821]: I1125 11:07:10.440299 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pxx4s" event={"ID":"1e083c7a-3c7c-4dd9-9072-a32e82c2a533","Type":"ContainerDied","Data":"c35903d5e4e451a4fa3f552f9af99113eecc727d52d9b5b327872c07de25bb76"} Nov 25 11:07:10 crc kubenswrapper[4821]: I1125 11:07:10.440327 4821 scope.go:117] "RemoveContainer" containerID="85f58d91063dfb06dbf7c2a3e3d2a366919e14563fed7b93f6d538232af80de5" Nov 25 11:07:10 crc kubenswrapper[4821]: I1125 11:07:10.473732 4821 scope.go:117] "RemoveContainer" containerID="62391c14eaccd54e29ed0c19408ed259a2d64436522ea260112df8e1493d9654" Nov 25 11:07:10 crc kubenswrapper[4821]: I1125 11:07:10.482035 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pxx4s"] Nov 25 11:07:10 crc kubenswrapper[4821]: I1125 11:07:10.497356 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-pxx4s"] Nov 25 11:07:10 crc kubenswrapper[4821]: I1125 11:07:10.498421 4821 scope.go:117] "RemoveContainer" containerID="81c714fe43b28018ef0e9c81f1aa133c52cea6a88d4fee3c99754da30c3d251f" Nov 25 11:07:10 crc kubenswrapper[4821]: I1125 11:07:10.540556 4821 scope.go:117] "RemoveContainer" containerID="85f58d91063dfb06dbf7c2a3e3d2a366919e14563fed7b93f6d538232af80de5" Nov 25 11:07:10 crc kubenswrapper[4821]: E1125 11:07:10.541773 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85f58d91063dfb06dbf7c2a3e3d2a366919e14563fed7b93f6d538232af80de5\": container with ID starting with 85f58d91063dfb06dbf7c2a3e3d2a366919e14563fed7b93f6d538232af80de5 not found: ID does not exist" containerID="85f58d91063dfb06dbf7c2a3e3d2a366919e14563fed7b93f6d538232af80de5" Nov 25 11:07:10 crc kubenswrapper[4821]: I1125 11:07:10.541825 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85f58d91063dfb06dbf7c2a3e3d2a366919e14563fed7b93f6d538232af80de5"} err="failed to get container status \"85f58d91063dfb06dbf7c2a3e3d2a366919e14563fed7b93f6d538232af80de5\": rpc error: code = NotFound desc = could not find container \"85f58d91063dfb06dbf7c2a3e3d2a366919e14563fed7b93f6d538232af80de5\": container with ID starting with 85f58d91063dfb06dbf7c2a3e3d2a366919e14563fed7b93f6d538232af80de5 not found: ID does not exist" Nov 25 11:07:10 crc kubenswrapper[4821]: I1125 11:07:10.541856 4821 scope.go:117] "RemoveContainer" containerID="62391c14eaccd54e29ed0c19408ed259a2d64436522ea260112df8e1493d9654" Nov 25 11:07:10 crc kubenswrapper[4821]: E1125 11:07:10.542207 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"62391c14eaccd54e29ed0c19408ed259a2d64436522ea260112df8e1493d9654\": container with ID starting with 62391c14eaccd54e29ed0c19408ed259a2d64436522ea260112df8e1493d9654 not found: ID does not exist" containerID="62391c14eaccd54e29ed0c19408ed259a2d64436522ea260112df8e1493d9654" Nov 25 11:07:10 crc kubenswrapper[4821]: I1125 11:07:10.542238 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62391c14eaccd54e29ed0c19408ed259a2d64436522ea260112df8e1493d9654"} err="failed to get container status \"62391c14eaccd54e29ed0c19408ed259a2d64436522ea260112df8e1493d9654\": rpc error: code = NotFound desc = could not find container \"62391c14eaccd54e29ed0c19408ed259a2d64436522ea260112df8e1493d9654\": container with ID starting with 62391c14eaccd54e29ed0c19408ed259a2d64436522ea260112df8e1493d9654 not found: ID does not exist" Nov 25 11:07:10 crc kubenswrapper[4821]: I1125 11:07:10.542255 4821 scope.go:117] "RemoveContainer" containerID="81c714fe43b28018ef0e9c81f1aa133c52cea6a88d4fee3c99754da30c3d251f" Nov 25 11:07:10 crc kubenswrapper[4821]: E1125 11:07:10.542578 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"81c714fe43b28018ef0e9c81f1aa133c52cea6a88d4fee3c99754da30c3d251f\": container with ID starting with 81c714fe43b28018ef0e9c81f1aa133c52cea6a88d4fee3c99754da30c3d251f not found: ID does not exist" containerID="81c714fe43b28018ef0e9c81f1aa133c52cea6a88d4fee3c99754da30c3d251f" Nov 25 11:07:10 crc kubenswrapper[4821]: I1125 11:07:10.542616 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81c714fe43b28018ef0e9c81f1aa133c52cea6a88d4fee3c99754da30c3d251f"} err="failed to get container status \"81c714fe43b28018ef0e9c81f1aa133c52cea6a88d4fee3c99754da30c3d251f\": rpc error: code = NotFound desc = could not find container \"81c714fe43b28018ef0e9c81f1aa133c52cea6a88d4fee3c99754da30c3d251f\": container with ID starting with 81c714fe43b28018ef0e9c81f1aa133c52cea6a88d4fee3c99754da30c3d251f not found: ID does not exist" Nov 25 11:07:12 crc kubenswrapper[4821]: I1125 11:07:12.130663 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e083c7a-3c7c-4dd9-9072-a32e82c2a533" path="/var/lib/kubelet/pods/1e083c7a-3c7c-4dd9-9072-a32e82c2a533/volumes" Nov 25 11:07:34 crc kubenswrapper[4821]: I1125 11:07:34.714671 4821 patch_prober.go:28] interesting pod/machine-config-daemon-2krbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:07:34 crc kubenswrapper[4821]: I1125 11:07:34.715155 4821 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:07:34 crc kubenswrapper[4821]: I1125 11:07:34.715234 4821 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" Nov 25 11:07:34 crc kubenswrapper[4821]: I1125 11:07:34.716013 4821 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"67625e9eb479a7295b3be108a852f5678838cf58bdadd2502f30ab844bb33b89"} pod="openshift-machine-config-operator/machine-config-daemon-2krbf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 11:07:34 crc kubenswrapper[4821]: I1125 11:07:34.716082 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerName="machine-config-daemon" containerID="cri-o://67625e9eb479a7295b3be108a852f5678838cf58bdadd2502f30ab844bb33b89" gracePeriod=600 Nov 25 11:07:35 crc kubenswrapper[4821]: I1125 11:07:35.665368 4821 generic.go:334] "Generic (PLEG): container finished" podID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerID="67625e9eb479a7295b3be108a852f5678838cf58bdadd2502f30ab844bb33b89" exitCode=0 Nov 25 11:07:35 crc kubenswrapper[4821]: I1125 11:07:35.665439 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" event={"ID":"5f948b87-ac86-4de6-ad64-c2ef947f84d4","Type":"ContainerDied","Data":"67625e9eb479a7295b3be108a852f5678838cf58bdadd2502f30ab844bb33b89"} Nov 25 11:07:35 crc kubenswrapper[4821]: I1125 11:07:35.665951 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" event={"ID":"5f948b87-ac86-4de6-ad64-c2ef947f84d4","Type":"ContainerStarted","Data":"61010995905bb7c40fff6417cf7e0b11d584b8a9a68f6fccf71aefe231f45291"} Nov 25 11:07:35 crc kubenswrapper[4821]: I1125 11:07:35.665984 4821 scope.go:117] "RemoveContainer" containerID="fc16b2ea0cd36da4cef842d21755ec621134a4c21c5c964f8856fb8e6a102b2d" Nov 25 11:08:53 crc kubenswrapper[4821]: I1125 11:08:53.308328 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-t6688"] Nov 25 11:08:53 crc kubenswrapper[4821]: E1125 11:08:53.309275 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e083c7a-3c7c-4dd9-9072-a32e82c2a533" containerName="extract-content" Nov 25 11:08:53 crc kubenswrapper[4821]: I1125 11:08:53.309291 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e083c7a-3c7c-4dd9-9072-a32e82c2a533" containerName="extract-content" Nov 25 11:08:53 crc kubenswrapper[4821]: E1125 11:08:53.309315 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e083c7a-3c7c-4dd9-9072-a32e82c2a533" containerName="registry-server" Nov 25 11:08:53 crc kubenswrapper[4821]: I1125 11:08:53.309324 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e083c7a-3c7c-4dd9-9072-a32e82c2a533" containerName="registry-server" Nov 25 11:08:53 crc kubenswrapper[4821]: E1125 11:08:53.309342 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e083c7a-3c7c-4dd9-9072-a32e82c2a533" containerName="extract-utilities" Nov 25 11:08:53 crc kubenswrapper[4821]: I1125 11:08:53.309350 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e083c7a-3c7c-4dd9-9072-a32e82c2a533" containerName="extract-utilities" Nov 25 11:08:53 crc kubenswrapper[4821]: I1125 11:08:53.309615 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e083c7a-3c7c-4dd9-9072-a32e82c2a533" containerName="registry-server" Nov 25 11:08:53 crc kubenswrapper[4821]: I1125 11:08:53.312323 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t6688" Nov 25 11:08:53 crc kubenswrapper[4821]: I1125 11:08:53.324910 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-t6688"] Nov 25 11:08:53 crc kubenswrapper[4821]: I1125 11:08:53.475021 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af82c304-6d71-4420-840a-c299a888d85a-catalog-content\") pod \"certified-operators-t6688\" (UID: \"af82c304-6d71-4420-840a-c299a888d85a\") " pod="openshift-marketplace/certified-operators-t6688" Nov 25 11:08:53 crc kubenswrapper[4821]: I1125 11:08:53.475078 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af82c304-6d71-4420-840a-c299a888d85a-utilities\") pod \"certified-operators-t6688\" (UID: \"af82c304-6d71-4420-840a-c299a888d85a\") " pod="openshift-marketplace/certified-operators-t6688" Nov 25 11:08:53 crc kubenswrapper[4821]: I1125 11:08:53.475139 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5wh5t\" (UniqueName: \"kubernetes.io/projected/af82c304-6d71-4420-840a-c299a888d85a-kube-api-access-5wh5t\") pod \"certified-operators-t6688\" (UID: \"af82c304-6d71-4420-840a-c299a888d85a\") " pod="openshift-marketplace/certified-operators-t6688" Nov 25 11:08:53 crc kubenswrapper[4821]: I1125 11:08:53.577565 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5wh5t\" (UniqueName: \"kubernetes.io/projected/af82c304-6d71-4420-840a-c299a888d85a-kube-api-access-5wh5t\") pod \"certified-operators-t6688\" (UID: \"af82c304-6d71-4420-840a-c299a888d85a\") " pod="openshift-marketplace/certified-operators-t6688" Nov 25 11:08:53 crc kubenswrapper[4821]: I1125 11:08:53.577801 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af82c304-6d71-4420-840a-c299a888d85a-catalog-content\") pod \"certified-operators-t6688\" (UID: \"af82c304-6d71-4420-840a-c299a888d85a\") " pod="openshift-marketplace/certified-operators-t6688" Nov 25 11:08:53 crc kubenswrapper[4821]: I1125 11:08:53.577832 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af82c304-6d71-4420-840a-c299a888d85a-utilities\") pod \"certified-operators-t6688\" (UID: \"af82c304-6d71-4420-840a-c299a888d85a\") " pod="openshift-marketplace/certified-operators-t6688" Nov 25 11:08:53 crc kubenswrapper[4821]: I1125 11:08:53.578395 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af82c304-6d71-4420-840a-c299a888d85a-utilities\") pod \"certified-operators-t6688\" (UID: \"af82c304-6d71-4420-840a-c299a888d85a\") " pod="openshift-marketplace/certified-operators-t6688" Nov 25 11:08:53 crc kubenswrapper[4821]: I1125 11:08:53.578435 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af82c304-6d71-4420-840a-c299a888d85a-catalog-content\") pod \"certified-operators-t6688\" (UID: \"af82c304-6d71-4420-840a-c299a888d85a\") " pod="openshift-marketplace/certified-operators-t6688" Nov 25 11:08:53 crc kubenswrapper[4821]: I1125 11:08:53.608069 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5wh5t\" (UniqueName: \"kubernetes.io/projected/af82c304-6d71-4420-840a-c299a888d85a-kube-api-access-5wh5t\") pod \"certified-operators-t6688\" (UID: \"af82c304-6d71-4420-840a-c299a888d85a\") " pod="openshift-marketplace/certified-operators-t6688" Nov 25 11:08:53 crc kubenswrapper[4821]: I1125 11:08:53.638070 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t6688" Nov 25 11:08:54 crc kubenswrapper[4821]: I1125 11:08:54.147392 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-t6688"] Nov 25 11:08:54 crc kubenswrapper[4821]: W1125 11:08:54.149149 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaf82c304_6d71_4420_840a_c299a888d85a.slice/crio-53427aa5a42d3e02aba2d4b1420b52c1e781378ab5b90958a450e604a1ef2f58 WatchSource:0}: Error finding container 53427aa5a42d3e02aba2d4b1420b52c1e781378ab5b90958a450e604a1ef2f58: Status 404 returned error can't find the container with id 53427aa5a42d3e02aba2d4b1420b52c1e781378ab5b90958a450e604a1ef2f58 Nov 25 11:08:54 crc kubenswrapper[4821]: I1125 11:08:54.342866 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t6688" event={"ID":"af82c304-6d71-4420-840a-c299a888d85a","Type":"ContainerStarted","Data":"53427aa5a42d3e02aba2d4b1420b52c1e781378ab5b90958a450e604a1ef2f58"} Nov 25 11:08:55 crc kubenswrapper[4821]: I1125 11:08:55.351783 4821 generic.go:334] "Generic (PLEG): container finished" podID="af82c304-6d71-4420-840a-c299a888d85a" containerID="a105ab6a3e6ea159d604ce1df2fa1f18fe22324605cf0e8b72b7e77c80651673" exitCode=0 Nov 25 11:08:55 crc kubenswrapper[4821]: I1125 11:08:55.351834 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t6688" event={"ID":"af82c304-6d71-4420-840a-c299a888d85a","Type":"ContainerDied","Data":"a105ab6a3e6ea159d604ce1df2fa1f18fe22324605cf0e8b72b7e77c80651673"} Nov 25 11:08:55 crc kubenswrapper[4821]: I1125 11:08:55.354177 4821 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 11:08:56 crc kubenswrapper[4821]: I1125 11:08:56.369638 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t6688" event={"ID":"af82c304-6d71-4420-840a-c299a888d85a","Type":"ContainerStarted","Data":"df81b569696b8ea77cfa371c18dc8642d22f86692ca996a5fb71062ec9670ee5"} Nov 25 11:08:57 crc kubenswrapper[4821]: I1125 11:08:57.380051 4821 generic.go:334] "Generic (PLEG): container finished" podID="af82c304-6d71-4420-840a-c299a888d85a" containerID="df81b569696b8ea77cfa371c18dc8642d22f86692ca996a5fb71062ec9670ee5" exitCode=0 Nov 25 11:08:57 crc kubenswrapper[4821]: I1125 11:08:57.380190 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t6688" event={"ID":"af82c304-6d71-4420-840a-c299a888d85a","Type":"ContainerDied","Data":"df81b569696b8ea77cfa371c18dc8642d22f86692ca996a5fb71062ec9670ee5"} Nov 25 11:08:58 crc kubenswrapper[4821]: I1125 11:08:58.389715 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t6688" event={"ID":"af82c304-6d71-4420-840a-c299a888d85a","Type":"ContainerStarted","Data":"32602e533073b25a5054e44bbd63f2a28eb06848389073c2c8c1fb06c3c8a8ec"} Nov 25 11:08:58 crc kubenswrapper[4821]: I1125 11:08:58.417786 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-t6688" podStartSLOduration=2.99452386 podStartE2EDuration="5.417765839s" podCreationTimestamp="2025-11-25 11:08:53 +0000 UTC" firstStartedPulling="2025-11-25 11:08:55.353974468 +0000 UTC m=+2205.890294315" lastFinishedPulling="2025-11-25 11:08:57.777216447 +0000 UTC m=+2208.313536294" observedRunningTime="2025-11-25 11:08:58.408481571 +0000 UTC m=+2208.944801428" watchObservedRunningTime="2025-11-25 11:08:58.417765839 +0000 UTC m=+2208.954085696" Nov 25 11:09:03 crc kubenswrapper[4821]: I1125 11:09:03.638714 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-t6688" Nov 25 11:09:03 crc kubenswrapper[4821]: I1125 11:09:03.639467 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-t6688" Nov 25 11:09:03 crc kubenswrapper[4821]: I1125 11:09:03.699271 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-t6688" Nov 25 11:09:04 crc kubenswrapper[4821]: I1125 11:09:04.485020 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-t6688" Nov 25 11:09:04 crc kubenswrapper[4821]: I1125 11:09:04.534298 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-t6688"] Nov 25 11:09:06 crc kubenswrapper[4821]: I1125 11:09:06.456941 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-t6688" podUID="af82c304-6d71-4420-840a-c299a888d85a" containerName="registry-server" containerID="cri-o://32602e533073b25a5054e44bbd63f2a28eb06848389073c2c8c1fb06c3c8a8ec" gracePeriod=2 Nov 25 11:09:06 crc kubenswrapper[4821]: I1125 11:09:06.916279 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t6688" Nov 25 11:09:07 crc kubenswrapper[4821]: I1125 11:09:07.033323 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af82c304-6d71-4420-840a-c299a888d85a-utilities\") pod \"af82c304-6d71-4420-840a-c299a888d85a\" (UID: \"af82c304-6d71-4420-840a-c299a888d85a\") " Nov 25 11:09:07 crc kubenswrapper[4821]: I1125 11:09:07.033422 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af82c304-6d71-4420-840a-c299a888d85a-catalog-content\") pod \"af82c304-6d71-4420-840a-c299a888d85a\" (UID: \"af82c304-6d71-4420-840a-c299a888d85a\") " Nov 25 11:09:07 crc kubenswrapper[4821]: I1125 11:09:07.033585 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5wh5t\" (UniqueName: \"kubernetes.io/projected/af82c304-6d71-4420-840a-c299a888d85a-kube-api-access-5wh5t\") pod \"af82c304-6d71-4420-840a-c299a888d85a\" (UID: \"af82c304-6d71-4420-840a-c299a888d85a\") " Nov 25 11:09:07 crc kubenswrapper[4821]: I1125 11:09:07.035013 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af82c304-6d71-4420-840a-c299a888d85a-utilities" (OuterVolumeSpecName: "utilities") pod "af82c304-6d71-4420-840a-c299a888d85a" (UID: "af82c304-6d71-4420-840a-c299a888d85a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:09:07 crc kubenswrapper[4821]: I1125 11:09:07.039359 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af82c304-6d71-4420-840a-c299a888d85a-kube-api-access-5wh5t" (OuterVolumeSpecName: "kube-api-access-5wh5t") pod "af82c304-6d71-4420-840a-c299a888d85a" (UID: "af82c304-6d71-4420-840a-c299a888d85a"). InnerVolumeSpecName "kube-api-access-5wh5t". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:09:07 crc kubenswrapper[4821]: I1125 11:09:07.135083 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5wh5t\" (UniqueName: \"kubernetes.io/projected/af82c304-6d71-4420-840a-c299a888d85a-kube-api-access-5wh5t\") on node \"crc\" DevicePath \"\"" Nov 25 11:09:07 crc kubenswrapper[4821]: I1125 11:09:07.135117 4821 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af82c304-6d71-4420-840a-c299a888d85a-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 11:09:07 crc kubenswrapper[4821]: I1125 11:09:07.400880 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af82c304-6d71-4420-840a-c299a888d85a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "af82c304-6d71-4420-840a-c299a888d85a" (UID: "af82c304-6d71-4420-840a-c299a888d85a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:09:07 crc kubenswrapper[4821]: I1125 11:09:07.438989 4821 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af82c304-6d71-4420-840a-c299a888d85a-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 11:09:07 crc kubenswrapper[4821]: I1125 11:09:07.466403 4821 generic.go:334] "Generic (PLEG): container finished" podID="af82c304-6d71-4420-840a-c299a888d85a" containerID="32602e533073b25a5054e44bbd63f2a28eb06848389073c2c8c1fb06c3c8a8ec" exitCode=0 Nov 25 11:09:07 crc kubenswrapper[4821]: I1125 11:09:07.466480 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t6688" Nov 25 11:09:07 crc kubenswrapper[4821]: I1125 11:09:07.466474 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t6688" event={"ID":"af82c304-6d71-4420-840a-c299a888d85a","Type":"ContainerDied","Data":"32602e533073b25a5054e44bbd63f2a28eb06848389073c2c8c1fb06c3c8a8ec"} Nov 25 11:09:07 crc kubenswrapper[4821]: I1125 11:09:07.466531 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t6688" event={"ID":"af82c304-6d71-4420-840a-c299a888d85a","Type":"ContainerDied","Data":"53427aa5a42d3e02aba2d4b1420b52c1e781378ab5b90958a450e604a1ef2f58"} Nov 25 11:09:07 crc kubenswrapper[4821]: I1125 11:09:07.466554 4821 scope.go:117] "RemoveContainer" containerID="32602e533073b25a5054e44bbd63f2a28eb06848389073c2c8c1fb06c3c8a8ec" Nov 25 11:09:07 crc kubenswrapper[4821]: I1125 11:09:07.485929 4821 scope.go:117] "RemoveContainer" containerID="df81b569696b8ea77cfa371c18dc8642d22f86692ca996a5fb71062ec9670ee5" Nov 25 11:09:07 crc kubenswrapper[4821]: I1125 11:09:07.501707 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-t6688"] Nov 25 11:09:07 crc kubenswrapper[4821]: I1125 11:09:07.508583 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-t6688"] Nov 25 11:09:07 crc kubenswrapper[4821]: I1125 11:09:07.526090 4821 scope.go:117] "RemoveContainer" containerID="a105ab6a3e6ea159d604ce1df2fa1f18fe22324605cf0e8b72b7e77c80651673" Nov 25 11:09:07 crc kubenswrapper[4821]: I1125 11:09:07.559281 4821 scope.go:117] "RemoveContainer" containerID="32602e533073b25a5054e44bbd63f2a28eb06848389073c2c8c1fb06c3c8a8ec" Nov 25 11:09:07 crc kubenswrapper[4821]: E1125 11:09:07.559672 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"32602e533073b25a5054e44bbd63f2a28eb06848389073c2c8c1fb06c3c8a8ec\": container with ID starting with 32602e533073b25a5054e44bbd63f2a28eb06848389073c2c8c1fb06c3c8a8ec not found: ID does not exist" containerID="32602e533073b25a5054e44bbd63f2a28eb06848389073c2c8c1fb06c3c8a8ec" Nov 25 11:09:07 crc kubenswrapper[4821]: I1125 11:09:07.559718 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32602e533073b25a5054e44bbd63f2a28eb06848389073c2c8c1fb06c3c8a8ec"} err="failed to get container status \"32602e533073b25a5054e44bbd63f2a28eb06848389073c2c8c1fb06c3c8a8ec\": rpc error: code = NotFound desc = could not find container \"32602e533073b25a5054e44bbd63f2a28eb06848389073c2c8c1fb06c3c8a8ec\": container with ID starting with 32602e533073b25a5054e44bbd63f2a28eb06848389073c2c8c1fb06c3c8a8ec not found: ID does not exist" Nov 25 11:09:07 crc kubenswrapper[4821]: I1125 11:09:07.559746 4821 scope.go:117] "RemoveContainer" containerID="df81b569696b8ea77cfa371c18dc8642d22f86692ca996a5fb71062ec9670ee5" Nov 25 11:09:07 crc kubenswrapper[4821]: E1125 11:09:07.560291 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df81b569696b8ea77cfa371c18dc8642d22f86692ca996a5fb71062ec9670ee5\": container with ID starting with df81b569696b8ea77cfa371c18dc8642d22f86692ca996a5fb71062ec9670ee5 not found: ID does not exist" containerID="df81b569696b8ea77cfa371c18dc8642d22f86692ca996a5fb71062ec9670ee5" Nov 25 11:09:07 crc kubenswrapper[4821]: I1125 11:09:07.560343 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df81b569696b8ea77cfa371c18dc8642d22f86692ca996a5fb71062ec9670ee5"} err="failed to get container status \"df81b569696b8ea77cfa371c18dc8642d22f86692ca996a5fb71062ec9670ee5\": rpc error: code = NotFound desc = could not find container \"df81b569696b8ea77cfa371c18dc8642d22f86692ca996a5fb71062ec9670ee5\": container with ID starting with df81b569696b8ea77cfa371c18dc8642d22f86692ca996a5fb71062ec9670ee5 not found: ID does not exist" Nov 25 11:09:07 crc kubenswrapper[4821]: I1125 11:09:07.560377 4821 scope.go:117] "RemoveContainer" containerID="a105ab6a3e6ea159d604ce1df2fa1f18fe22324605cf0e8b72b7e77c80651673" Nov 25 11:09:07 crc kubenswrapper[4821]: E1125 11:09:07.560673 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a105ab6a3e6ea159d604ce1df2fa1f18fe22324605cf0e8b72b7e77c80651673\": container with ID starting with a105ab6a3e6ea159d604ce1df2fa1f18fe22324605cf0e8b72b7e77c80651673 not found: ID does not exist" containerID="a105ab6a3e6ea159d604ce1df2fa1f18fe22324605cf0e8b72b7e77c80651673" Nov 25 11:09:07 crc kubenswrapper[4821]: I1125 11:09:07.560710 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a105ab6a3e6ea159d604ce1df2fa1f18fe22324605cf0e8b72b7e77c80651673"} err="failed to get container status \"a105ab6a3e6ea159d604ce1df2fa1f18fe22324605cf0e8b72b7e77c80651673\": rpc error: code = NotFound desc = could not find container \"a105ab6a3e6ea159d604ce1df2fa1f18fe22324605cf0e8b72b7e77c80651673\": container with ID starting with a105ab6a3e6ea159d604ce1df2fa1f18fe22324605cf0e8b72b7e77c80651673 not found: ID does not exist" Nov 25 11:09:08 crc kubenswrapper[4821]: I1125 11:09:08.131509 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af82c304-6d71-4420-840a-c299a888d85a" path="/var/lib/kubelet/pods/af82c304-6d71-4420-840a-c299a888d85a/volumes" Nov 25 11:10:04 crc kubenswrapper[4821]: I1125 11:10:04.714706 4821 patch_prober.go:28] interesting pod/machine-config-daemon-2krbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:10:04 crc kubenswrapper[4821]: I1125 11:10:04.715423 4821 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:10:34 crc kubenswrapper[4821]: I1125 11:10:34.714819 4821 patch_prober.go:28] interesting pod/machine-config-daemon-2krbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:10:34 crc kubenswrapper[4821]: I1125 11:10:34.715476 4821 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:11:04 crc kubenswrapper[4821]: I1125 11:11:04.715332 4821 patch_prober.go:28] interesting pod/machine-config-daemon-2krbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:11:04 crc kubenswrapper[4821]: I1125 11:11:04.715855 4821 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:11:04 crc kubenswrapper[4821]: I1125 11:11:04.715896 4821 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" Nov 25 11:11:04 crc kubenswrapper[4821]: I1125 11:11:04.716569 4821 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"61010995905bb7c40fff6417cf7e0b11d584b8a9a68f6fccf71aefe231f45291"} pod="openshift-machine-config-operator/machine-config-daemon-2krbf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 11:11:04 crc kubenswrapper[4821]: I1125 11:11:04.716620 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerName="machine-config-daemon" containerID="cri-o://61010995905bb7c40fff6417cf7e0b11d584b8a9a68f6fccf71aefe231f45291" gracePeriod=600 Nov 25 11:11:04 crc kubenswrapper[4821]: E1125 11:11:04.866748 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:11:05 crc kubenswrapper[4821]: I1125 11:11:05.472964 4821 generic.go:334] "Generic (PLEG): container finished" podID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerID="61010995905bb7c40fff6417cf7e0b11d584b8a9a68f6fccf71aefe231f45291" exitCode=0 Nov 25 11:11:05 crc kubenswrapper[4821]: I1125 11:11:05.473021 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" event={"ID":"5f948b87-ac86-4de6-ad64-c2ef947f84d4","Type":"ContainerDied","Data":"61010995905bb7c40fff6417cf7e0b11d584b8a9a68f6fccf71aefe231f45291"} Nov 25 11:11:05 crc kubenswrapper[4821]: I1125 11:11:05.473064 4821 scope.go:117] "RemoveContainer" containerID="67625e9eb479a7295b3be108a852f5678838cf58bdadd2502f30ab844bb33b89" Nov 25 11:11:05 crc kubenswrapper[4821]: I1125 11:11:05.473668 4821 scope.go:117] "RemoveContainer" containerID="61010995905bb7c40fff6417cf7e0b11d584b8a9a68f6fccf71aefe231f45291" Nov 25 11:11:05 crc kubenswrapper[4821]: E1125 11:11:05.473915 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:11:19 crc kubenswrapper[4821]: I1125 11:11:19.115390 4821 scope.go:117] "RemoveContainer" containerID="61010995905bb7c40fff6417cf7e0b11d584b8a9a68f6fccf71aefe231f45291" Nov 25 11:11:19 crc kubenswrapper[4821]: E1125 11:11:19.116417 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:11:23 crc kubenswrapper[4821]: I1125 11:11:23.628936 4821 generic.go:334] "Generic (PLEG): container finished" podID="0ee4d7f9-162b-4882-b189-13c53bcdc274" containerID="94388636806b204fde9e3bf85b805d500b31f590fc25a3b91d46ab2294fde1cf" exitCode=0 Nov 25 11:11:23 crc kubenswrapper[4821]: I1125 11:11:23.629051 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9w8v8" event={"ID":"0ee4d7f9-162b-4882-b189-13c53bcdc274","Type":"ContainerDied","Data":"94388636806b204fde9e3bf85b805d500b31f590fc25a3b91d46ab2294fde1cf"} Nov 25 11:11:25 crc kubenswrapper[4821]: I1125 11:11:25.090987 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9w8v8" Nov 25 11:11:25 crc kubenswrapper[4821]: I1125 11:11:25.144707 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0ee4d7f9-162b-4882-b189-13c53bcdc274-inventory\") pod \"0ee4d7f9-162b-4882-b189-13c53bcdc274\" (UID: \"0ee4d7f9-162b-4882-b189-13c53bcdc274\") " Nov 25 11:11:25 crc kubenswrapper[4821]: I1125 11:11:25.144761 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0ee4d7f9-162b-4882-b189-13c53bcdc274-ssh-key\") pod \"0ee4d7f9-162b-4882-b189-13c53bcdc274\" (UID: \"0ee4d7f9-162b-4882-b189-13c53bcdc274\") " Nov 25 11:11:25 crc kubenswrapper[4821]: I1125 11:11:25.144804 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ee4d7f9-162b-4882-b189-13c53bcdc274-libvirt-combined-ca-bundle\") pod \"0ee4d7f9-162b-4882-b189-13c53bcdc274\" (UID: \"0ee4d7f9-162b-4882-b189-13c53bcdc274\") " Nov 25 11:11:25 crc kubenswrapper[4821]: I1125 11:11:25.144842 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r7j8r\" (UniqueName: \"kubernetes.io/projected/0ee4d7f9-162b-4882-b189-13c53bcdc274-kube-api-access-r7j8r\") pod \"0ee4d7f9-162b-4882-b189-13c53bcdc274\" (UID: \"0ee4d7f9-162b-4882-b189-13c53bcdc274\") " Nov 25 11:11:25 crc kubenswrapper[4821]: I1125 11:11:25.144930 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/0ee4d7f9-162b-4882-b189-13c53bcdc274-libvirt-secret-0\") pod \"0ee4d7f9-162b-4882-b189-13c53bcdc274\" (UID: \"0ee4d7f9-162b-4882-b189-13c53bcdc274\") " Nov 25 11:11:25 crc kubenswrapper[4821]: I1125 11:11:25.151462 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ee4d7f9-162b-4882-b189-13c53bcdc274-kube-api-access-r7j8r" (OuterVolumeSpecName: "kube-api-access-r7j8r") pod "0ee4d7f9-162b-4882-b189-13c53bcdc274" (UID: "0ee4d7f9-162b-4882-b189-13c53bcdc274"). InnerVolumeSpecName "kube-api-access-r7j8r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:11:25 crc kubenswrapper[4821]: I1125 11:11:25.151691 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ee4d7f9-162b-4882-b189-13c53bcdc274-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "0ee4d7f9-162b-4882-b189-13c53bcdc274" (UID: "0ee4d7f9-162b-4882-b189-13c53bcdc274"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:11:25 crc kubenswrapper[4821]: I1125 11:11:25.173867 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ee4d7f9-162b-4882-b189-13c53bcdc274-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0ee4d7f9-162b-4882-b189-13c53bcdc274" (UID: "0ee4d7f9-162b-4882-b189-13c53bcdc274"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:11:25 crc kubenswrapper[4821]: I1125 11:11:25.175759 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ee4d7f9-162b-4882-b189-13c53bcdc274-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "0ee4d7f9-162b-4882-b189-13c53bcdc274" (UID: "0ee4d7f9-162b-4882-b189-13c53bcdc274"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:11:25 crc kubenswrapper[4821]: I1125 11:11:25.179318 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ee4d7f9-162b-4882-b189-13c53bcdc274-inventory" (OuterVolumeSpecName: "inventory") pod "0ee4d7f9-162b-4882-b189-13c53bcdc274" (UID: "0ee4d7f9-162b-4882-b189-13c53bcdc274"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:11:25 crc kubenswrapper[4821]: I1125 11:11:25.247489 4821 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0ee4d7f9-162b-4882-b189-13c53bcdc274-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 11:11:25 crc kubenswrapper[4821]: I1125 11:11:25.247527 4821 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0ee4d7f9-162b-4882-b189-13c53bcdc274-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 11:11:25 crc kubenswrapper[4821]: I1125 11:11:25.247537 4821 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ee4d7f9-162b-4882-b189-13c53bcdc274-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:11:25 crc kubenswrapper[4821]: I1125 11:11:25.247553 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r7j8r\" (UniqueName: \"kubernetes.io/projected/0ee4d7f9-162b-4882-b189-13c53bcdc274-kube-api-access-r7j8r\") on node \"crc\" DevicePath \"\"" Nov 25 11:11:25 crc kubenswrapper[4821]: I1125 11:11:25.247565 4821 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/0ee4d7f9-162b-4882-b189-13c53bcdc274-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Nov 25 11:11:25 crc kubenswrapper[4821]: I1125 11:11:25.650644 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9w8v8" Nov 25 11:11:25 crc kubenswrapper[4821]: I1125 11:11:25.651403 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9w8v8" event={"ID":"0ee4d7f9-162b-4882-b189-13c53bcdc274","Type":"ContainerDied","Data":"313214dcc72f79983d312f27b9a9eb73bb5a1d376206544ebba551b0902efe58"} Nov 25 11:11:25 crc kubenswrapper[4821]: I1125 11:11:25.651553 4821 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="313214dcc72f79983d312f27b9a9eb73bb5a1d376206544ebba551b0902efe58" Nov 25 11:11:25 crc kubenswrapper[4821]: I1125 11:11:25.748368 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-mt9vk"] Nov 25 11:11:25 crc kubenswrapper[4821]: E1125 11:11:25.748905 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af82c304-6d71-4420-840a-c299a888d85a" containerName="registry-server" Nov 25 11:11:25 crc kubenswrapper[4821]: I1125 11:11:25.748923 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="af82c304-6d71-4420-840a-c299a888d85a" containerName="registry-server" Nov 25 11:11:25 crc kubenswrapper[4821]: E1125 11:11:25.748940 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af82c304-6d71-4420-840a-c299a888d85a" containerName="extract-utilities" Nov 25 11:11:25 crc kubenswrapper[4821]: I1125 11:11:25.748946 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="af82c304-6d71-4420-840a-c299a888d85a" containerName="extract-utilities" Nov 25 11:11:25 crc kubenswrapper[4821]: E1125 11:11:25.748954 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ee4d7f9-162b-4882-b189-13c53bcdc274" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 25 11:11:25 crc kubenswrapper[4821]: I1125 11:11:25.748978 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ee4d7f9-162b-4882-b189-13c53bcdc274" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 25 11:11:25 crc kubenswrapper[4821]: E1125 11:11:25.748996 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af82c304-6d71-4420-840a-c299a888d85a" containerName="extract-content" Nov 25 11:11:25 crc kubenswrapper[4821]: I1125 11:11:25.749002 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="af82c304-6d71-4420-840a-c299a888d85a" containerName="extract-content" Nov 25 11:11:25 crc kubenswrapper[4821]: I1125 11:11:25.749267 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="af82c304-6d71-4420-840a-c299a888d85a" containerName="registry-server" Nov 25 11:11:25 crc kubenswrapper[4821]: I1125 11:11:25.749291 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ee4d7f9-162b-4882-b189-13c53bcdc274" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 25 11:11:25 crc kubenswrapper[4821]: I1125 11:11:25.750152 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mt9vk" Nov 25 11:11:25 crc kubenswrapper[4821]: I1125 11:11:25.753057 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 11:11:25 crc kubenswrapper[4821]: I1125 11:11:25.753799 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Nov 25 11:11:25 crc kubenswrapper[4821]: I1125 11:11:25.754131 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-z764x" Nov 25 11:11:25 crc kubenswrapper[4821]: I1125 11:11:25.754197 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Nov 25 11:11:25 crc kubenswrapper[4821]: I1125 11:11:25.754242 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Nov 25 11:11:25 crc kubenswrapper[4821]: I1125 11:11:25.754641 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 11:11:25 crc kubenswrapper[4821]: I1125 11:11:25.756650 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 11:11:25 crc kubenswrapper[4821]: I1125 11:11:25.770570 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-mt9vk"] Nov 25 11:11:25 crc kubenswrapper[4821]: I1125 11:11:25.859469 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40682cac-f449-4465-9ad3-3bfa28b3800d-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mt9vk\" (UID: \"40682cac-f449-4465-9ad3-3bfa28b3800d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mt9vk" Nov 25 11:11:25 crc kubenswrapper[4821]: I1125 11:11:25.859529 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/40682cac-f449-4465-9ad3-3bfa28b3800d-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mt9vk\" (UID: \"40682cac-f449-4465-9ad3-3bfa28b3800d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mt9vk" Nov 25 11:11:25 crc kubenswrapper[4821]: I1125 11:11:25.859576 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/40682cac-f449-4465-9ad3-3bfa28b3800d-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mt9vk\" (UID: \"40682cac-f449-4465-9ad3-3bfa28b3800d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mt9vk" Nov 25 11:11:25 crc kubenswrapper[4821]: I1125 11:11:25.859601 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/40682cac-f449-4465-9ad3-3bfa28b3800d-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mt9vk\" (UID: \"40682cac-f449-4465-9ad3-3bfa28b3800d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mt9vk" Nov 25 11:11:25 crc kubenswrapper[4821]: I1125 11:11:25.859738 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9xrb6\" (UniqueName: \"kubernetes.io/projected/40682cac-f449-4465-9ad3-3bfa28b3800d-kube-api-access-9xrb6\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mt9vk\" (UID: \"40682cac-f449-4465-9ad3-3bfa28b3800d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mt9vk" Nov 25 11:11:25 crc kubenswrapper[4821]: I1125 11:11:25.859802 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/40682cac-f449-4465-9ad3-3bfa28b3800d-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mt9vk\" (UID: \"40682cac-f449-4465-9ad3-3bfa28b3800d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mt9vk" Nov 25 11:11:25 crc kubenswrapper[4821]: I1125 11:11:25.860030 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/40682cac-f449-4465-9ad3-3bfa28b3800d-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mt9vk\" (UID: \"40682cac-f449-4465-9ad3-3bfa28b3800d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mt9vk" Nov 25 11:11:25 crc kubenswrapper[4821]: I1125 11:11:25.860076 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/40682cac-f449-4465-9ad3-3bfa28b3800d-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mt9vk\" (UID: \"40682cac-f449-4465-9ad3-3bfa28b3800d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mt9vk" Nov 25 11:11:25 crc kubenswrapper[4821]: I1125 11:11:25.860108 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/40682cac-f449-4465-9ad3-3bfa28b3800d-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mt9vk\" (UID: \"40682cac-f449-4465-9ad3-3bfa28b3800d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mt9vk" Nov 25 11:11:25 crc kubenswrapper[4821]: I1125 11:11:25.962274 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40682cac-f449-4465-9ad3-3bfa28b3800d-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mt9vk\" (UID: \"40682cac-f449-4465-9ad3-3bfa28b3800d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mt9vk" Nov 25 11:11:25 crc kubenswrapper[4821]: I1125 11:11:25.962647 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/40682cac-f449-4465-9ad3-3bfa28b3800d-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mt9vk\" (UID: \"40682cac-f449-4465-9ad3-3bfa28b3800d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mt9vk" Nov 25 11:11:25 crc kubenswrapper[4821]: I1125 11:11:25.962707 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/40682cac-f449-4465-9ad3-3bfa28b3800d-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mt9vk\" (UID: \"40682cac-f449-4465-9ad3-3bfa28b3800d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mt9vk" Nov 25 11:11:25 crc kubenswrapper[4821]: I1125 11:11:25.962740 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/40682cac-f449-4465-9ad3-3bfa28b3800d-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mt9vk\" (UID: \"40682cac-f449-4465-9ad3-3bfa28b3800d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mt9vk" Nov 25 11:11:25 crc kubenswrapper[4821]: I1125 11:11:25.962769 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9xrb6\" (UniqueName: \"kubernetes.io/projected/40682cac-f449-4465-9ad3-3bfa28b3800d-kube-api-access-9xrb6\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mt9vk\" (UID: \"40682cac-f449-4465-9ad3-3bfa28b3800d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mt9vk" Nov 25 11:11:25 crc kubenswrapper[4821]: I1125 11:11:25.962795 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/40682cac-f449-4465-9ad3-3bfa28b3800d-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mt9vk\" (UID: \"40682cac-f449-4465-9ad3-3bfa28b3800d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mt9vk" Nov 25 11:11:25 crc kubenswrapper[4821]: I1125 11:11:25.962884 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/40682cac-f449-4465-9ad3-3bfa28b3800d-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mt9vk\" (UID: \"40682cac-f449-4465-9ad3-3bfa28b3800d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mt9vk" Nov 25 11:11:25 crc kubenswrapper[4821]: I1125 11:11:25.962911 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/40682cac-f449-4465-9ad3-3bfa28b3800d-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mt9vk\" (UID: \"40682cac-f449-4465-9ad3-3bfa28b3800d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mt9vk" Nov 25 11:11:25 crc kubenswrapper[4821]: I1125 11:11:25.962933 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/40682cac-f449-4465-9ad3-3bfa28b3800d-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mt9vk\" (UID: \"40682cac-f449-4465-9ad3-3bfa28b3800d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mt9vk" Nov 25 11:11:25 crc kubenswrapper[4821]: I1125 11:11:25.963997 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/40682cac-f449-4465-9ad3-3bfa28b3800d-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mt9vk\" (UID: \"40682cac-f449-4465-9ad3-3bfa28b3800d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mt9vk" Nov 25 11:11:25 crc kubenswrapper[4821]: I1125 11:11:25.966758 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/40682cac-f449-4465-9ad3-3bfa28b3800d-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mt9vk\" (UID: \"40682cac-f449-4465-9ad3-3bfa28b3800d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mt9vk" Nov 25 11:11:25 crc kubenswrapper[4821]: I1125 11:11:25.967072 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/40682cac-f449-4465-9ad3-3bfa28b3800d-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mt9vk\" (UID: \"40682cac-f449-4465-9ad3-3bfa28b3800d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mt9vk" Nov 25 11:11:25 crc kubenswrapper[4821]: I1125 11:11:25.967343 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40682cac-f449-4465-9ad3-3bfa28b3800d-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mt9vk\" (UID: \"40682cac-f449-4465-9ad3-3bfa28b3800d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mt9vk" Nov 25 11:11:25 crc kubenswrapper[4821]: I1125 11:11:25.967429 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/40682cac-f449-4465-9ad3-3bfa28b3800d-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mt9vk\" (UID: \"40682cac-f449-4465-9ad3-3bfa28b3800d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mt9vk" Nov 25 11:11:25 crc kubenswrapper[4821]: I1125 11:11:25.967776 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/40682cac-f449-4465-9ad3-3bfa28b3800d-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mt9vk\" (UID: \"40682cac-f449-4465-9ad3-3bfa28b3800d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mt9vk" Nov 25 11:11:25 crc kubenswrapper[4821]: I1125 11:11:25.968318 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/40682cac-f449-4465-9ad3-3bfa28b3800d-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mt9vk\" (UID: \"40682cac-f449-4465-9ad3-3bfa28b3800d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mt9vk" Nov 25 11:11:25 crc kubenswrapper[4821]: I1125 11:11:25.968768 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/40682cac-f449-4465-9ad3-3bfa28b3800d-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mt9vk\" (UID: \"40682cac-f449-4465-9ad3-3bfa28b3800d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mt9vk" Nov 25 11:11:25 crc kubenswrapper[4821]: I1125 11:11:25.981785 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9xrb6\" (UniqueName: \"kubernetes.io/projected/40682cac-f449-4465-9ad3-3bfa28b3800d-kube-api-access-9xrb6\") pod \"nova-edpm-deployment-openstack-edpm-ipam-mt9vk\" (UID: \"40682cac-f449-4465-9ad3-3bfa28b3800d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mt9vk" Nov 25 11:11:26 crc kubenswrapper[4821]: I1125 11:11:26.066667 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mt9vk" Nov 25 11:11:26 crc kubenswrapper[4821]: I1125 11:11:26.596732 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-mt9vk"] Nov 25 11:11:26 crc kubenswrapper[4821]: I1125 11:11:26.660091 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mt9vk" event={"ID":"40682cac-f449-4465-9ad3-3bfa28b3800d","Type":"ContainerStarted","Data":"6dbfe0d593c06cbf5fe45a675965cd46a5035d296e4f25bcb7654c5f7b3b1b31"} Nov 25 11:11:27 crc kubenswrapper[4821]: I1125 11:11:27.669579 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mt9vk" event={"ID":"40682cac-f449-4465-9ad3-3bfa28b3800d","Type":"ContainerStarted","Data":"f2cd1db11e19e48bc051f4ae1b8efe7ddf5f20b7b64c39a3e32f6aba00aa1a72"} Nov 25 11:11:27 crc kubenswrapper[4821]: I1125 11:11:27.689739 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mt9vk" podStartSLOduration=2.255941061 podStartE2EDuration="2.689722848s" podCreationTimestamp="2025-11-25 11:11:25 +0000 UTC" firstStartedPulling="2025-11-25 11:11:26.605929509 +0000 UTC m=+2357.142249356" lastFinishedPulling="2025-11-25 11:11:27.039711296 +0000 UTC m=+2357.576031143" observedRunningTime="2025-11-25 11:11:27.689450892 +0000 UTC m=+2358.225770739" watchObservedRunningTime="2025-11-25 11:11:27.689722848 +0000 UTC m=+2358.226042695" Nov 25 11:11:30 crc kubenswrapper[4821]: I1125 11:11:30.122331 4821 scope.go:117] "RemoveContainer" containerID="61010995905bb7c40fff6417cf7e0b11d584b8a9a68f6fccf71aefe231f45291" Nov 25 11:11:30 crc kubenswrapper[4821]: E1125 11:11:30.123130 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:11:43 crc kubenswrapper[4821]: I1125 11:11:43.115147 4821 scope.go:117] "RemoveContainer" containerID="61010995905bb7c40fff6417cf7e0b11d584b8a9a68f6fccf71aefe231f45291" Nov 25 11:11:43 crc kubenswrapper[4821]: E1125 11:11:43.115875 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:11:57 crc kubenswrapper[4821]: I1125 11:11:57.115294 4821 scope.go:117] "RemoveContainer" containerID="61010995905bb7c40fff6417cf7e0b11d584b8a9a68f6fccf71aefe231f45291" Nov 25 11:11:57 crc kubenswrapper[4821]: E1125 11:11:57.116054 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:12:09 crc kubenswrapper[4821]: I1125 11:12:09.114008 4821 scope.go:117] "RemoveContainer" containerID="61010995905bb7c40fff6417cf7e0b11d584b8a9a68f6fccf71aefe231f45291" Nov 25 11:12:09 crc kubenswrapper[4821]: E1125 11:12:09.117017 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:12:21 crc kubenswrapper[4821]: I1125 11:12:21.115979 4821 scope.go:117] "RemoveContainer" containerID="61010995905bb7c40fff6417cf7e0b11d584b8a9a68f6fccf71aefe231f45291" Nov 25 11:12:21 crc kubenswrapper[4821]: E1125 11:12:21.117052 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:12:36 crc kubenswrapper[4821]: I1125 11:12:36.114627 4821 scope.go:117] "RemoveContainer" containerID="61010995905bb7c40fff6417cf7e0b11d584b8a9a68f6fccf71aefe231f45291" Nov 25 11:12:36 crc kubenswrapper[4821]: E1125 11:12:36.115420 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:12:51 crc kubenswrapper[4821]: I1125 11:12:51.114428 4821 scope.go:117] "RemoveContainer" containerID="61010995905bb7c40fff6417cf7e0b11d584b8a9a68f6fccf71aefe231f45291" Nov 25 11:12:51 crc kubenswrapper[4821]: E1125 11:12:51.115211 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:13:03 crc kubenswrapper[4821]: I1125 11:13:03.118608 4821 scope.go:117] "RemoveContainer" containerID="61010995905bb7c40fff6417cf7e0b11d584b8a9a68f6fccf71aefe231f45291" Nov 25 11:13:03 crc kubenswrapper[4821]: E1125 11:13:03.138272 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:13:15 crc kubenswrapper[4821]: I1125 11:13:15.115496 4821 scope.go:117] "RemoveContainer" containerID="61010995905bb7c40fff6417cf7e0b11d584b8a9a68f6fccf71aefe231f45291" Nov 25 11:13:15 crc kubenswrapper[4821]: E1125 11:13:15.116259 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:13:26 crc kubenswrapper[4821]: I1125 11:13:26.114204 4821 scope.go:117] "RemoveContainer" containerID="61010995905bb7c40fff6417cf7e0b11d584b8a9a68f6fccf71aefe231f45291" Nov 25 11:13:26 crc kubenswrapper[4821]: E1125 11:13:26.114964 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:13:39 crc kubenswrapper[4821]: I1125 11:13:39.114723 4821 scope.go:117] "RemoveContainer" containerID="61010995905bb7c40fff6417cf7e0b11d584b8a9a68f6fccf71aefe231f45291" Nov 25 11:13:39 crc kubenswrapper[4821]: E1125 11:13:39.115547 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:13:53 crc kubenswrapper[4821]: I1125 11:13:53.115340 4821 scope.go:117] "RemoveContainer" containerID="61010995905bb7c40fff6417cf7e0b11d584b8a9a68f6fccf71aefe231f45291" Nov 25 11:13:53 crc kubenswrapper[4821]: E1125 11:13:53.116381 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:14:05 crc kubenswrapper[4821]: I1125 11:14:05.114417 4821 scope.go:117] "RemoveContainer" containerID="61010995905bb7c40fff6417cf7e0b11d584b8a9a68f6fccf71aefe231f45291" Nov 25 11:14:05 crc kubenswrapper[4821]: E1125 11:14:05.115192 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:14:13 crc kubenswrapper[4821]: I1125 11:14:13.076665 4821 generic.go:334] "Generic (PLEG): container finished" podID="40682cac-f449-4465-9ad3-3bfa28b3800d" containerID="f2cd1db11e19e48bc051f4ae1b8efe7ddf5f20b7b64c39a3e32f6aba00aa1a72" exitCode=0 Nov 25 11:14:13 crc kubenswrapper[4821]: I1125 11:14:13.076726 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mt9vk" event={"ID":"40682cac-f449-4465-9ad3-3bfa28b3800d","Type":"ContainerDied","Data":"f2cd1db11e19e48bc051f4ae1b8efe7ddf5f20b7b64c39a3e32f6aba00aa1a72"} Nov 25 11:14:14 crc kubenswrapper[4821]: I1125 11:14:14.536992 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mt9vk" Nov 25 11:14:14 crc kubenswrapper[4821]: I1125 11:14:14.739947 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/40682cac-f449-4465-9ad3-3bfa28b3800d-ssh-key\") pod \"40682cac-f449-4465-9ad3-3bfa28b3800d\" (UID: \"40682cac-f449-4465-9ad3-3bfa28b3800d\") " Nov 25 11:14:14 crc kubenswrapper[4821]: I1125 11:14:14.740303 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/40682cac-f449-4465-9ad3-3bfa28b3800d-nova-cell1-compute-config-1\") pod \"40682cac-f449-4465-9ad3-3bfa28b3800d\" (UID: \"40682cac-f449-4465-9ad3-3bfa28b3800d\") " Nov 25 11:14:14 crc kubenswrapper[4821]: I1125 11:14:14.740382 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/40682cac-f449-4465-9ad3-3bfa28b3800d-inventory\") pod \"40682cac-f449-4465-9ad3-3bfa28b3800d\" (UID: \"40682cac-f449-4465-9ad3-3bfa28b3800d\") " Nov 25 11:14:14 crc kubenswrapper[4821]: I1125 11:14:14.740415 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40682cac-f449-4465-9ad3-3bfa28b3800d-nova-combined-ca-bundle\") pod \"40682cac-f449-4465-9ad3-3bfa28b3800d\" (UID: \"40682cac-f449-4465-9ad3-3bfa28b3800d\") " Nov 25 11:14:14 crc kubenswrapper[4821]: I1125 11:14:14.740463 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/40682cac-f449-4465-9ad3-3bfa28b3800d-nova-cell1-compute-config-0\") pod \"40682cac-f449-4465-9ad3-3bfa28b3800d\" (UID: \"40682cac-f449-4465-9ad3-3bfa28b3800d\") " Nov 25 11:14:14 crc kubenswrapper[4821]: I1125 11:14:14.740513 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/40682cac-f449-4465-9ad3-3bfa28b3800d-nova-extra-config-0\") pod \"40682cac-f449-4465-9ad3-3bfa28b3800d\" (UID: \"40682cac-f449-4465-9ad3-3bfa28b3800d\") " Nov 25 11:14:14 crc kubenswrapper[4821]: I1125 11:14:14.740579 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xrb6\" (UniqueName: \"kubernetes.io/projected/40682cac-f449-4465-9ad3-3bfa28b3800d-kube-api-access-9xrb6\") pod \"40682cac-f449-4465-9ad3-3bfa28b3800d\" (UID: \"40682cac-f449-4465-9ad3-3bfa28b3800d\") " Nov 25 11:14:14 crc kubenswrapper[4821]: I1125 11:14:14.740617 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/40682cac-f449-4465-9ad3-3bfa28b3800d-nova-migration-ssh-key-0\") pod \"40682cac-f449-4465-9ad3-3bfa28b3800d\" (UID: \"40682cac-f449-4465-9ad3-3bfa28b3800d\") " Nov 25 11:14:14 crc kubenswrapper[4821]: I1125 11:14:14.740643 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/40682cac-f449-4465-9ad3-3bfa28b3800d-nova-migration-ssh-key-1\") pod \"40682cac-f449-4465-9ad3-3bfa28b3800d\" (UID: \"40682cac-f449-4465-9ad3-3bfa28b3800d\") " Nov 25 11:14:14 crc kubenswrapper[4821]: I1125 11:14:14.745663 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40682cac-f449-4465-9ad3-3bfa28b3800d-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "40682cac-f449-4465-9ad3-3bfa28b3800d" (UID: "40682cac-f449-4465-9ad3-3bfa28b3800d"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:14:14 crc kubenswrapper[4821]: I1125 11:14:14.766726 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40682cac-f449-4465-9ad3-3bfa28b3800d-kube-api-access-9xrb6" (OuterVolumeSpecName: "kube-api-access-9xrb6") pod "40682cac-f449-4465-9ad3-3bfa28b3800d" (UID: "40682cac-f449-4465-9ad3-3bfa28b3800d"). InnerVolumeSpecName "kube-api-access-9xrb6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:14:14 crc kubenswrapper[4821]: I1125 11:14:14.770298 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40682cac-f449-4465-9ad3-3bfa28b3800d-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "40682cac-f449-4465-9ad3-3bfa28b3800d" (UID: "40682cac-f449-4465-9ad3-3bfa28b3800d"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:14:14 crc kubenswrapper[4821]: I1125 11:14:14.772735 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40682cac-f449-4465-9ad3-3bfa28b3800d-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "40682cac-f449-4465-9ad3-3bfa28b3800d" (UID: "40682cac-f449-4465-9ad3-3bfa28b3800d"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:14:14 crc kubenswrapper[4821]: I1125 11:14:14.772778 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40682cac-f449-4465-9ad3-3bfa28b3800d-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "40682cac-f449-4465-9ad3-3bfa28b3800d" (UID: "40682cac-f449-4465-9ad3-3bfa28b3800d"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:14:14 crc kubenswrapper[4821]: I1125 11:14:14.774773 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/40682cac-f449-4465-9ad3-3bfa28b3800d-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "40682cac-f449-4465-9ad3-3bfa28b3800d" (UID: "40682cac-f449-4465-9ad3-3bfa28b3800d"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:14:14 crc kubenswrapper[4821]: I1125 11:14:14.775158 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40682cac-f449-4465-9ad3-3bfa28b3800d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "40682cac-f449-4465-9ad3-3bfa28b3800d" (UID: "40682cac-f449-4465-9ad3-3bfa28b3800d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:14:14 crc kubenswrapper[4821]: I1125 11:14:14.789027 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40682cac-f449-4465-9ad3-3bfa28b3800d-inventory" (OuterVolumeSpecName: "inventory") pod "40682cac-f449-4465-9ad3-3bfa28b3800d" (UID: "40682cac-f449-4465-9ad3-3bfa28b3800d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:14:14 crc kubenswrapper[4821]: I1125 11:14:14.796962 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40682cac-f449-4465-9ad3-3bfa28b3800d-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "40682cac-f449-4465-9ad3-3bfa28b3800d" (UID: "40682cac-f449-4465-9ad3-3bfa28b3800d"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:14:14 crc kubenswrapper[4821]: I1125 11:14:14.843027 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xrb6\" (UniqueName: \"kubernetes.io/projected/40682cac-f449-4465-9ad3-3bfa28b3800d-kube-api-access-9xrb6\") on node \"crc\" DevicePath \"\"" Nov 25 11:14:14 crc kubenswrapper[4821]: I1125 11:14:14.843066 4821 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/40682cac-f449-4465-9ad3-3bfa28b3800d-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Nov 25 11:14:14 crc kubenswrapper[4821]: I1125 11:14:14.843082 4821 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/40682cac-f449-4465-9ad3-3bfa28b3800d-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Nov 25 11:14:14 crc kubenswrapper[4821]: I1125 11:14:14.843093 4821 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/40682cac-f449-4465-9ad3-3bfa28b3800d-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 11:14:14 crc kubenswrapper[4821]: I1125 11:14:14.843104 4821 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/40682cac-f449-4465-9ad3-3bfa28b3800d-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Nov 25 11:14:14 crc kubenswrapper[4821]: I1125 11:14:14.843116 4821 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/40682cac-f449-4465-9ad3-3bfa28b3800d-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 11:14:14 crc kubenswrapper[4821]: I1125 11:14:14.843126 4821 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40682cac-f449-4465-9ad3-3bfa28b3800d-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:14:14 crc kubenswrapper[4821]: I1125 11:14:14.843138 4821 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/40682cac-f449-4465-9ad3-3bfa28b3800d-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Nov 25 11:14:14 crc kubenswrapper[4821]: I1125 11:14:14.843149 4821 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/40682cac-f449-4465-9ad3-3bfa28b3800d-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Nov 25 11:14:15 crc kubenswrapper[4821]: I1125 11:14:15.103048 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mt9vk" event={"ID":"40682cac-f449-4465-9ad3-3bfa28b3800d","Type":"ContainerDied","Data":"6dbfe0d593c06cbf5fe45a675965cd46a5035d296e4f25bcb7654c5f7b3b1b31"} Nov 25 11:14:15 crc kubenswrapper[4821]: I1125 11:14:15.103089 4821 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6dbfe0d593c06cbf5fe45a675965cd46a5035d296e4f25bcb7654c5f7b3b1b31" Nov 25 11:14:15 crc kubenswrapper[4821]: I1125 11:14:15.103183 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-mt9vk" Nov 25 11:14:15 crc kubenswrapper[4821]: I1125 11:14:15.202547 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-w75cf"] Nov 25 11:14:15 crc kubenswrapper[4821]: E1125 11:14:15.202940 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40682cac-f449-4465-9ad3-3bfa28b3800d" containerName="nova-edpm-deployment-openstack-edpm-ipam" Nov 25 11:14:15 crc kubenswrapper[4821]: I1125 11:14:15.202956 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="40682cac-f449-4465-9ad3-3bfa28b3800d" containerName="nova-edpm-deployment-openstack-edpm-ipam" Nov 25 11:14:15 crc kubenswrapper[4821]: I1125 11:14:15.203173 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="40682cac-f449-4465-9ad3-3bfa28b3800d" containerName="nova-edpm-deployment-openstack-edpm-ipam" Nov 25 11:14:15 crc kubenswrapper[4821]: I1125 11:14:15.203773 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-w75cf" Nov 25 11:14:15 crc kubenswrapper[4821]: I1125 11:14:15.205671 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 11:14:15 crc kubenswrapper[4821]: I1125 11:14:15.208109 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 11:14:15 crc kubenswrapper[4821]: I1125 11:14:15.208226 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-z764x" Nov 25 11:14:15 crc kubenswrapper[4821]: I1125 11:14:15.208109 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Nov 25 11:14:15 crc kubenswrapper[4821]: I1125 11:14:15.214365 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 11:14:15 crc kubenswrapper[4821]: I1125 11:14:15.215582 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-w75cf"] Nov 25 11:14:15 crc kubenswrapper[4821]: I1125 11:14:15.352826 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8mtjj\" (UniqueName: \"kubernetes.io/projected/814ef062-0cc8-40bf-ac77-c87010466f7d-kube-api-access-8mtjj\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-w75cf\" (UID: \"814ef062-0cc8-40bf-ac77-c87010466f7d\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-w75cf" Nov 25 11:14:15 crc kubenswrapper[4821]: I1125 11:14:15.352873 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/814ef062-0cc8-40bf-ac77-c87010466f7d-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-w75cf\" (UID: \"814ef062-0cc8-40bf-ac77-c87010466f7d\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-w75cf" Nov 25 11:14:15 crc kubenswrapper[4821]: I1125 11:14:15.352907 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/814ef062-0cc8-40bf-ac77-c87010466f7d-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-w75cf\" (UID: \"814ef062-0cc8-40bf-ac77-c87010466f7d\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-w75cf" Nov 25 11:14:15 crc kubenswrapper[4821]: I1125 11:14:15.352940 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/814ef062-0cc8-40bf-ac77-c87010466f7d-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-w75cf\" (UID: \"814ef062-0cc8-40bf-ac77-c87010466f7d\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-w75cf" Nov 25 11:14:15 crc kubenswrapper[4821]: I1125 11:14:15.353090 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/814ef062-0cc8-40bf-ac77-c87010466f7d-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-w75cf\" (UID: \"814ef062-0cc8-40bf-ac77-c87010466f7d\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-w75cf" Nov 25 11:14:15 crc kubenswrapper[4821]: I1125 11:14:15.353126 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/814ef062-0cc8-40bf-ac77-c87010466f7d-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-w75cf\" (UID: \"814ef062-0cc8-40bf-ac77-c87010466f7d\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-w75cf" Nov 25 11:14:15 crc kubenswrapper[4821]: I1125 11:14:15.353239 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/814ef062-0cc8-40bf-ac77-c87010466f7d-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-w75cf\" (UID: \"814ef062-0cc8-40bf-ac77-c87010466f7d\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-w75cf" Nov 25 11:14:15 crc kubenswrapper[4821]: I1125 11:14:15.455207 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/814ef062-0cc8-40bf-ac77-c87010466f7d-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-w75cf\" (UID: \"814ef062-0cc8-40bf-ac77-c87010466f7d\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-w75cf" Nov 25 11:14:15 crc kubenswrapper[4821]: I1125 11:14:15.455305 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/814ef062-0cc8-40bf-ac77-c87010466f7d-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-w75cf\" (UID: \"814ef062-0cc8-40bf-ac77-c87010466f7d\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-w75cf" Nov 25 11:14:15 crc kubenswrapper[4821]: I1125 11:14:15.455368 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8mtjj\" (UniqueName: \"kubernetes.io/projected/814ef062-0cc8-40bf-ac77-c87010466f7d-kube-api-access-8mtjj\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-w75cf\" (UID: \"814ef062-0cc8-40bf-ac77-c87010466f7d\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-w75cf" Nov 25 11:14:15 crc kubenswrapper[4821]: I1125 11:14:15.455410 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/814ef062-0cc8-40bf-ac77-c87010466f7d-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-w75cf\" (UID: \"814ef062-0cc8-40bf-ac77-c87010466f7d\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-w75cf" Nov 25 11:14:15 crc kubenswrapper[4821]: I1125 11:14:15.455440 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/814ef062-0cc8-40bf-ac77-c87010466f7d-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-w75cf\" (UID: \"814ef062-0cc8-40bf-ac77-c87010466f7d\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-w75cf" Nov 25 11:14:15 crc kubenswrapper[4821]: I1125 11:14:15.455467 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/814ef062-0cc8-40bf-ac77-c87010466f7d-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-w75cf\" (UID: \"814ef062-0cc8-40bf-ac77-c87010466f7d\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-w75cf" Nov 25 11:14:15 crc kubenswrapper[4821]: I1125 11:14:15.455527 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/814ef062-0cc8-40bf-ac77-c87010466f7d-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-w75cf\" (UID: \"814ef062-0cc8-40bf-ac77-c87010466f7d\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-w75cf" Nov 25 11:14:15 crc kubenswrapper[4821]: I1125 11:14:15.460576 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/814ef062-0cc8-40bf-ac77-c87010466f7d-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-w75cf\" (UID: \"814ef062-0cc8-40bf-ac77-c87010466f7d\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-w75cf" Nov 25 11:14:15 crc kubenswrapper[4821]: I1125 11:14:15.460681 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/814ef062-0cc8-40bf-ac77-c87010466f7d-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-w75cf\" (UID: \"814ef062-0cc8-40bf-ac77-c87010466f7d\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-w75cf" Nov 25 11:14:15 crc kubenswrapper[4821]: I1125 11:14:15.460776 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/814ef062-0cc8-40bf-ac77-c87010466f7d-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-w75cf\" (UID: \"814ef062-0cc8-40bf-ac77-c87010466f7d\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-w75cf" Nov 25 11:14:15 crc kubenswrapper[4821]: I1125 11:14:15.461891 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/814ef062-0cc8-40bf-ac77-c87010466f7d-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-w75cf\" (UID: \"814ef062-0cc8-40bf-ac77-c87010466f7d\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-w75cf" Nov 25 11:14:15 crc kubenswrapper[4821]: I1125 11:14:15.463266 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/814ef062-0cc8-40bf-ac77-c87010466f7d-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-w75cf\" (UID: \"814ef062-0cc8-40bf-ac77-c87010466f7d\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-w75cf" Nov 25 11:14:15 crc kubenswrapper[4821]: I1125 11:14:15.468596 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/814ef062-0cc8-40bf-ac77-c87010466f7d-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-w75cf\" (UID: \"814ef062-0cc8-40bf-ac77-c87010466f7d\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-w75cf" Nov 25 11:14:15 crc kubenswrapper[4821]: I1125 11:14:15.479305 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8mtjj\" (UniqueName: \"kubernetes.io/projected/814ef062-0cc8-40bf-ac77-c87010466f7d-kube-api-access-8mtjj\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-w75cf\" (UID: \"814ef062-0cc8-40bf-ac77-c87010466f7d\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-w75cf" Nov 25 11:14:15 crc kubenswrapper[4821]: I1125 11:14:15.522725 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-w75cf" Nov 25 11:14:16 crc kubenswrapper[4821]: I1125 11:14:16.049174 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-w75cf"] Nov 25 11:14:16 crc kubenswrapper[4821]: I1125 11:14:16.064200 4821 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 11:14:16 crc kubenswrapper[4821]: I1125 11:14:16.127374 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-w75cf" event={"ID":"814ef062-0cc8-40bf-ac77-c87010466f7d","Type":"ContainerStarted","Data":"d935cc1cc0651a73a0e6410818217c84442d66d01211c09b8c1933733f2ef551"} Nov 25 11:14:17 crc kubenswrapper[4821]: I1125 11:14:17.127483 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-w75cf" event={"ID":"814ef062-0cc8-40bf-ac77-c87010466f7d","Type":"ContainerStarted","Data":"aa3aa650ec47be2eb1cbf081799d97db55d73c28c1b732305c2c41b87ccb2ea0"} Nov 25 11:14:17 crc kubenswrapper[4821]: I1125 11:14:17.149104 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-w75cf" podStartSLOduration=1.736321161 podStartE2EDuration="2.149083696s" podCreationTimestamp="2025-11-25 11:14:15 +0000 UTC" firstStartedPulling="2025-11-25 11:14:16.06387142 +0000 UTC m=+2526.600191277" lastFinishedPulling="2025-11-25 11:14:16.476633925 +0000 UTC m=+2527.012953812" observedRunningTime="2025-11-25 11:14:17.140968993 +0000 UTC m=+2527.677288850" watchObservedRunningTime="2025-11-25 11:14:17.149083696 +0000 UTC m=+2527.685403543" Nov 25 11:14:18 crc kubenswrapper[4821]: I1125 11:14:18.115214 4821 scope.go:117] "RemoveContainer" containerID="61010995905bb7c40fff6417cf7e0b11d584b8a9a68f6fccf71aefe231f45291" Nov 25 11:14:18 crc kubenswrapper[4821]: E1125 11:14:18.115585 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:14:32 crc kubenswrapper[4821]: I1125 11:14:32.114506 4821 scope.go:117] "RemoveContainer" containerID="61010995905bb7c40fff6417cf7e0b11d584b8a9a68f6fccf71aefe231f45291" Nov 25 11:14:32 crc kubenswrapper[4821]: E1125 11:14:32.115551 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:14:47 crc kubenswrapper[4821]: I1125 11:14:47.115136 4821 scope.go:117] "RemoveContainer" containerID="61010995905bb7c40fff6417cf7e0b11d584b8a9a68f6fccf71aefe231f45291" Nov 25 11:14:47 crc kubenswrapper[4821]: E1125 11:14:47.116555 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:15:00 crc kubenswrapper[4821]: I1125 11:15:00.155858 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401155-wzwhl"] Nov 25 11:15:00 crc kubenswrapper[4821]: I1125 11:15:00.158192 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401155-wzwhl" Nov 25 11:15:00 crc kubenswrapper[4821]: I1125 11:15:00.159875 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 25 11:15:00 crc kubenswrapper[4821]: I1125 11:15:00.160095 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 25 11:15:00 crc kubenswrapper[4821]: I1125 11:15:00.166416 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401155-wzwhl"] Nov 25 11:15:00 crc kubenswrapper[4821]: I1125 11:15:00.216373 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cfczc\" (UniqueName: \"kubernetes.io/projected/c4375c27-1d24-436d-8d20-35e8b70727d3-kube-api-access-cfczc\") pod \"collect-profiles-29401155-wzwhl\" (UID: \"c4375c27-1d24-436d-8d20-35e8b70727d3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401155-wzwhl" Nov 25 11:15:00 crc kubenswrapper[4821]: I1125 11:15:00.216729 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c4375c27-1d24-436d-8d20-35e8b70727d3-secret-volume\") pod \"collect-profiles-29401155-wzwhl\" (UID: \"c4375c27-1d24-436d-8d20-35e8b70727d3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401155-wzwhl" Nov 25 11:15:00 crc kubenswrapper[4821]: I1125 11:15:00.217181 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c4375c27-1d24-436d-8d20-35e8b70727d3-config-volume\") pod \"collect-profiles-29401155-wzwhl\" (UID: \"c4375c27-1d24-436d-8d20-35e8b70727d3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401155-wzwhl" Nov 25 11:15:00 crc kubenswrapper[4821]: I1125 11:15:00.318692 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c4375c27-1d24-436d-8d20-35e8b70727d3-secret-volume\") pod \"collect-profiles-29401155-wzwhl\" (UID: \"c4375c27-1d24-436d-8d20-35e8b70727d3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401155-wzwhl" Nov 25 11:15:00 crc kubenswrapper[4821]: I1125 11:15:00.318805 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c4375c27-1d24-436d-8d20-35e8b70727d3-config-volume\") pod \"collect-profiles-29401155-wzwhl\" (UID: \"c4375c27-1d24-436d-8d20-35e8b70727d3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401155-wzwhl" Nov 25 11:15:00 crc kubenswrapper[4821]: I1125 11:15:00.318855 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cfczc\" (UniqueName: \"kubernetes.io/projected/c4375c27-1d24-436d-8d20-35e8b70727d3-kube-api-access-cfczc\") pod \"collect-profiles-29401155-wzwhl\" (UID: \"c4375c27-1d24-436d-8d20-35e8b70727d3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401155-wzwhl" Nov 25 11:15:00 crc kubenswrapper[4821]: I1125 11:15:00.319954 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c4375c27-1d24-436d-8d20-35e8b70727d3-config-volume\") pod \"collect-profiles-29401155-wzwhl\" (UID: \"c4375c27-1d24-436d-8d20-35e8b70727d3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401155-wzwhl" Nov 25 11:15:00 crc kubenswrapper[4821]: I1125 11:15:00.324822 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c4375c27-1d24-436d-8d20-35e8b70727d3-secret-volume\") pod \"collect-profiles-29401155-wzwhl\" (UID: \"c4375c27-1d24-436d-8d20-35e8b70727d3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401155-wzwhl" Nov 25 11:15:00 crc kubenswrapper[4821]: I1125 11:15:00.336607 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cfczc\" (UniqueName: \"kubernetes.io/projected/c4375c27-1d24-436d-8d20-35e8b70727d3-kube-api-access-cfczc\") pod \"collect-profiles-29401155-wzwhl\" (UID: \"c4375c27-1d24-436d-8d20-35e8b70727d3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401155-wzwhl" Nov 25 11:15:00 crc kubenswrapper[4821]: I1125 11:15:00.492716 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401155-wzwhl" Nov 25 11:15:00 crc kubenswrapper[4821]: I1125 11:15:00.908528 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401155-wzwhl"] Nov 25 11:15:01 crc kubenswrapper[4821]: I1125 11:15:01.563364 4821 generic.go:334] "Generic (PLEG): container finished" podID="c4375c27-1d24-436d-8d20-35e8b70727d3" containerID="d8ff54ed641dc706d4c8fb469bfec24910451714c9a8d28075d6129905e755a0" exitCode=0 Nov 25 11:15:01 crc kubenswrapper[4821]: I1125 11:15:01.563428 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401155-wzwhl" event={"ID":"c4375c27-1d24-436d-8d20-35e8b70727d3","Type":"ContainerDied","Data":"d8ff54ed641dc706d4c8fb469bfec24910451714c9a8d28075d6129905e755a0"} Nov 25 11:15:01 crc kubenswrapper[4821]: I1125 11:15:01.564498 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401155-wzwhl" event={"ID":"c4375c27-1d24-436d-8d20-35e8b70727d3","Type":"ContainerStarted","Data":"c56d9f1b4d49076255d9f2ea7c15ad3f8b9ea33308a8824a4b37a32678614eb7"} Nov 25 11:15:02 crc kubenswrapper[4821]: I1125 11:15:02.115086 4821 scope.go:117] "RemoveContainer" containerID="61010995905bb7c40fff6417cf7e0b11d584b8a9a68f6fccf71aefe231f45291" Nov 25 11:15:02 crc kubenswrapper[4821]: E1125 11:15:02.115456 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:15:02 crc kubenswrapper[4821]: I1125 11:15:02.933381 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401155-wzwhl" Nov 25 11:15:03 crc kubenswrapper[4821]: I1125 11:15:03.069128 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c4375c27-1d24-436d-8d20-35e8b70727d3-secret-volume\") pod \"c4375c27-1d24-436d-8d20-35e8b70727d3\" (UID: \"c4375c27-1d24-436d-8d20-35e8b70727d3\") " Nov 25 11:15:03 crc kubenswrapper[4821]: I1125 11:15:03.069231 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfczc\" (UniqueName: \"kubernetes.io/projected/c4375c27-1d24-436d-8d20-35e8b70727d3-kube-api-access-cfczc\") pod \"c4375c27-1d24-436d-8d20-35e8b70727d3\" (UID: \"c4375c27-1d24-436d-8d20-35e8b70727d3\") " Nov 25 11:15:03 crc kubenswrapper[4821]: I1125 11:15:03.069307 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c4375c27-1d24-436d-8d20-35e8b70727d3-config-volume\") pod \"c4375c27-1d24-436d-8d20-35e8b70727d3\" (UID: \"c4375c27-1d24-436d-8d20-35e8b70727d3\") " Nov 25 11:15:03 crc kubenswrapper[4821]: I1125 11:15:03.069983 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4375c27-1d24-436d-8d20-35e8b70727d3-config-volume" (OuterVolumeSpecName: "config-volume") pod "c4375c27-1d24-436d-8d20-35e8b70727d3" (UID: "c4375c27-1d24-436d-8d20-35e8b70727d3"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:15:03 crc kubenswrapper[4821]: I1125 11:15:03.075147 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4375c27-1d24-436d-8d20-35e8b70727d3-kube-api-access-cfczc" (OuterVolumeSpecName: "kube-api-access-cfczc") pod "c4375c27-1d24-436d-8d20-35e8b70727d3" (UID: "c4375c27-1d24-436d-8d20-35e8b70727d3"). InnerVolumeSpecName "kube-api-access-cfczc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:15:03 crc kubenswrapper[4821]: I1125 11:15:03.075299 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4375c27-1d24-436d-8d20-35e8b70727d3-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c4375c27-1d24-436d-8d20-35e8b70727d3" (UID: "c4375c27-1d24-436d-8d20-35e8b70727d3"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:15:03 crc kubenswrapper[4821]: I1125 11:15:03.171322 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfczc\" (UniqueName: \"kubernetes.io/projected/c4375c27-1d24-436d-8d20-35e8b70727d3-kube-api-access-cfczc\") on node \"crc\" DevicePath \"\"" Nov 25 11:15:03 crc kubenswrapper[4821]: I1125 11:15:03.171634 4821 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c4375c27-1d24-436d-8d20-35e8b70727d3-config-volume\") on node \"crc\" DevicePath \"\"" Nov 25 11:15:03 crc kubenswrapper[4821]: I1125 11:15:03.171643 4821 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c4375c27-1d24-436d-8d20-35e8b70727d3-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 25 11:15:03 crc kubenswrapper[4821]: I1125 11:15:03.596924 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401155-wzwhl" event={"ID":"c4375c27-1d24-436d-8d20-35e8b70727d3","Type":"ContainerDied","Data":"c56d9f1b4d49076255d9f2ea7c15ad3f8b9ea33308a8824a4b37a32678614eb7"} Nov 25 11:15:03 crc kubenswrapper[4821]: I1125 11:15:03.596973 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401155-wzwhl" Nov 25 11:15:03 crc kubenswrapper[4821]: I1125 11:15:03.596998 4821 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c56d9f1b4d49076255d9f2ea7c15ad3f8b9ea33308a8824a4b37a32678614eb7" Nov 25 11:15:04 crc kubenswrapper[4821]: I1125 11:15:04.005988 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401110-dm86l"] Nov 25 11:15:04 crc kubenswrapper[4821]: I1125 11:15:04.013513 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401110-dm86l"] Nov 25 11:15:04 crc kubenswrapper[4821]: I1125 11:15:04.124273 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5033129-ba8c-488f-982e-78e12b2f1664" path="/var/lib/kubelet/pods/a5033129-ba8c-488f-982e-78e12b2f1664/volumes" Nov 25 11:15:16 crc kubenswrapper[4821]: I1125 11:15:16.114650 4821 scope.go:117] "RemoveContainer" containerID="61010995905bb7c40fff6417cf7e0b11d584b8a9a68f6fccf71aefe231f45291" Nov 25 11:15:16 crc kubenswrapper[4821]: E1125 11:15:16.115432 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:15:25 crc kubenswrapper[4821]: I1125 11:15:25.318914 4821 scope.go:117] "RemoveContainer" containerID="21e6691e92fd38f9fb69206ae744da84aa6fde06b42b6560d4a335470deec9e3" Nov 25 11:15:28 crc kubenswrapper[4821]: I1125 11:15:28.117238 4821 scope.go:117] "RemoveContainer" containerID="61010995905bb7c40fff6417cf7e0b11d584b8a9a68f6fccf71aefe231f45291" Nov 25 11:15:28 crc kubenswrapper[4821]: E1125 11:15:28.118060 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:15:39 crc kubenswrapper[4821]: I1125 11:15:39.114731 4821 scope.go:117] "RemoveContainer" containerID="61010995905bb7c40fff6417cf7e0b11d584b8a9a68f6fccf71aefe231f45291" Nov 25 11:15:39 crc kubenswrapper[4821]: E1125 11:15:39.115905 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:15:52 crc kubenswrapper[4821]: I1125 11:15:52.115226 4821 scope.go:117] "RemoveContainer" containerID="61010995905bb7c40fff6417cf7e0b11d584b8a9a68f6fccf71aefe231f45291" Nov 25 11:15:52 crc kubenswrapper[4821]: E1125 11:15:52.116954 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:16:07 crc kubenswrapper[4821]: I1125 11:16:07.115117 4821 scope.go:117] "RemoveContainer" containerID="61010995905bb7c40fff6417cf7e0b11d584b8a9a68f6fccf71aefe231f45291" Nov 25 11:16:08 crc kubenswrapper[4821]: I1125 11:16:08.165113 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" event={"ID":"5f948b87-ac86-4de6-ad64-c2ef947f84d4","Type":"ContainerStarted","Data":"d5e77cf7ba2ff0993f9dc84aaf718cec52c804bab193ec24676471e97efd1f7b"} Nov 25 11:16:35 crc kubenswrapper[4821]: I1125 11:16:35.389835 4821 generic.go:334] "Generic (PLEG): container finished" podID="814ef062-0cc8-40bf-ac77-c87010466f7d" containerID="aa3aa650ec47be2eb1cbf081799d97db55d73c28c1b732305c2c41b87ccb2ea0" exitCode=0 Nov 25 11:16:35 crc kubenswrapper[4821]: I1125 11:16:35.389989 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-w75cf" event={"ID":"814ef062-0cc8-40bf-ac77-c87010466f7d","Type":"ContainerDied","Data":"aa3aa650ec47be2eb1cbf081799d97db55d73c28c1b732305c2c41b87ccb2ea0"} Nov 25 11:16:36 crc kubenswrapper[4821]: I1125 11:16:36.807414 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-w75cf" Nov 25 11:16:36 crc kubenswrapper[4821]: I1125 11:16:36.877155 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8mtjj\" (UniqueName: \"kubernetes.io/projected/814ef062-0cc8-40bf-ac77-c87010466f7d-kube-api-access-8mtjj\") pod \"814ef062-0cc8-40bf-ac77-c87010466f7d\" (UID: \"814ef062-0cc8-40bf-ac77-c87010466f7d\") " Nov 25 11:16:36 crc kubenswrapper[4821]: I1125 11:16:36.877502 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/814ef062-0cc8-40bf-ac77-c87010466f7d-ceilometer-compute-config-data-1\") pod \"814ef062-0cc8-40bf-ac77-c87010466f7d\" (UID: \"814ef062-0cc8-40bf-ac77-c87010466f7d\") " Nov 25 11:16:36 crc kubenswrapper[4821]: I1125 11:16:36.877594 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/814ef062-0cc8-40bf-ac77-c87010466f7d-ceilometer-compute-config-data-0\") pod \"814ef062-0cc8-40bf-ac77-c87010466f7d\" (UID: \"814ef062-0cc8-40bf-ac77-c87010466f7d\") " Nov 25 11:16:36 crc kubenswrapper[4821]: I1125 11:16:36.877708 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/814ef062-0cc8-40bf-ac77-c87010466f7d-ssh-key\") pod \"814ef062-0cc8-40bf-ac77-c87010466f7d\" (UID: \"814ef062-0cc8-40bf-ac77-c87010466f7d\") " Nov 25 11:16:36 crc kubenswrapper[4821]: I1125 11:16:36.877834 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/814ef062-0cc8-40bf-ac77-c87010466f7d-telemetry-combined-ca-bundle\") pod \"814ef062-0cc8-40bf-ac77-c87010466f7d\" (UID: \"814ef062-0cc8-40bf-ac77-c87010466f7d\") " Nov 25 11:16:36 crc kubenswrapper[4821]: I1125 11:16:36.877920 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/814ef062-0cc8-40bf-ac77-c87010466f7d-inventory\") pod \"814ef062-0cc8-40bf-ac77-c87010466f7d\" (UID: \"814ef062-0cc8-40bf-ac77-c87010466f7d\") " Nov 25 11:16:36 crc kubenswrapper[4821]: I1125 11:16:36.877989 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/814ef062-0cc8-40bf-ac77-c87010466f7d-ceilometer-compute-config-data-2\") pod \"814ef062-0cc8-40bf-ac77-c87010466f7d\" (UID: \"814ef062-0cc8-40bf-ac77-c87010466f7d\") " Nov 25 11:16:36 crc kubenswrapper[4821]: I1125 11:16:36.882994 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/814ef062-0cc8-40bf-ac77-c87010466f7d-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "814ef062-0cc8-40bf-ac77-c87010466f7d" (UID: "814ef062-0cc8-40bf-ac77-c87010466f7d"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:16:36 crc kubenswrapper[4821]: I1125 11:16:36.884273 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/814ef062-0cc8-40bf-ac77-c87010466f7d-kube-api-access-8mtjj" (OuterVolumeSpecName: "kube-api-access-8mtjj") pod "814ef062-0cc8-40bf-ac77-c87010466f7d" (UID: "814ef062-0cc8-40bf-ac77-c87010466f7d"). InnerVolumeSpecName "kube-api-access-8mtjj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:16:36 crc kubenswrapper[4821]: I1125 11:16:36.908702 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/814ef062-0cc8-40bf-ac77-c87010466f7d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "814ef062-0cc8-40bf-ac77-c87010466f7d" (UID: "814ef062-0cc8-40bf-ac77-c87010466f7d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:16:36 crc kubenswrapper[4821]: I1125 11:16:36.908761 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/814ef062-0cc8-40bf-ac77-c87010466f7d-inventory" (OuterVolumeSpecName: "inventory") pod "814ef062-0cc8-40bf-ac77-c87010466f7d" (UID: "814ef062-0cc8-40bf-ac77-c87010466f7d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:16:36 crc kubenswrapper[4821]: I1125 11:16:36.909288 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/814ef062-0cc8-40bf-ac77-c87010466f7d-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "814ef062-0cc8-40bf-ac77-c87010466f7d" (UID: "814ef062-0cc8-40bf-ac77-c87010466f7d"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:16:36 crc kubenswrapper[4821]: I1125 11:16:36.913658 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/814ef062-0cc8-40bf-ac77-c87010466f7d-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "814ef062-0cc8-40bf-ac77-c87010466f7d" (UID: "814ef062-0cc8-40bf-ac77-c87010466f7d"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:16:36 crc kubenswrapper[4821]: I1125 11:16:36.915350 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/814ef062-0cc8-40bf-ac77-c87010466f7d-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "814ef062-0cc8-40bf-ac77-c87010466f7d" (UID: "814ef062-0cc8-40bf-ac77-c87010466f7d"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:16:36 crc kubenswrapper[4821]: I1125 11:16:36.980763 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8mtjj\" (UniqueName: \"kubernetes.io/projected/814ef062-0cc8-40bf-ac77-c87010466f7d-kube-api-access-8mtjj\") on node \"crc\" DevicePath \"\"" Nov 25 11:16:36 crc kubenswrapper[4821]: I1125 11:16:36.980795 4821 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/814ef062-0cc8-40bf-ac77-c87010466f7d-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Nov 25 11:16:36 crc kubenswrapper[4821]: I1125 11:16:36.980807 4821 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/814ef062-0cc8-40bf-ac77-c87010466f7d-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Nov 25 11:16:36 crc kubenswrapper[4821]: I1125 11:16:36.980818 4821 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/814ef062-0cc8-40bf-ac77-c87010466f7d-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 11:16:36 crc kubenswrapper[4821]: I1125 11:16:36.980828 4821 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/814ef062-0cc8-40bf-ac77-c87010466f7d-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:16:36 crc kubenswrapper[4821]: I1125 11:16:36.980836 4821 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/814ef062-0cc8-40bf-ac77-c87010466f7d-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 11:16:36 crc kubenswrapper[4821]: I1125 11:16:36.980845 4821 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/814ef062-0cc8-40bf-ac77-c87010466f7d-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Nov 25 11:16:37 crc kubenswrapper[4821]: I1125 11:16:37.409691 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-w75cf" event={"ID":"814ef062-0cc8-40bf-ac77-c87010466f7d","Type":"ContainerDied","Data":"d935cc1cc0651a73a0e6410818217c84442d66d01211c09b8c1933733f2ef551"} Nov 25 11:16:37 crc kubenswrapper[4821]: I1125 11:16:37.410038 4821 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d935cc1cc0651a73a0e6410818217c84442d66d01211c09b8c1933733f2ef551" Nov 25 11:16:37 crc kubenswrapper[4821]: I1125 11:16:37.409746 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-w75cf" Nov 25 11:16:49 crc kubenswrapper[4821]: I1125 11:16:49.036102 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-2q5h7"] Nov 25 11:16:49 crc kubenswrapper[4821]: E1125 11:16:49.036983 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4375c27-1d24-436d-8d20-35e8b70727d3" containerName="collect-profiles" Nov 25 11:16:49 crc kubenswrapper[4821]: I1125 11:16:49.036995 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4375c27-1d24-436d-8d20-35e8b70727d3" containerName="collect-profiles" Nov 25 11:16:49 crc kubenswrapper[4821]: E1125 11:16:49.037009 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="814ef062-0cc8-40bf-ac77-c87010466f7d" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Nov 25 11:16:49 crc kubenswrapper[4821]: I1125 11:16:49.037017 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="814ef062-0cc8-40bf-ac77-c87010466f7d" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Nov 25 11:16:49 crc kubenswrapper[4821]: I1125 11:16:49.037249 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="814ef062-0cc8-40bf-ac77-c87010466f7d" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Nov 25 11:16:49 crc kubenswrapper[4821]: I1125 11:16:49.037265 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4375c27-1d24-436d-8d20-35e8b70727d3" containerName="collect-profiles" Nov 25 11:16:49 crc kubenswrapper[4821]: I1125 11:16:49.038543 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2q5h7" Nov 25 11:16:49 crc kubenswrapper[4821]: I1125 11:16:49.054467 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2q5h7"] Nov 25 11:16:49 crc kubenswrapper[4821]: I1125 11:16:49.200677 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/583b83b9-f7b0-451d-96ce-81b4f7a2041a-utilities\") pod \"redhat-marketplace-2q5h7\" (UID: \"583b83b9-f7b0-451d-96ce-81b4f7a2041a\") " pod="openshift-marketplace/redhat-marketplace-2q5h7" Nov 25 11:16:49 crc kubenswrapper[4821]: I1125 11:16:49.200744 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nks8h\" (UniqueName: \"kubernetes.io/projected/583b83b9-f7b0-451d-96ce-81b4f7a2041a-kube-api-access-nks8h\") pod \"redhat-marketplace-2q5h7\" (UID: \"583b83b9-f7b0-451d-96ce-81b4f7a2041a\") " pod="openshift-marketplace/redhat-marketplace-2q5h7" Nov 25 11:16:49 crc kubenswrapper[4821]: I1125 11:16:49.201481 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/583b83b9-f7b0-451d-96ce-81b4f7a2041a-catalog-content\") pod \"redhat-marketplace-2q5h7\" (UID: \"583b83b9-f7b0-451d-96ce-81b4f7a2041a\") " pod="openshift-marketplace/redhat-marketplace-2q5h7" Nov 25 11:16:49 crc kubenswrapper[4821]: I1125 11:16:49.302738 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/583b83b9-f7b0-451d-96ce-81b4f7a2041a-catalog-content\") pod \"redhat-marketplace-2q5h7\" (UID: \"583b83b9-f7b0-451d-96ce-81b4f7a2041a\") " pod="openshift-marketplace/redhat-marketplace-2q5h7" Nov 25 11:16:49 crc kubenswrapper[4821]: I1125 11:16:49.302783 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/583b83b9-f7b0-451d-96ce-81b4f7a2041a-utilities\") pod \"redhat-marketplace-2q5h7\" (UID: \"583b83b9-f7b0-451d-96ce-81b4f7a2041a\") " pod="openshift-marketplace/redhat-marketplace-2q5h7" Nov 25 11:16:49 crc kubenswrapper[4821]: I1125 11:16:49.302825 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nks8h\" (UniqueName: \"kubernetes.io/projected/583b83b9-f7b0-451d-96ce-81b4f7a2041a-kube-api-access-nks8h\") pod \"redhat-marketplace-2q5h7\" (UID: \"583b83b9-f7b0-451d-96ce-81b4f7a2041a\") " pod="openshift-marketplace/redhat-marketplace-2q5h7" Nov 25 11:16:49 crc kubenswrapper[4821]: I1125 11:16:49.303428 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/583b83b9-f7b0-451d-96ce-81b4f7a2041a-utilities\") pod \"redhat-marketplace-2q5h7\" (UID: \"583b83b9-f7b0-451d-96ce-81b4f7a2041a\") " pod="openshift-marketplace/redhat-marketplace-2q5h7" Nov 25 11:16:49 crc kubenswrapper[4821]: I1125 11:16:49.303984 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/583b83b9-f7b0-451d-96ce-81b4f7a2041a-catalog-content\") pod \"redhat-marketplace-2q5h7\" (UID: \"583b83b9-f7b0-451d-96ce-81b4f7a2041a\") " pod="openshift-marketplace/redhat-marketplace-2q5h7" Nov 25 11:16:49 crc kubenswrapper[4821]: I1125 11:16:49.321131 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nks8h\" (UniqueName: \"kubernetes.io/projected/583b83b9-f7b0-451d-96ce-81b4f7a2041a-kube-api-access-nks8h\") pod \"redhat-marketplace-2q5h7\" (UID: \"583b83b9-f7b0-451d-96ce-81b4f7a2041a\") " pod="openshift-marketplace/redhat-marketplace-2q5h7" Nov 25 11:16:49 crc kubenswrapper[4821]: I1125 11:16:49.355816 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2q5h7" Nov 25 11:16:49 crc kubenswrapper[4821]: I1125 11:16:49.800651 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2q5h7"] Nov 25 11:16:50 crc kubenswrapper[4821]: I1125 11:16:50.516613 4821 generic.go:334] "Generic (PLEG): container finished" podID="583b83b9-f7b0-451d-96ce-81b4f7a2041a" containerID="4a567484677ce930428412f7ba8d1ea0aa750275f76259d03b6d3447b0c80cef" exitCode=0 Nov 25 11:16:50 crc kubenswrapper[4821]: I1125 11:16:50.516670 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2q5h7" event={"ID":"583b83b9-f7b0-451d-96ce-81b4f7a2041a","Type":"ContainerDied","Data":"4a567484677ce930428412f7ba8d1ea0aa750275f76259d03b6d3447b0c80cef"} Nov 25 11:16:50 crc kubenswrapper[4821]: I1125 11:16:50.516966 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2q5h7" event={"ID":"583b83b9-f7b0-451d-96ce-81b4f7a2041a","Type":"ContainerStarted","Data":"bc4dc420782b550b45f8b7a93844824877a0bb67d709fd15599ac4eb7fafe3f0"} Nov 25 11:16:52 crc kubenswrapper[4821]: I1125 11:16:52.540823 4821 generic.go:334] "Generic (PLEG): container finished" podID="583b83b9-f7b0-451d-96ce-81b4f7a2041a" containerID="b8fc1ada99d9bd0a58b30efcca01f466e10a46647178117c6d16b3bbd865f207" exitCode=0 Nov 25 11:16:52 crc kubenswrapper[4821]: I1125 11:16:52.540911 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2q5h7" event={"ID":"583b83b9-f7b0-451d-96ce-81b4f7a2041a","Type":"ContainerDied","Data":"b8fc1ada99d9bd0a58b30efcca01f466e10a46647178117c6d16b3bbd865f207"} Nov 25 11:16:53 crc kubenswrapper[4821]: I1125 11:16:53.554436 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2q5h7" event={"ID":"583b83b9-f7b0-451d-96ce-81b4f7a2041a","Type":"ContainerStarted","Data":"2335ea31c7ae0e4881551d6d3259d1d2e29f13cb5df67217dca7c688b4a3d485"} Nov 25 11:16:53 crc kubenswrapper[4821]: I1125 11:16:53.580660 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-2q5h7" podStartSLOduration=2.1556376999999998 podStartE2EDuration="4.580641878s" podCreationTimestamp="2025-11-25 11:16:49 +0000 UTC" firstStartedPulling="2025-11-25 11:16:50.518073156 +0000 UTC m=+2681.054393003" lastFinishedPulling="2025-11-25 11:16:52.943077334 +0000 UTC m=+2683.479397181" observedRunningTime="2025-11-25 11:16:53.57170509 +0000 UTC m=+2684.108024937" watchObservedRunningTime="2025-11-25 11:16:53.580641878 +0000 UTC m=+2684.116961725" Nov 25 11:16:59 crc kubenswrapper[4821]: I1125 11:16:59.356659 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-2q5h7" Nov 25 11:16:59 crc kubenswrapper[4821]: I1125 11:16:59.357361 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-2q5h7" Nov 25 11:16:59 crc kubenswrapper[4821]: I1125 11:16:59.401025 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-2q5h7" Nov 25 11:16:59 crc kubenswrapper[4821]: I1125 11:16:59.655984 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-2q5h7" Nov 25 11:16:59 crc kubenswrapper[4821]: I1125 11:16:59.704045 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2q5h7"] Nov 25 11:17:01 crc kubenswrapper[4821]: I1125 11:17:01.623693 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-2q5h7" podUID="583b83b9-f7b0-451d-96ce-81b4f7a2041a" containerName="registry-server" containerID="cri-o://2335ea31c7ae0e4881551d6d3259d1d2e29f13cb5df67217dca7c688b4a3d485" gracePeriod=2 Nov 25 11:17:02 crc kubenswrapper[4821]: I1125 11:17:02.172283 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2q5h7" Nov 25 11:17:02 crc kubenswrapper[4821]: I1125 11:17:02.372280 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nks8h\" (UniqueName: \"kubernetes.io/projected/583b83b9-f7b0-451d-96ce-81b4f7a2041a-kube-api-access-nks8h\") pod \"583b83b9-f7b0-451d-96ce-81b4f7a2041a\" (UID: \"583b83b9-f7b0-451d-96ce-81b4f7a2041a\") " Nov 25 11:17:02 crc kubenswrapper[4821]: I1125 11:17:02.372345 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/583b83b9-f7b0-451d-96ce-81b4f7a2041a-utilities\") pod \"583b83b9-f7b0-451d-96ce-81b4f7a2041a\" (UID: \"583b83b9-f7b0-451d-96ce-81b4f7a2041a\") " Nov 25 11:17:02 crc kubenswrapper[4821]: I1125 11:17:02.372389 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/583b83b9-f7b0-451d-96ce-81b4f7a2041a-catalog-content\") pod \"583b83b9-f7b0-451d-96ce-81b4f7a2041a\" (UID: \"583b83b9-f7b0-451d-96ce-81b4f7a2041a\") " Nov 25 11:17:02 crc kubenswrapper[4821]: I1125 11:17:02.373451 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/583b83b9-f7b0-451d-96ce-81b4f7a2041a-utilities" (OuterVolumeSpecName: "utilities") pod "583b83b9-f7b0-451d-96ce-81b4f7a2041a" (UID: "583b83b9-f7b0-451d-96ce-81b4f7a2041a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:17:02 crc kubenswrapper[4821]: I1125 11:17:02.380429 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/583b83b9-f7b0-451d-96ce-81b4f7a2041a-kube-api-access-nks8h" (OuterVolumeSpecName: "kube-api-access-nks8h") pod "583b83b9-f7b0-451d-96ce-81b4f7a2041a" (UID: "583b83b9-f7b0-451d-96ce-81b4f7a2041a"). InnerVolumeSpecName "kube-api-access-nks8h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:17:02 crc kubenswrapper[4821]: I1125 11:17:02.392313 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/583b83b9-f7b0-451d-96ce-81b4f7a2041a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "583b83b9-f7b0-451d-96ce-81b4f7a2041a" (UID: "583b83b9-f7b0-451d-96ce-81b4f7a2041a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:17:02 crc kubenswrapper[4821]: I1125 11:17:02.474554 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nks8h\" (UniqueName: \"kubernetes.io/projected/583b83b9-f7b0-451d-96ce-81b4f7a2041a-kube-api-access-nks8h\") on node \"crc\" DevicePath \"\"" Nov 25 11:17:02 crc kubenswrapper[4821]: I1125 11:17:02.474592 4821 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/583b83b9-f7b0-451d-96ce-81b4f7a2041a-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 11:17:02 crc kubenswrapper[4821]: I1125 11:17:02.474601 4821 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/583b83b9-f7b0-451d-96ce-81b4f7a2041a-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 11:17:02 crc kubenswrapper[4821]: I1125 11:17:02.637187 4821 generic.go:334] "Generic (PLEG): container finished" podID="583b83b9-f7b0-451d-96ce-81b4f7a2041a" containerID="2335ea31c7ae0e4881551d6d3259d1d2e29f13cb5df67217dca7c688b4a3d485" exitCode=0 Nov 25 11:17:02 crc kubenswrapper[4821]: I1125 11:17:02.637243 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2q5h7" Nov 25 11:17:02 crc kubenswrapper[4821]: I1125 11:17:02.637245 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2q5h7" event={"ID":"583b83b9-f7b0-451d-96ce-81b4f7a2041a","Type":"ContainerDied","Data":"2335ea31c7ae0e4881551d6d3259d1d2e29f13cb5df67217dca7c688b4a3d485"} Nov 25 11:17:02 crc kubenswrapper[4821]: I1125 11:17:02.637412 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2q5h7" event={"ID":"583b83b9-f7b0-451d-96ce-81b4f7a2041a","Type":"ContainerDied","Data":"bc4dc420782b550b45f8b7a93844824877a0bb67d709fd15599ac4eb7fafe3f0"} Nov 25 11:17:02 crc kubenswrapper[4821]: I1125 11:17:02.637442 4821 scope.go:117] "RemoveContainer" containerID="2335ea31c7ae0e4881551d6d3259d1d2e29f13cb5df67217dca7c688b4a3d485" Nov 25 11:17:02 crc kubenswrapper[4821]: I1125 11:17:02.657667 4821 scope.go:117] "RemoveContainer" containerID="b8fc1ada99d9bd0a58b30efcca01f466e10a46647178117c6d16b3bbd865f207" Nov 25 11:17:02 crc kubenswrapper[4821]: I1125 11:17:02.686738 4821 scope.go:117] "RemoveContainer" containerID="4a567484677ce930428412f7ba8d1ea0aa750275f76259d03b6d3447b0c80cef" Nov 25 11:17:02 crc kubenswrapper[4821]: I1125 11:17:02.699324 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2q5h7"] Nov 25 11:17:02 crc kubenswrapper[4821]: I1125 11:17:02.707947 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-2q5h7"] Nov 25 11:17:02 crc kubenswrapper[4821]: I1125 11:17:02.741106 4821 scope.go:117] "RemoveContainer" containerID="2335ea31c7ae0e4881551d6d3259d1d2e29f13cb5df67217dca7c688b4a3d485" Nov 25 11:17:02 crc kubenswrapper[4821]: E1125 11:17:02.741688 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2335ea31c7ae0e4881551d6d3259d1d2e29f13cb5df67217dca7c688b4a3d485\": container with ID starting with 2335ea31c7ae0e4881551d6d3259d1d2e29f13cb5df67217dca7c688b4a3d485 not found: ID does not exist" containerID="2335ea31c7ae0e4881551d6d3259d1d2e29f13cb5df67217dca7c688b4a3d485" Nov 25 11:17:02 crc kubenswrapper[4821]: I1125 11:17:02.741718 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2335ea31c7ae0e4881551d6d3259d1d2e29f13cb5df67217dca7c688b4a3d485"} err="failed to get container status \"2335ea31c7ae0e4881551d6d3259d1d2e29f13cb5df67217dca7c688b4a3d485\": rpc error: code = NotFound desc = could not find container \"2335ea31c7ae0e4881551d6d3259d1d2e29f13cb5df67217dca7c688b4a3d485\": container with ID starting with 2335ea31c7ae0e4881551d6d3259d1d2e29f13cb5df67217dca7c688b4a3d485 not found: ID does not exist" Nov 25 11:17:02 crc kubenswrapper[4821]: I1125 11:17:02.741737 4821 scope.go:117] "RemoveContainer" containerID="b8fc1ada99d9bd0a58b30efcca01f466e10a46647178117c6d16b3bbd865f207" Nov 25 11:17:02 crc kubenswrapper[4821]: E1125 11:17:02.742206 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b8fc1ada99d9bd0a58b30efcca01f466e10a46647178117c6d16b3bbd865f207\": container with ID starting with b8fc1ada99d9bd0a58b30efcca01f466e10a46647178117c6d16b3bbd865f207 not found: ID does not exist" containerID="b8fc1ada99d9bd0a58b30efcca01f466e10a46647178117c6d16b3bbd865f207" Nov 25 11:17:02 crc kubenswrapper[4821]: I1125 11:17:02.742238 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b8fc1ada99d9bd0a58b30efcca01f466e10a46647178117c6d16b3bbd865f207"} err="failed to get container status \"b8fc1ada99d9bd0a58b30efcca01f466e10a46647178117c6d16b3bbd865f207\": rpc error: code = NotFound desc = could not find container \"b8fc1ada99d9bd0a58b30efcca01f466e10a46647178117c6d16b3bbd865f207\": container with ID starting with b8fc1ada99d9bd0a58b30efcca01f466e10a46647178117c6d16b3bbd865f207 not found: ID does not exist" Nov 25 11:17:02 crc kubenswrapper[4821]: I1125 11:17:02.742251 4821 scope.go:117] "RemoveContainer" containerID="4a567484677ce930428412f7ba8d1ea0aa750275f76259d03b6d3447b0c80cef" Nov 25 11:17:02 crc kubenswrapper[4821]: E1125 11:17:02.742537 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a567484677ce930428412f7ba8d1ea0aa750275f76259d03b6d3447b0c80cef\": container with ID starting with 4a567484677ce930428412f7ba8d1ea0aa750275f76259d03b6d3447b0c80cef not found: ID does not exist" containerID="4a567484677ce930428412f7ba8d1ea0aa750275f76259d03b6d3447b0c80cef" Nov 25 11:17:02 crc kubenswrapper[4821]: I1125 11:17:02.742566 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a567484677ce930428412f7ba8d1ea0aa750275f76259d03b6d3447b0c80cef"} err="failed to get container status \"4a567484677ce930428412f7ba8d1ea0aa750275f76259d03b6d3447b0c80cef\": rpc error: code = NotFound desc = could not find container \"4a567484677ce930428412f7ba8d1ea0aa750275f76259d03b6d3447b0c80cef\": container with ID starting with 4a567484677ce930428412f7ba8d1ea0aa750275f76259d03b6d3447b0c80cef not found: ID does not exist" Nov 25 11:17:04 crc kubenswrapper[4821]: I1125 11:17:04.126122 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="583b83b9-f7b0-451d-96ce-81b4f7a2041a" path="/var/lib/kubelet/pods/583b83b9-f7b0-451d-96ce-81b4f7a2041a/volumes" Nov 25 11:17:37 crc kubenswrapper[4821]: I1125 11:17:37.443981 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Nov 25 11:17:37 crc kubenswrapper[4821]: E1125 11:17:37.444997 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="583b83b9-f7b0-451d-96ce-81b4f7a2041a" containerName="registry-server" Nov 25 11:17:37 crc kubenswrapper[4821]: I1125 11:17:37.445012 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="583b83b9-f7b0-451d-96ce-81b4f7a2041a" containerName="registry-server" Nov 25 11:17:37 crc kubenswrapper[4821]: E1125 11:17:37.445034 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="583b83b9-f7b0-451d-96ce-81b4f7a2041a" containerName="extract-utilities" Nov 25 11:17:37 crc kubenswrapper[4821]: I1125 11:17:37.445041 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="583b83b9-f7b0-451d-96ce-81b4f7a2041a" containerName="extract-utilities" Nov 25 11:17:37 crc kubenswrapper[4821]: E1125 11:17:37.445058 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="583b83b9-f7b0-451d-96ce-81b4f7a2041a" containerName="extract-content" Nov 25 11:17:37 crc kubenswrapper[4821]: I1125 11:17:37.445065 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="583b83b9-f7b0-451d-96ce-81b4f7a2041a" containerName="extract-content" Nov 25 11:17:37 crc kubenswrapper[4821]: I1125 11:17:37.445316 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="583b83b9-f7b0-451d-96ce-81b4f7a2041a" containerName="registry-server" Nov 25 11:17:37 crc kubenswrapper[4821]: I1125 11:17:37.446047 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 25 11:17:37 crc kubenswrapper[4821]: I1125 11:17:37.448367 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Nov 25 11:17:37 crc kubenswrapper[4821]: I1125 11:17:37.449360 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-qrs9n" Nov 25 11:17:37 crc kubenswrapper[4821]: I1125 11:17:37.449437 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Nov 25 11:17:37 crc kubenswrapper[4821]: I1125 11:17:37.449708 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Nov 25 11:17:37 crc kubenswrapper[4821]: I1125 11:17:37.459680 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Nov 25 11:17:37 crc kubenswrapper[4821]: I1125 11:17:37.547088 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/d8aea755-bf2f-4443-8196-46936ce3e2fd-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"d8aea755-bf2f-4443-8196-46936ce3e2fd\") " pod="openstack/tempest-tests-tempest" Nov 25 11:17:37 crc kubenswrapper[4821]: I1125 11:17:37.547453 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zz66w\" (UniqueName: \"kubernetes.io/projected/d8aea755-bf2f-4443-8196-46936ce3e2fd-kube-api-access-zz66w\") pod \"tempest-tests-tempest\" (UID: \"d8aea755-bf2f-4443-8196-46936ce3e2fd\") " pod="openstack/tempest-tests-tempest" Nov 25 11:17:37 crc kubenswrapper[4821]: I1125 11:17:37.547606 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d8aea755-bf2f-4443-8196-46936ce3e2fd-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"d8aea755-bf2f-4443-8196-46936ce3e2fd\") " pod="openstack/tempest-tests-tempest" Nov 25 11:17:37 crc kubenswrapper[4821]: I1125 11:17:37.547839 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d8aea755-bf2f-4443-8196-46936ce3e2fd-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"d8aea755-bf2f-4443-8196-46936ce3e2fd\") " pod="openstack/tempest-tests-tempest" Nov 25 11:17:37 crc kubenswrapper[4821]: I1125 11:17:37.547998 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"tempest-tests-tempest\" (UID: \"d8aea755-bf2f-4443-8196-46936ce3e2fd\") " pod="openstack/tempest-tests-tempest" Nov 25 11:17:37 crc kubenswrapper[4821]: I1125 11:17:37.548111 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d8aea755-bf2f-4443-8196-46936ce3e2fd-config-data\") pod \"tempest-tests-tempest\" (UID: \"d8aea755-bf2f-4443-8196-46936ce3e2fd\") " pod="openstack/tempest-tests-tempest" Nov 25 11:17:37 crc kubenswrapper[4821]: I1125 11:17:37.548279 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/d8aea755-bf2f-4443-8196-46936ce3e2fd-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"d8aea755-bf2f-4443-8196-46936ce3e2fd\") " pod="openstack/tempest-tests-tempest" Nov 25 11:17:37 crc kubenswrapper[4821]: I1125 11:17:37.548394 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d8aea755-bf2f-4443-8196-46936ce3e2fd-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"d8aea755-bf2f-4443-8196-46936ce3e2fd\") " pod="openstack/tempest-tests-tempest" Nov 25 11:17:37 crc kubenswrapper[4821]: I1125 11:17:37.548497 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/d8aea755-bf2f-4443-8196-46936ce3e2fd-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"d8aea755-bf2f-4443-8196-46936ce3e2fd\") " pod="openstack/tempest-tests-tempest" Nov 25 11:17:37 crc kubenswrapper[4821]: I1125 11:17:37.650110 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/d8aea755-bf2f-4443-8196-46936ce3e2fd-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"d8aea755-bf2f-4443-8196-46936ce3e2fd\") " pod="openstack/tempest-tests-tempest" Nov 25 11:17:37 crc kubenswrapper[4821]: I1125 11:17:37.650187 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zz66w\" (UniqueName: \"kubernetes.io/projected/d8aea755-bf2f-4443-8196-46936ce3e2fd-kube-api-access-zz66w\") pod \"tempest-tests-tempest\" (UID: \"d8aea755-bf2f-4443-8196-46936ce3e2fd\") " pod="openstack/tempest-tests-tempest" Nov 25 11:17:37 crc kubenswrapper[4821]: I1125 11:17:37.650222 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d8aea755-bf2f-4443-8196-46936ce3e2fd-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"d8aea755-bf2f-4443-8196-46936ce3e2fd\") " pod="openstack/tempest-tests-tempest" Nov 25 11:17:37 crc kubenswrapper[4821]: I1125 11:17:37.650263 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d8aea755-bf2f-4443-8196-46936ce3e2fd-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"d8aea755-bf2f-4443-8196-46936ce3e2fd\") " pod="openstack/tempest-tests-tempest" Nov 25 11:17:37 crc kubenswrapper[4821]: I1125 11:17:37.650316 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"tempest-tests-tempest\" (UID: \"d8aea755-bf2f-4443-8196-46936ce3e2fd\") " pod="openstack/tempest-tests-tempest" Nov 25 11:17:37 crc kubenswrapper[4821]: I1125 11:17:37.650346 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d8aea755-bf2f-4443-8196-46936ce3e2fd-config-data\") pod \"tempest-tests-tempest\" (UID: \"d8aea755-bf2f-4443-8196-46936ce3e2fd\") " pod="openstack/tempest-tests-tempest" Nov 25 11:17:37 crc kubenswrapper[4821]: I1125 11:17:37.650374 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/d8aea755-bf2f-4443-8196-46936ce3e2fd-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"d8aea755-bf2f-4443-8196-46936ce3e2fd\") " pod="openstack/tempest-tests-tempest" Nov 25 11:17:37 crc kubenswrapper[4821]: I1125 11:17:37.650404 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d8aea755-bf2f-4443-8196-46936ce3e2fd-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"d8aea755-bf2f-4443-8196-46936ce3e2fd\") " pod="openstack/tempest-tests-tempest" Nov 25 11:17:37 crc kubenswrapper[4821]: I1125 11:17:37.650432 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/d8aea755-bf2f-4443-8196-46936ce3e2fd-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"d8aea755-bf2f-4443-8196-46936ce3e2fd\") " pod="openstack/tempest-tests-tempest" Nov 25 11:17:37 crc kubenswrapper[4821]: I1125 11:17:37.650929 4821 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"tempest-tests-tempest\" (UID: \"d8aea755-bf2f-4443-8196-46936ce3e2fd\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/tempest-tests-tempest" Nov 25 11:17:37 crc kubenswrapper[4821]: I1125 11:17:37.651392 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/d8aea755-bf2f-4443-8196-46936ce3e2fd-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"d8aea755-bf2f-4443-8196-46936ce3e2fd\") " pod="openstack/tempest-tests-tempest" Nov 25 11:17:37 crc kubenswrapper[4821]: I1125 11:17:37.652499 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d8aea755-bf2f-4443-8196-46936ce3e2fd-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"d8aea755-bf2f-4443-8196-46936ce3e2fd\") " pod="openstack/tempest-tests-tempest" Nov 25 11:17:37 crc kubenswrapper[4821]: I1125 11:17:37.652907 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d8aea755-bf2f-4443-8196-46936ce3e2fd-config-data\") pod \"tempest-tests-tempest\" (UID: \"d8aea755-bf2f-4443-8196-46936ce3e2fd\") " pod="openstack/tempest-tests-tempest" Nov 25 11:17:37 crc kubenswrapper[4821]: I1125 11:17:37.653575 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/d8aea755-bf2f-4443-8196-46936ce3e2fd-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"d8aea755-bf2f-4443-8196-46936ce3e2fd\") " pod="openstack/tempest-tests-tempest" Nov 25 11:17:37 crc kubenswrapper[4821]: I1125 11:17:37.656007 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d8aea755-bf2f-4443-8196-46936ce3e2fd-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"d8aea755-bf2f-4443-8196-46936ce3e2fd\") " pod="openstack/tempest-tests-tempest" Nov 25 11:17:37 crc kubenswrapper[4821]: I1125 11:17:37.656624 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/d8aea755-bf2f-4443-8196-46936ce3e2fd-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"d8aea755-bf2f-4443-8196-46936ce3e2fd\") " pod="openstack/tempest-tests-tempest" Nov 25 11:17:37 crc kubenswrapper[4821]: I1125 11:17:37.657480 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d8aea755-bf2f-4443-8196-46936ce3e2fd-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"d8aea755-bf2f-4443-8196-46936ce3e2fd\") " pod="openstack/tempest-tests-tempest" Nov 25 11:17:37 crc kubenswrapper[4821]: I1125 11:17:37.668405 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zz66w\" (UniqueName: \"kubernetes.io/projected/d8aea755-bf2f-4443-8196-46936ce3e2fd-kube-api-access-zz66w\") pod \"tempest-tests-tempest\" (UID: \"d8aea755-bf2f-4443-8196-46936ce3e2fd\") " pod="openstack/tempest-tests-tempest" Nov 25 11:17:37 crc kubenswrapper[4821]: I1125 11:17:37.678710 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"tempest-tests-tempest\" (UID: \"d8aea755-bf2f-4443-8196-46936ce3e2fd\") " pod="openstack/tempest-tests-tempest" Nov 25 11:17:37 crc kubenswrapper[4821]: I1125 11:17:37.817563 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 25 11:17:38 crc kubenswrapper[4821]: I1125 11:17:38.276335 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Nov 25 11:17:38 crc kubenswrapper[4821]: W1125 11:17:38.295586 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd8aea755_bf2f_4443_8196_46936ce3e2fd.slice/crio-2dc6c27588da93a980af00c534f445a80ec1a579e213d38d1d64998bc6cebf78 WatchSource:0}: Error finding container 2dc6c27588da93a980af00c534f445a80ec1a579e213d38d1d64998bc6cebf78: Status 404 returned error can't find the container with id 2dc6c27588da93a980af00c534f445a80ec1a579e213d38d1d64998bc6cebf78 Nov 25 11:17:38 crc kubenswrapper[4821]: I1125 11:17:38.963732 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"d8aea755-bf2f-4443-8196-46936ce3e2fd","Type":"ContainerStarted","Data":"2dc6c27588da93a980af00c534f445a80ec1a579e213d38d1d64998bc6cebf78"} Nov 25 11:17:52 crc kubenswrapper[4821]: I1125 11:17:52.547634 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-2mzwd"] Nov 25 11:17:52 crc kubenswrapper[4821]: I1125 11:17:52.550714 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2mzwd" Nov 25 11:17:52 crc kubenswrapper[4821]: I1125 11:17:52.556631 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2mzwd"] Nov 25 11:17:52 crc kubenswrapper[4821]: I1125 11:17:52.649565 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dflws\" (UniqueName: \"kubernetes.io/projected/8dcb623e-e543-4129-a6ea-695c314bad9d-kube-api-access-dflws\") pod \"redhat-operators-2mzwd\" (UID: \"8dcb623e-e543-4129-a6ea-695c314bad9d\") " pod="openshift-marketplace/redhat-operators-2mzwd" Nov 25 11:17:52 crc kubenswrapper[4821]: I1125 11:17:52.649634 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8dcb623e-e543-4129-a6ea-695c314bad9d-catalog-content\") pod \"redhat-operators-2mzwd\" (UID: \"8dcb623e-e543-4129-a6ea-695c314bad9d\") " pod="openshift-marketplace/redhat-operators-2mzwd" Nov 25 11:17:52 crc kubenswrapper[4821]: I1125 11:17:52.649689 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8dcb623e-e543-4129-a6ea-695c314bad9d-utilities\") pod \"redhat-operators-2mzwd\" (UID: \"8dcb623e-e543-4129-a6ea-695c314bad9d\") " pod="openshift-marketplace/redhat-operators-2mzwd" Nov 25 11:17:52 crc kubenswrapper[4821]: I1125 11:17:52.751078 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dflws\" (UniqueName: \"kubernetes.io/projected/8dcb623e-e543-4129-a6ea-695c314bad9d-kube-api-access-dflws\") pod \"redhat-operators-2mzwd\" (UID: \"8dcb623e-e543-4129-a6ea-695c314bad9d\") " pod="openshift-marketplace/redhat-operators-2mzwd" Nov 25 11:17:52 crc kubenswrapper[4821]: I1125 11:17:52.751127 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8dcb623e-e543-4129-a6ea-695c314bad9d-catalog-content\") pod \"redhat-operators-2mzwd\" (UID: \"8dcb623e-e543-4129-a6ea-695c314bad9d\") " pod="openshift-marketplace/redhat-operators-2mzwd" Nov 25 11:17:52 crc kubenswrapper[4821]: I1125 11:17:52.751177 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8dcb623e-e543-4129-a6ea-695c314bad9d-utilities\") pod \"redhat-operators-2mzwd\" (UID: \"8dcb623e-e543-4129-a6ea-695c314bad9d\") " pod="openshift-marketplace/redhat-operators-2mzwd" Nov 25 11:17:52 crc kubenswrapper[4821]: I1125 11:17:52.757351 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8dcb623e-e543-4129-a6ea-695c314bad9d-utilities\") pod \"redhat-operators-2mzwd\" (UID: \"8dcb623e-e543-4129-a6ea-695c314bad9d\") " pod="openshift-marketplace/redhat-operators-2mzwd" Nov 25 11:17:52 crc kubenswrapper[4821]: I1125 11:17:52.757390 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8dcb623e-e543-4129-a6ea-695c314bad9d-catalog-content\") pod \"redhat-operators-2mzwd\" (UID: \"8dcb623e-e543-4129-a6ea-695c314bad9d\") " pod="openshift-marketplace/redhat-operators-2mzwd" Nov 25 11:17:52 crc kubenswrapper[4821]: I1125 11:17:52.792617 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dflws\" (UniqueName: \"kubernetes.io/projected/8dcb623e-e543-4129-a6ea-695c314bad9d-kube-api-access-dflws\") pod \"redhat-operators-2mzwd\" (UID: \"8dcb623e-e543-4129-a6ea-695c314bad9d\") " pod="openshift-marketplace/redhat-operators-2mzwd" Nov 25 11:17:52 crc kubenswrapper[4821]: I1125 11:17:52.916529 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2mzwd" Nov 25 11:17:58 crc kubenswrapper[4821]: I1125 11:17:58.833487 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-ncn9h"] Nov 25 11:17:58 crc kubenswrapper[4821]: I1125 11:17:58.837456 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ncn9h" Nov 25 11:17:58 crc kubenswrapper[4821]: I1125 11:17:58.842552 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ncn9h"] Nov 25 11:17:58 crc kubenswrapper[4821]: I1125 11:17:58.877862 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5e191cd-eb4d-41d7-b634-a7ee707fd960-catalog-content\") pod \"community-operators-ncn9h\" (UID: \"f5e191cd-eb4d-41d7-b634-a7ee707fd960\") " pod="openshift-marketplace/community-operators-ncn9h" Nov 25 11:17:58 crc kubenswrapper[4821]: I1125 11:17:58.878062 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5e191cd-eb4d-41d7-b634-a7ee707fd960-utilities\") pod \"community-operators-ncn9h\" (UID: \"f5e191cd-eb4d-41d7-b634-a7ee707fd960\") " pod="openshift-marketplace/community-operators-ncn9h" Nov 25 11:17:58 crc kubenswrapper[4821]: I1125 11:17:58.878103 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h88t9\" (UniqueName: \"kubernetes.io/projected/f5e191cd-eb4d-41d7-b634-a7ee707fd960-kube-api-access-h88t9\") pod \"community-operators-ncn9h\" (UID: \"f5e191cd-eb4d-41d7-b634-a7ee707fd960\") " pod="openshift-marketplace/community-operators-ncn9h" Nov 25 11:17:58 crc kubenswrapper[4821]: I1125 11:17:58.980206 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5e191cd-eb4d-41d7-b634-a7ee707fd960-utilities\") pod \"community-operators-ncn9h\" (UID: \"f5e191cd-eb4d-41d7-b634-a7ee707fd960\") " pod="openshift-marketplace/community-operators-ncn9h" Nov 25 11:17:58 crc kubenswrapper[4821]: I1125 11:17:58.980266 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h88t9\" (UniqueName: \"kubernetes.io/projected/f5e191cd-eb4d-41d7-b634-a7ee707fd960-kube-api-access-h88t9\") pod \"community-operators-ncn9h\" (UID: \"f5e191cd-eb4d-41d7-b634-a7ee707fd960\") " pod="openshift-marketplace/community-operators-ncn9h" Nov 25 11:17:58 crc kubenswrapper[4821]: I1125 11:17:58.980373 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5e191cd-eb4d-41d7-b634-a7ee707fd960-catalog-content\") pod \"community-operators-ncn9h\" (UID: \"f5e191cd-eb4d-41d7-b634-a7ee707fd960\") " pod="openshift-marketplace/community-operators-ncn9h" Nov 25 11:17:58 crc kubenswrapper[4821]: I1125 11:17:58.980749 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5e191cd-eb4d-41d7-b634-a7ee707fd960-utilities\") pod \"community-operators-ncn9h\" (UID: \"f5e191cd-eb4d-41d7-b634-a7ee707fd960\") " pod="openshift-marketplace/community-operators-ncn9h" Nov 25 11:17:58 crc kubenswrapper[4821]: I1125 11:17:58.980978 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5e191cd-eb4d-41d7-b634-a7ee707fd960-catalog-content\") pod \"community-operators-ncn9h\" (UID: \"f5e191cd-eb4d-41d7-b634-a7ee707fd960\") " pod="openshift-marketplace/community-operators-ncn9h" Nov 25 11:17:59 crc kubenswrapper[4821]: I1125 11:17:59.022131 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h88t9\" (UniqueName: \"kubernetes.io/projected/f5e191cd-eb4d-41d7-b634-a7ee707fd960-kube-api-access-h88t9\") pod \"community-operators-ncn9h\" (UID: \"f5e191cd-eb4d-41d7-b634-a7ee707fd960\") " pod="openshift-marketplace/community-operators-ncn9h" Nov 25 11:17:59 crc kubenswrapper[4821]: I1125 11:17:59.173094 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ncn9h" Nov 25 11:18:10 crc kubenswrapper[4821]: E1125 11:18:10.363916 4821 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Nov 25 11:18:10 crc kubenswrapper[4821]: E1125 11:18:10.364821 4821 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zz66w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(d8aea755-bf2f-4443-8196-46936ce3e2fd): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 11:18:10 crc kubenswrapper[4821]: E1125 11:18:10.366024 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="d8aea755-bf2f-4443-8196-46936ce3e2fd" Nov 25 11:18:10 crc kubenswrapper[4821]: I1125 11:18:10.752450 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ncn9h"] Nov 25 11:18:10 crc kubenswrapper[4821]: W1125 11:18:10.800533 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8dcb623e_e543_4129_a6ea_695c314bad9d.slice/crio-1218517c6ab7061ed69d5b0ba0ac353e041563c5972496659c4d5c7984b8c6e7 WatchSource:0}: Error finding container 1218517c6ab7061ed69d5b0ba0ac353e041563c5972496659c4d5c7984b8c6e7: Status 404 returned error can't find the container with id 1218517c6ab7061ed69d5b0ba0ac353e041563c5972496659c4d5c7984b8c6e7 Nov 25 11:18:10 crc kubenswrapper[4821]: I1125 11:18:10.803824 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2mzwd"] Nov 25 11:18:11 crc kubenswrapper[4821]: I1125 11:18:11.301343 4821 generic.go:334] "Generic (PLEG): container finished" podID="8dcb623e-e543-4129-a6ea-695c314bad9d" containerID="2813b4502c262487863adf34670dcb5be394f14dcc0b0ec905905b0a485775f7" exitCode=0 Nov 25 11:18:11 crc kubenswrapper[4821]: I1125 11:18:11.301456 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2mzwd" event={"ID":"8dcb623e-e543-4129-a6ea-695c314bad9d","Type":"ContainerDied","Data":"2813b4502c262487863adf34670dcb5be394f14dcc0b0ec905905b0a485775f7"} Nov 25 11:18:11 crc kubenswrapper[4821]: I1125 11:18:11.301506 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2mzwd" event={"ID":"8dcb623e-e543-4129-a6ea-695c314bad9d","Type":"ContainerStarted","Data":"1218517c6ab7061ed69d5b0ba0ac353e041563c5972496659c4d5c7984b8c6e7"} Nov 25 11:18:11 crc kubenswrapper[4821]: I1125 11:18:11.309232 4821 generic.go:334] "Generic (PLEG): container finished" podID="f5e191cd-eb4d-41d7-b634-a7ee707fd960" containerID="81e99aee9a5302cd1905ad34f90c330f4133a7c0232759f6b5f74de9957e7abe" exitCode=0 Nov 25 11:18:11 crc kubenswrapper[4821]: I1125 11:18:11.309338 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ncn9h" event={"ID":"f5e191cd-eb4d-41d7-b634-a7ee707fd960","Type":"ContainerDied","Data":"81e99aee9a5302cd1905ad34f90c330f4133a7c0232759f6b5f74de9957e7abe"} Nov 25 11:18:11 crc kubenswrapper[4821]: I1125 11:18:11.309401 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ncn9h" event={"ID":"f5e191cd-eb4d-41d7-b634-a7ee707fd960","Type":"ContainerStarted","Data":"a2971de0f8cbe82afb051c553655a6515b886a75d2da276cf7a23fbf403ca1ab"} Nov 25 11:18:11 crc kubenswrapper[4821]: E1125 11:18:11.311682 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="d8aea755-bf2f-4443-8196-46936ce3e2fd" Nov 25 11:18:13 crc kubenswrapper[4821]: I1125 11:18:13.328863 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2mzwd" event={"ID":"8dcb623e-e543-4129-a6ea-695c314bad9d","Type":"ContainerStarted","Data":"252aefe9a538cea2e00a9a9d44d771e5951e7789b8fe16e82b5d4d405661c49d"} Nov 25 11:18:13 crc kubenswrapper[4821]: I1125 11:18:13.331114 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ncn9h" event={"ID":"f5e191cd-eb4d-41d7-b634-a7ee707fd960","Type":"ContainerStarted","Data":"0a1680a5af79c8f6fed01fa20b577f56dc9d3aa91c7e45a682e818f7426a2caf"} Nov 25 11:18:14 crc kubenswrapper[4821]: I1125 11:18:14.339773 4821 generic.go:334] "Generic (PLEG): container finished" podID="8dcb623e-e543-4129-a6ea-695c314bad9d" containerID="252aefe9a538cea2e00a9a9d44d771e5951e7789b8fe16e82b5d4d405661c49d" exitCode=0 Nov 25 11:18:14 crc kubenswrapper[4821]: I1125 11:18:14.339897 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2mzwd" event={"ID":"8dcb623e-e543-4129-a6ea-695c314bad9d","Type":"ContainerDied","Data":"252aefe9a538cea2e00a9a9d44d771e5951e7789b8fe16e82b5d4d405661c49d"} Nov 25 11:18:14 crc kubenswrapper[4821]: I1125 11:18:14.345847 4821 generic.go:334] "Generic (PLEG): container finished" podID="f5e191cd-eb4d-41d7-b634-a7ee707fd960" containerID="0a1680a5af79c8f6fed01fa20b577f56dc9d3aa91c7e45a682e818f7426a2caf" exitCode=0 Nov 25 11:18:14 crc kubenswrapper[4821]: I1125 11:18:14.346048 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ncn9h" event={"ID":"f5e191cd-eb4d-41d7-b634-a7ee707fd960","Type":"ContainerDied","Data":"0a1680a5af79c8f6fed01fa20b577f56dc9d3aa91c7e45a682e818f7426a2caf"} Nov 25 11:18:16 crc kubenswrapper[4821]: I1125 11:18:16.368144 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2mzwd" event={"ID":"8dcb623e-e543-4129-a6ea-695c314bad9d","Type":"ContainerStarted","Data":"281a63a60823b3cc2329d6235521646e367fe5d71785d3886299fd410bc11171"} Nov 25 11:18:16 crc kubenswrapper[4821]: I1125 11:18:16.370355 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ncn9h" event={"ID":"f5e191cd-eb4d-41d7-b634-a7ee707fd960","Type":"ContainerStarted","Data":"4a5c8744a3b8c13ab40ae81676b89a7be3d6bfd90939cb924cdde2bb4bae1697"} Nov 25 11:18:16 crc kubenswrapper[4821]: I1125 11:18:16.388328 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-2mzwd" podStartSLOduration=20.340571193 podStartE2EDuration="24.388312481s" podCreationTimestamp="2025-11-25 11:17:52 +0000 UTC" firstStartedPulling="2025-11-25 11:18:11.306226751 +0000 UTC m=+2761.842546598" lastFinishedPulling="2025-11-25 11:18:15.353968039 +0000 UTC m=+2765.890287886" observedRunningTime="2025-11-25 11:18:16.385800229 +0000 UTC m=+2766.922120076" watchObservedRunningTime="2025-11-25 11:18:16.388312481 +0000 UTC m=+2766.924632328" Nov 25 11:18:16 crc kubenswrapper[4821]: I1125 11:18:16.409218 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-ncn9h" podStartSLOduration=14.07223311 podStartE2EDuration="18.40919773s" podCreationTimestamp="2025-11-25 11:17:58 +0000 UTC" firstStartedPulling="2025-11-25 11:18:11.311598371 +0000 UTC m=+2761.847918218" lastFinishedPulling="2025-11-25 11:18:15.648562991 +0000 UTC m=+2766.184882838" observedRunningTime="2025-11-25 11:18:16.400263833 +0000 UTC m=+2766.936583700" watchObservedRunningTime="2025-11-25 11:18:16.40919773 +0000 UTC m=+2766.945517597" Nov 25 11:18:19 crc kubenswrapper[4821]: I1125 11:18:19.174267 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-ncn9h" Nov 25 11:18:19 crc kubenswrapper[4821]: I1125 11:18:19.174540 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-ncn9h" Nov 25 11:18:19 crc kubenswrapper[4821]: I1125 11:18:19.217284 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-ncn9h" Nov 25 11:18:22 crc kubenswrapper[4821]: I1125 11:18:22.917524 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-2mzwd" Nov 25 11:18:22 crc kubenswrapper[4821]: I1125 11:18:22.918008 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-2mzwd" Nov 25 11:18:22 crc kubenswrapper[4821]: I1125 11:18:22.966973 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-2mzwd" Nov 25 11:18:23 crc kubenswrapper[4821]: I1125 11:18:23.478071 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-2mzwd" Nov 25 11:18:23 crc kubenswrapper[4821]: I1125 11:18:23.738951 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2mzwd"] Nov 25 11:18:25 crc kubenswrapper[4821]: I1125 11:18:25.449624 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-2mzwd" podUID="8dcb623e-e543-4129-a6ea-695c314bad9d" containerName="registry-server" containerID="cri-o://281a63a60823b3cc2329d6235521646e367fe5d71785d3886299fd410bc11171" gracePeriod=2 Nov 25 11:18:26 crc kubenswrapper[4821]: I1125 11:18:26.461939 4821 generic.go:334] "Generic (PLEG): container finished" podID="8dcb623e-e543-4129-a6ea-695c314bad9d" containerID="281a63a60823b3cc2329d6235521646e367fe5d71785d3886299fd410bc11171" exitCode=0 Nov 25 11:18:26 crc kubenswrapper[4821]: I1125 11:18:26.462277 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2mzwd" event={"ID":"8dcb623e-e543-4129-a6ea-695c314bad9d","Type":"ContainerDied","Data":"281a63a60823b3cc2329d6235521646e367fe5d71785d3886299fd410bc11171"} Nov 25 11:18:26 crc kubenswrapper[4821]: I1125 11:18:26.475287 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Nov 25 11:18:26 crc kubenswrapper[4821]: I1125 11:18:26.682032 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2mzwd" Nov 25 11:18:26 crc kubenswrapper[4821]: I1125 11:18:26.849219 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8dcb623e-e543-4129-a6ea-695c314bad9d-utilities\") pod \"8dcb623e-e543-4129-a6ea-695c314bad9d\" (UID: \"8dcb623e-e543-4129-a6ea-695c314bad9d\") " Nov 25 11:18:26 crc kubenswrapper[4821]: I1125 11:18:26.849283 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8dcb623e-e543-4129-a6ea-695c314bad9d-catalog-content\") pod \"8dcb623e-e543-4129-a6ea-695c314bad9d\" (UID: \"8dcb623e-e543-4129-a6ea-695c314bad9d\") " Nov 25 11:18:26 crc kubenswrapper[4821]: I1125 11:18:26.849323 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dflws\" (UniqueName: \"kubernetes.io/projected/8dcb623e-e543-4129-a6ea-695c314bad9d-kube-api-access-dflws\") pod \"8dcb623e-e543-4129-a6ea-695c314bad9d\" (UID: \"8dcb623e-e543-4129-a6ea-695c314bad9d\") " Nov 25 11:18:26 crc kubenswrapper[4821]: I1125 11:18:26.851221 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8dcb623e-e543-4129-a6ea-695c314bad9d-utilities" (OuterVolumeSpecName: "utilities") pod "8dcb623e-e543-4129-a6ea-695c314bad9d" (UID: "8dcb623e-e543-4129-a6ea-695c314bad9d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:18:26 crc kubenswrapper[4821]: I1125 11:18:26.867817 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8dcb623e-e543-4129-a6ea-695c314bad9d-kube-api-access-dflws" (OuterVolumeSpecName: "kube-api-access-dflws") pod "8dcb623e-e543-4129-a6ea-695c314bad9d" (UID: "8dcb623e-e543-4129-a6ea-695c314bad9d"). InnerVolumeSpecName "kube-api-access-dflws". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:18:26 crc kubenswrapper[4821]: I1125 11:18:26.938787 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8dcb623e-e543-4129-a6ea-695c314bad9d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8dcb623e-e543-4129-a6ea-695c314bad9d" (UID: "8dcb623e-e543-4129-a6ea-695c314bad9d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:18:26 crc kubenswrapper[4821]: I1125 11:18:26.952032 4821 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8dcb623e-e543-4129-a6ea-695c314bad9d-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 11:18:26 crc kubenswrapper[4821]: I1125 11:18:26.952073 4821 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8dcb623e-e543-4129-a6ea-695c314bad9d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 11:18:26 crc kubenswrapper[4821]: I1125 11:18:26.952086 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dflws\" (UniqueName: \"kubernetes.io/projected/8dcb623e-e543-4129-a6ea-695c314bad9d-kube-api-access-dflws\") on node \"crc\" DevicePath \"\"" Nov 25 11:18:27 crc kubenswrapper[4821]: I1125 11:18:27.477832 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2mzwd" event={"ID":"8dcb623e-e543-4129-a6ea-695c314bad9d","Type":"ContainerDied","Data":"1218517c6ab7061ed69d5b0ba0ac353e041563c5972496659c4d5c7984b8c6e7"} Nov 25 11:18:27 crc kubenswrapper[4821]: I1125 11:18:27.477892 4821 scope.go:117] "RemoveContainer" containerID="281a63a60823b3cc2329d6235521646e367fe5d71785d3886299fd410bc11171" Nov 25 11:18:27 crc kubenswrapper[4821]: I1125 11:18:27.477924 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2mzwd" Nov 25 11:18:27 crc kubenswrapper[4821]: I1125 11:18:27.507692 4821 scope.go:117] "RemoveContainer" containerID="252aefe9a538cea2e00a9a9d44d771e5951e7789b8fe16e82b5d4d405661c49d" Nov 25 11:18:27 crc kubenswrapper[4821]: I1125 11:18:27.515453 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2mzwd"] Nov 25 11:18:27 crc kubenswrapper[4821]: I1125 11:18:27.523835 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-2mzwd"] Nov 25 11:18:27 crc kubenswrapper[4821]: I1125 11:18:27.547632 4821 scope.go:117] "RemoveContainer" containerID="2813b4502c262487863adf34670dcb5be394f14dcc0b0ec905905b0a485775f7" Nov 25 11:18:28 crc kubenswrapper[4821]: I1125 11:18:28.125715 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8dcb623e-e543-4129-a6ea-695c314bad9d" path="/var/lib/kubelet/pods/8dcb623e-e543-4129-a6ea-695c314bad9d/volumes" Nov 25 11:18:28 crc kubenswrapper[4821]: I1125 11:18:28.494518 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"d8aea755-bf2f-4443-8196-46936ce3e2fd","Type":"ContainerStarted","Data":"2e8913f0f951c82e4bd46360606b86de76140c25a93d213dd39ebaae984f2756"} Nov 25 11:18:28 crc kubenswrapper[4821]: I1125 11:18:28.518374 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=4.347419189 podStartE2EDuration="52.518355331s" podCreationTimestamp="2025-11-25 11:17:36 +0000 UTC" firstStartedPulling="2025-11-25 11:17:38.299761919 +0000 UTC m=+2728.836081766" lastFinishedPulling="2025-11-25 11:18:26.470698061 +0000 UTC m=+2777.007017908" observedRunningTime="2025-11-25 11:18:28.510195732 +0000 UTC m=+2779.046515579" watchObservedRunningTime="2025-11-25 11:18:28.518355331 +0000 UTC m=+2779.054675188" Nov 25 11:18:29 crc kubenswrapper[4821]: I1125 11:18:29.238253 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-ncn9h" Nov 25 11:18:29 crc kubenswrapper[4821]: I1125 11:18:29.287909 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ncn9h"] Nov 25 11:18:29 crc kubenswrapper[4821]: I1125 11:18:29.502887 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-ncn9h" podUID="f5e191cd-eb4d-41d7-b634-a7ee707fd960" containerName="registry-server" containerID="cri-o://4a5c8744a3b8c13ab40ae81676b89a7be3d6bfd90939cb924cdde2bb4bae1697" gracePeriod=2 Nov 25 11:18:29 crc kubenswrapper[4821]: I1125 11:18:29.951895 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ncn9h" Nov 25 11:18:30 crc kubenswrapper[4821]: I1125 11:18:30.015529 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h88t9\" (UniqueName: \"kubernetes.io/projected/f5e191cd-eb4d-41d7-b634-a7ee707fd960-kube-api-access-h88t9\") pod \"f5e191cd-eb4d-41d7-b634-a7ee707fd960\" (UID: \"f5e191cd-eb4d-41d7-b634-a7ee707fd960\") " Nov 25 11:18:30 crc kubenswrapper[4821]: I1125 11:18:30.015619 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5e191cd-eb4d-41d7-b634-a7ee707fd960-catalog-content\") pod \"f5e191cd-eb4d-41d7-b634-a7ee707fd960\" (UID: \"f5e191cd-eb4d-41d7-b634-a7ee707fd960\") " Nov 25 11:18:30 crc kubenswrapper[4821]: I1125 11:18:30.015791 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5e191cd-eb4d-41d7-b634-a7ee707fd960-utilities\") pod \"f5e191cd-eb4d-41d7-b634-a7ee707fd960\" (UID: \"f5e191cd-eb4d-41d7-b634-a7ee707fd960\") " Nov 25 11:18:30 crc kubenswrapper[4821]: I1125 11:18:30.017134 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5e191cd-eb4d-41d7-b634-a7ee707fd960-utilities" (OuterVolumeSpecName: "utilities") pod "f5e191cd-eb4d-41d7-b634-a7ee707fd960" (UID: "f5e191cd-eb4d-41d7-b634-a7ee707fd960"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:18:30 crc kubenswrapper[4821]: I1125 11:18:30.020714 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5e191cd-eb4d-41d7-b634-a7ee707fd960-kube-api-access-h88t9" (OuterVolumeSpecName: "kube-api-access-h88t9") pod "f5e191cd-eb4d-41d7-b634-a7ee707fd960" (UID: "f5e191cd-eb4d-41d7-b634-a7ee707fd960"). InnerVolumeSpecName "kube-api-access-h88t9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:18:30 crc kubenswrapper[4821]: I1125 11:18:30.077277 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5e191cd-eb4d-41d7-b634-a7ee707fd960-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f5e191cd-eb4d-41d7-b634-a7ee707fd960" (UID: "f5e191cd-eb4d-41d7-b634-a7ee707fd960"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:18:30 crc kubenswrapper[4821]: I1125 11:18:30.118849 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h88t9\" (UniqueName: \"kubernetes.io/projected/f5e191cd-eb4d-41d7-b634-a7ee707fd960-kube-api-access-h88t9\") on node \"crc\" DevicePath \"\"" Nov 25 11:18:30 crc kubenswrapper[4821]: I1125 11:18:30.118886 4821 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5e191cd-eb4d-41d7-b634-a7ee707fd960-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 11:18:30 crc kubenswrapper[4821]: I1125 11:18:30.118895 4821 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5e191cd-eb4d-41d7-b634-a7ee707fd960-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 11:18:30 crc kubenswrapper[4821]: I1125 11:18:30.515012 4821 generic.go:334] "Generic (PLEG): container finished" podID="f5e191cd-eb4d-41d7-b634-a7ee707fd960" containerID="4a5c8744a3b8c13ab40ae81676b89a7be3d6bfd90939cb924cdde2bb4bae1697" exitCode=0 Nov 25 11:18:30 crc kubenswrapper[4821]: I1125 11:18:30.515086 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ncn9h" event={"ID":"f5e191cd-eb4d-41d7-b634-a7ee707fd960","Type":"ContainerDied","Data":"4a5c8744a3b8c13ab40ae81676b89a7be3d6bfd90939cb924cdde2bb4bae1697"} Nov 25 11:18:30 crc kubenswrapper[4821]: I1125 11:18:30.515176 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ncn9h" event={"ID":"f5e191cd-eb4d-41d7-b634-a7ee707fd960","Type":"ContainerDied","Data":"a2971de0f8cbe82afb051c553655a6515b886a75d2da276cf7a23fbf403ca1ab"} Nov 25 11:18:30 crc kubenswrapper[4821]: I1125 11:18:30.515200 4821 scope.go:117] "RemoveContainer" containerID="4a5c8744a3b8c13ab40ae81676b89a7be3d6bfd90939cb924cdde2bb4bae1697" Nov 25 11:18:30 crc kubenswrapper[4821]: I1125 11:18:30.515116 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ncn9h" Nov 25 11:18:30 crc kubenswrapper[4821]: I1125 11:18:30.541484 4821 scope.go:117] "RemoveContainer" containerID="0a1680a5af79c8f6fed01fa20b577f56dc9d3aa91c7e45a682e818f7426a2caf" Nov 25 11:18:30 crc kubenswrapper[4821]: I1125 11:18:30.544735 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ncn9h"] Nov 25 11:18:30 crc kubenswrapper[4821]: I1125 11:18:30.553985 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-ncn9h"] Nov 25 11:18:30 crc kubenswrapper[4821]: I1125 11:18:30.567626 4821 scope.go:117] "RemoveContainer" containerID="81e99aee9a5302cd1905ad34f90c330f4133a7c0232759f6b5f74de9957e7abe" Nov 25 11:18:30 crc kubenswrapper[4821]: I1125 11:18:30.613001 4821 scope.go:117] "RemoveContainer" containerID="4a5c8744a3b8c13ab40ae81676b89a7be3d6bfd90939cb924cdde2bb4bae1697" Nov 25 11:18:30 crc kubenswrapper[4821]: E1125 11:18:30.613808 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a5c8744a3b8c13ab40ae81676b89a7be3d6bfd90939cb924cdde2bb4bae1697\": container with ID starting with 4a5c8744a3b8c13ab40ae81676b89a7be3d6bfd90939cb924cdde2bb4bae1697 not found: ID does not exist" containerID="4a5c8744a3b8c13ab40ae81676b89a7be3d6bfd90939cb924cdde2bb4bae1697" Nov 25 11:18:30 crc kubenswrapper[4821]: I1125 11:18:30.613869 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a5c8744a3b8c13ab40ae81676b89a7be3d6bfd90939cb924cdde2bb4bae1697"} err="failed to get container status \"4a5c8744a3b8c13ab40ae81676b89a7be3d6bfd90939cb924cdde2bb4bae1697\": rpc error: code = NotFound desc = could not find container \"4a5c8744a3b8c13ab40ae81676b89a7be3d6bfd90939cb924cdde2bb4bae1697\": container with ID starting with 4a5c8744a3b8c13ab40ae81676b89a7be3d6bfd90939cb924cdde2bb4bae1697 not found: ID does not exist" Nov 25 11:18:30 crc kubenswrapper[4821]: I1125 11:18:30.613896 4821 scope.go:117] "RemoveContainer" containerID="0a1680a5af79c8f6fed01fa20b577f56dc9d3aa91c7e45a682e818f7426a2caf" Nov 25 11:18:30 crc kubenswrapper[4821]: E1125 11:18:30.614170 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a1680a5af79c8f6fed01fa20b577f56dc9d3aa91c7e45a682e818f7426a2caf\": container with ID starting with 0a1680a5af79c8f6fed01fa20b577f56dc9d3aa91c7e45a682e818f7426a2caf not found: ID does not exist" containerID="0a1680a5af79c8f6fed01fa20b577f56dc9d3aa91c7e45a682e818f7426a2caf" Nov 25 11:18:30 crc kubenswrapper[4821]: I1125 11:18:30.614216 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a1680a5af79c8f6fed01fa20b577f56dc9d3aa91c7e45a682e818f7426a2caf"} err="failed to get container status \"0a1680a5af79c8f6fed01fa20b577f56dc9d3aa91c7e45a682e818f7426a2caf\": rpc error: code = NotFound desc = could not find container \"0a1680a5af79c8f6fed01fa20b577f56dc9d3aa91c7e45a682e818f7426a2caf\": container with ID starting with 0a1680a5af79c8f6fed01fa20b577f56dc9d3aa91c7e45a682e818f7426a2caf not found: ID does not exist" Nov 25 11:18:30 crc kubenswrapper[4821]: I1125 11:18:30.614234 4821 scope.go:117] "RemoveContainer" containerID="81e99aee9a5302cd1905ad34f90c330f4133a7c0232759f6b5f74de9957e7abe" Nov 25 11:18:30 crc kubenswrapper[4821]: E1125 11:18:30.614421 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"81e99aee9a5302cd1905ad34f90c330f4133a7c0232759f6b5f74de9957e7abe\": container with ID starting with 81e99aee9a5302cd1905ad34f90c330f4133a7c0232759f6b5f74de9957e7abe not found: ID does not exist" containerID="81e99aee9a5302cd1905ad34f90c330f4133a7c0232759f6b5f74de9957e7abe" Nov 25 11:18:30 crc kubenswrapper[4821]: I1125 11:18:30.614455 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81e99aee9a5302cd1905ad34f90c330f4133a7c0232759f6b5f74de9957e7abe"} err="failed to get container status \"81e99aee9a5302cd1905ad34f90c330f4133a7c0232759f6b5f74de9957e7abe\": rpc error: code = NotFound desc = could not find container \"81e99aee9a5302cd1905ad34f90c330f4133a7c0232759f6b5f74de9957e7abe\": container with ID starting with 81e99aee9a5302cd1905ad34f90c330f4133a7c0232759f6b5f74de9957e7abe not found: ID does not exist" Nov 25 11:18:32 crc kubenswrapper[4821]: I1125 11:18:32.139348 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5e191cd-eb4d-41d7-b634-a7ee707fd960" path="/var/lib/kubelet/pods/f5e191cd-eb4d-41d7-b634-a7ee707fd960/volumes" Nov 25 11:18:34 crc kubenswrapper[4821]: I1125 11:18:34.714748 4821 patch_prober.go:28] interesting pod/machine-config-daemon-2krbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:18:34 crc kubenswrapper[4821]: I1125 11:18:34.715367 4821 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:19:04 crc kubenswrapper[4821]: I1125 11:19:04.715295 4821 patch_prober.go:28] interesting pod/machine-config-daemon-2krbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:19:04 crc kubenswrapper[4821]: I1125 11:19:04.715966 4821 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:19:13 crc kubenswrapper[4821]: I1125 11:19:13.031662 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-htdmk"] Nov 25 11:19:13 crc kubenswrapper[4821]: E1125 11:19:13.033340 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dcb623e-e543-4129-a6ea-695c314bad9d" containerName="extract-utilities" Nov 25 11:19:13 crc kubenswrapper[4821]: I1125 11:19:13.034119 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dcb623e-e543-4129-a6ea-695c314bad9d" containerName="extract-utilities" Nov 25 11:19:13 crc kubenswrapper[4821]: E1125 11:19:13.034179 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dcb623e-e543-4129-a6ea-695c314bad9d" containerName="registry-server" Nov 25 11:19:13 crc kubenswrapper[4821]: I1125 11:19:13.034191 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dcb623e-e543-4129-a6ea-695c314bad9d" containerName="registry-server" Nov 25 11:19:13 crc kubenswrapper[4821]: E1125 11:19:13.034214 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5e191cd-eb4d-41d7-b634-a7ee707fd960" containerName="extract-utilities" Nov 25 11:19:13 crc kubenswrapper[4821]: I1125 11:19:13.034223 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5e191cd-eb4d-41d7-b634-a7ee707fd960" containerName="extract-utilities" Nov 25 11:19:13 crc kubenswrapper[4821]: E1125 11:19:13.034248 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dcb623e-e543-4129-a6ea-695c314bad9d" containerName="extract-content" Nov 25 11:19:13 crc kubenswrapper[4821]: I1125 11:19:13.034258 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dcb623e-e543-4129-a6ea-695c314bad9d" containerName="extract-content" Nov 25 11:19:13 crc kubenswrapper[4821]: E1125 11:19:13.034278 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5e191cd-eb4d-41d7-b634-a7ee707fd960" containerName="registry-server" Nov 25 11:19:13 crc kubenswrapper[4821]: I1125 11:19:13.034286 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5e191cd-eb4d-41d7-b634-a7ee707fd960" containerName="registry-server" Nov 25 11:19:13 crc kubenswrapper[4821]: E1125 11:19:13.034312 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5e191cd-eb4d-41d7-b634-a7ee707fd960" containerName="extract-content" Nov 25 11:19:13 crc kubenswrapper[4821]: I1125 11:19:13.034320 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5e191cd-eb4d-41d7-b634-a7ee707fd960" containerName="extract-content" Nov 25 11:19:13 crc kubenswrapper[4821]: I1125 11:19:13.035070 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5e191cd-eb4d-41d7-b634-a7ee707fd960" containerName="registry-server" Nov 25 11:19:13 crc kubenswrapper[4821]: I1125 11:19:13.035111 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="8dcb623e-e543-4129-a6ea-695c314bad9d" containerName="registry-server" Nov 25 11:19:13 crc kubenswrapper[4821]: I1125 11:19:13.049471 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-htdmk" Nov 25 11:19:13 crc kubenswrapper[4821]: I1125 11:19:13.068574 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-htdmk"] Nov 25 11:19:13 crc kubenswrapper[4821]: I1125 11:19:13.099606 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jwx5k\" (UniqueName: \"kubernetes.io/projected/7650a517-b31a-4b5e-acd5-01694c9e00b9-kube-api-access-jwx5k\") pod \"certified-operators-htdmk\" (UID: \"7650a517-b31a-4b5e-acd5-01694c9e00b9\") " pod="openshift-marketplace/certified-operators-htdmk" Nov 25 11:19:13 crc kubenswrapper[4821]: I1125 11:19:13.099681 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7650a517-b31a-4b5e-acd5-01694c9e00b9-catalog-content\") pod \"certified-operators-htdmk\" (UID: \"7650a517-b31a-4b5e-acd5-01694c9e00b9\") " pod="openshift-marketplace/certified-operators-htdmk" Nov 25 11:19:13 crc kubenswrapper[4821]: I1125 11:19:13.099796 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7650a517-b31a-4b5e-acd5-01694c9e00b9-utilities\") pod \"certified-operators-htdmk\" (UID: \"7650a517-b31a-4b5e-acd5-01694c9e00b9\") " pod="openshift-marketplace/certified-operators-htdmk" Nov 25 11:19:13 crc kubenswrapper[4821]: I1125 11:19:13.201777 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jwx5k\" (UniqueName: \"kubernetes.io/projected/7650a517-b31a-4b5e-acd5-01694c9e00b9-kube-api-access-jwx5k\") pod \"certified-operators-htdmk\" (UID: \"7650a517-b31a-4b5e-acd5-01694c9e00b9\") " pod="openshift-marketplace/certified-operators-htdmk" Nov 25 11:19:13 crc kubenswrapper[4821]: I1125 11:19:13.202191 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7650a517-b31a-4b5e-acd5-01694c9e00b9-catalog-content\") pod \"certified-operators-htdmk\" (UID: \"7650a517-b31a-4b5e-acd5-01694c9e00b9\") " pod="openshift-marketplace/certified-operators-htdmk" Nov 25 11:19:13 crc kubenswrapper[4821]: I1125 11:19:13.202331 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7650a517-b31a-4b5e-acd5-01694c9e00b9-utilities\") pod \"certified-operators-htdmk\" (UID: \"7650a517-b31a-4b5e-acd5-01694c9e00b9\") " pod="openshift-marketplace/certified-operators-htdmk" Nov 25 11:19:13 crc kubenswrapper[4821]: I1125 11:19:13.202903 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7650a517-b31a-4b5e-acd5-01694c9e00b9-catalog-content\") pod \"certified-operators-htdmk\" (UID: \"7650a517-b31a-4b5e-acd5-01694c9e00b9\") " pod="openshift-marketplace/certified-operators-htdmk" Nov 25 11:19:13 crc kubenswrapper[4821]: I1125 11:19:13.203194 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7650a517-b31a-4b5e-acd5-01694c9e00b9-utilities\") pod \"certified-operators-htdmk\" (UID: \"7650a517-b31a-4b5e-acd5-01694c9e00b9\") " pod="openshift-marketplace/certified-operators-htdmk" Nov 25 11:19:13 crc kubenswrapper[4821]: I1125 11:19:13.232109 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jwx5k\" (UniqueName: \"kubernetes.io/projected/7650a517-b31a-4b5e-acd5-01694c9e00b9-kube-api-access-jwx5k\") pod \"certified-operators-htdmk\" (UID: \"7650a517-b31a-4b5e-acd5-01694c9e00b9\") " pod="openshift-marketplace/certified-operators-htdmk" Nov 25 11:19:13 crc kubenswrapper[4821]: I1125 11:19:13.375207 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-htdmk" Nov 25 11:19:13 crc kubenswrapper[4821]: I1125 11:19:13.909333 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-htdmk"] Nov 25 11:19:13 crc kubenswrapper[4821]: W1125 11:19:13.913348 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7650a517_b31a_4b5e_acd5_01694c9e00b9.slice/crio-923d7d90e450b8c1be6c38ab1ec9dd02852a76dbf62ba0f1864d81a41b7a7561 WatchSource:0}: Error finding container 923d7d90e450b8c1be6c38ab1ec9dd02852a76dbf62ba0f1864d81a41b7a7561: Status 404 returned error can't find the container with id 923d7d90e450b8c1be6c38ab1ec9dd02852a76dbf62ba0f1864d81a41b7a7561 Nov 25 11:19:13 crc kubenswrapper[4821]: I1125 11:19:13.968205 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-htdmk" event={"ID":"7650a517-b31a-4b5e-acd5-01694c9e00b9","Type":"ContainerStarted","Data":"923d7d90e450b8c1be6c38ab1ec9dd02852a76dbf62ba0f1864d81a41b7a7561"} Nov 25 11:19:14 crc kubenswrapper[4821]: I1125 11:19:14.979571 4821 generic.go:334] "Generic (PLEG): container finished" podID="7650a517-b31a-4b5e-acd5-01694c9e00b9" containerID="e3847e40103f7a0d6475177ef9832dfef36e1ca85e4060559af9a8dd7d4329e2" exitCode=0 Nov 25 11:19:14 crc kubenswrapper[4821]: I1125 11:19:14.979833 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-htdmk" event={"ID":"7650a517-b31a-4b5e-acd5-01694c9e00b9","Type":"ContainerDied","Data":"e3847e40103f7a0d6475177ef9832dfef36e1ca85e4060559af9a8dd7d4329e2"} Nov 25 11:19:15 crc kubenswrapper[4821]: I1125 11:19:15.991005 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-htdmk" event={"ID":"7650a517-b31a-4b5e-acd5-01694c9e00b9","Type":"ContainerStarted","Data":"b3865bf837a4c4748798ff95bd40961f70e25b6aa6547621bc6514504cb27823"} Nov 25 11:19:17 crc kubenswrapper[4821]: I1125 11:19:17.003244 4821 generic.go:334] "Generic (PLEG): container finished" podID="7650a517-b31a-4b5e-acd5-01694c9e00b9" containerID="b3865bf837a4c4748798ff95bd40961f70e25b6aa6547621bc6514504cb27823" exitCode=0 Nov 25 11:19:17 crc kubenswrapper[4821]: I1125 11:19:17.003301 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-htdmk" event={"ID":"7650a517-b31a-4b5e-acd5-01694c9e00b9","Type":"ContainerDied","Data":"b3865bf837a4c4748798ff95bd40961f70e25b6aa6547621bc6514504cb27823"} Nov 25 11:19:17 crc kubenswrapper[4821]: I1125 11:19:17.005760 4821 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 11:19:18 crc kubenswrapper[4821]: I1125 11:19:18.015144 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-htdmk" event={"ID":"7650a517-b31a-4b5e-acd5-01694c9e00b9","Type":"ContainerStarted","Data":"0679d9549e1ddfb7a85c68566ca54117423cf3591e0382ab5220323f087f329d"} Nov 25 11:19:18 crc kubenswrapper[4821]: I1125 11:19:18.035926 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-htdmk" podStartSLOduration=3.579208659 podStartE2EDuration="6.035906338s" podCreationTimestamp="2025-11-25 11:19:12 +0000 UTC" firstStartedPulling="2025-11-25 11:19:14.98388606 +0000 UTC m=+2825.520205927" lastFinishedPulling="2025-11-25 11:19:17.440583759 +0000 UTC m=+2827.976903606" observedRunningTime="2025-11-25 11:19:18.033062889 +0000 UTC m=+2828.569382766" watchObservedRunningTime="2025-11-25 11:19:18.035906338 +0000 UTC m=+2828.572226185" Nov 25 11:19:23 crc kubenswrapper[4821]: I1125 11:19:23.376179 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-htdmk" Nov 25 11:19:23 crc kubenswrapper[4821]: I1125 11:19:23.376772 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-htdmk" Nov 25 11:19:23 crc kubenswrapper[4821]: I1125 11:19:23.420536 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-htdmk" Nov 25 11:19:24 crc kubenswrapper[4821]: I1125 11:19:24.124004 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-htdmk" Nov 25 11:19:24 crc kubenswrapper[4821]: I1125 11:19:24.173902 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-htdmk"] Nov 25 11:19:26 crc kubenswrapper[4821]: I1125 11:19:26.086760 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-htdmk" podUID="7650a517-b31a-4b5e-acd5-01694c9e00b9" containerName="registry-server" containerID="cri-o://0679d9549e1ddfb7a85c68566ca54117423cf3591e0382ab5220323f087f329d" gracePeriod=2 Nov 25 11:19:26 crc kubenswrapper[4821]: I1125 11:19:26.575937 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-htdmk" Nov 25 11:19:26 crc kubenswrapper[4821]: I1125 11:19:26.766720 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jwx5k\" (UniqueName: \"kubernetes.io/projected/7650a517-b31a-4b5e-acd5-01694c9e00b9-kube-api-access-jwx5k\") pod \"7650a517-b31a-4b5e-acd5-01694c9e00b9\" (UID: \"7650a517-b31a-4b5e-acd5-01694c9e00b9\") " Nov 25 11:19:26 crc kubenswrapper[4821]: I1125 11:19:26.767379 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7650a517-b31a-4b5e-acd5-01694c9e00b9-utilities\") pod \"7650a517-b31a-4b5e-acd5-01694c9e00b9\" (UID: \"7650a517-b31a-4b5e-acd5-01694c9e00b9\") " Nov 25 11:19:26 crc kubenswrapper[4821]: I1125 11:19:26.767407 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7650a517-b31a-4b5e-acd5-01694c9e00b9-catalog-content\") pod \"7650a517-b31a-4b5e-acd5-01694c9e00b9\" (UID: \"7650a517-b31a-4b5e-acd5-01694c9e00b9\") " Nov 25 11:19:26 crc kubenswrapper[4821]: I1125 11:19:26.768124 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7650a517-b31a-4b5e-acd5-01694c9e00b9-utilities" (OuterVolumeSpecName: "utilities") pod "7650a517-b31a-4b5e-acd5-01694c9e00b9" (UID: "7650a517-b31a-4b5e-acd5-01694c9e00b9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:19:26 crc kubenswrapper[4821]: I1125 11:19:26.768535 4821 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7650a517-b31a-4b5e-acd5-01694c9e00b9-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 11:19:26 crc kubenswrapper[4821]: I1125 11:19:26.775369 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7650a517-b31a-4b5e-acd5-01694c9e00b9-kube-api-access-jwx5k" (OuterVolumeSpecName: "kube-api-access-jwx5k") pod "7650a517-b31a-4b5e-acd5-01694c9e00b9" (UID: "7650a517-b31a-4b5e-acd5-01694c9e00b9"). InnerVolumeSpecName "kube-api-access-jwx5k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:19:26 crc kubenswrapper[4821]: I1125 11:19:26.817472 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7650a517-b31a-4b5e-acd5-01694c9e00b9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7650a517-b31a-4b5e-acd5-01694c9e00b9" (UID: "7650a517-b31a-4b5e-acd5-01694c9e00b9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:19:26 crc kubenswrapper[4821]: I1125 11:19:26.871050 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jwx5k\" (UniqueName: \"kubernetes.io/projected/7650a517-b31a-4b5e-acd5-01694c9e00b9-kube-api-access-jwx5k\") on node \"crc\" DevicePath \"\"" Nov 25 11:19:26 crc kubenswrapper[4821]: I1125 11:19:26.872225 4821 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7650a517-b31a-4b5e-acd5-01694c9e00b9-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 11:19:27 crc kubenswrapper[4821]: I1125 11:19:27.100215 4821 generic.go:334] "Generic (PLEG): container finished" podID="7650a517-b31a-4b5e-acd5-01694c9e00b9" containerID="0679d9549e1ddfb7a85c68566ca54117423cf3591e0382ab5220323f087f329d" exitCode=0 Nov 25 11:19:27 crc kubenswrapper[4821]: I1125 11:19:27.100306 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-htdmk" Nov 25 11:19:27 crc kubenswrapper[4821]: I1125 11:19:27.100350 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-htdmk" event={"ID":"7650a517-b31a-4b5e-acd5-01694c9e00b9","Type":"ContainerDied","Data":"0679d9549e1ddfb7a85c68566ca54117423cf3591e0382ab5220323f087f329d"} Nov 25 11:19:27 crc kubenswrapper[4821]: I1125 11:19:27.101557 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-htdmk" event={"ID":"7650a517-b31a-4b5e-acd5-01694c9e00b9","Type":"ContainerDied","Data":"923d7d90e450b8c1be6c38ab1ec9dd02852a76dbf62ba0f1864d81a41b7a7561"} Nov 25 11:19:27 crc kubenswrapper[4821]: I1125 11:19:27.101595 4821 scope.go:117] "RemoveContainer" containerID="0679d9549e1ddfb7a85c68566ca54117423cf3591e0382ab5220323f087f329d" Nov 25 11:19:27 crc kubenswrapper[4821]: I1125 11:19:27.130883 4821 scope.go:117] "RemoveContainer" containerID="b3865bf837a4c4748798ff95bd40961f70e25b6aa6547621bc6514504cb27823" Nov 25 11:19:27 crc kubenswrapper[4821]: I1125 11:19:27.153200 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-htdmk"] Nov 25 11:19:27 crc kubenswrapper[4821]: I1125 11:19:27.160618 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-htdmk"] Nov 25 11:19:27 crc kubenswrapper[4821]: I1125 11:19:27.184150 4821 scope.go:117] "RemoveContainer" containerID="e3847e40103f7a0d6475177ef9832dfef36e1ca85e4060559af9a8dd7d4329e2" Nov 25 11:19:27 crc kubenswrapper[4821]: I1125 11:19:27.212878 4821 scope.go:117] "RemoveContainer" containerID="0679d9549e1ddfb7a85c68566ca54117423cf3591e0382ab5220323f087f329d" Nov 25 11:19:27 crc kubenswrapper[4821]: E1125 11:19:27.213507 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0679d9549e1ddfb7a85c68566ca54117423cf3591e0382ab5220323f087f329d\": container with ID starting with 0679d9549e1ddfb7a85c68566ca54117423cf3591e0382ab5220323f087f329d not found: ID does not exist" containerID="0679d9549e1ddfb7a85c68566ca54117423cf3591e0382ab5220323f087f329d" Nov 25 11:19:27 crc kubenswrapper[4821]: I1125 11:19:27.213557 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0679d9549e1ddfb7a85c68566ca54117423cf3591e0382ab5220323f087f329d"} err="failed to get container status \"0679d9549e1ddfb7a85c68566ca54117423cf3591e0382ab5220323f087f329d\": rpc error: code = NotFound desc = could not find container \"0679d9549e1ddfb7a85c68566ca54117423cf3591e0382ab5220323f087f329d\": container with ID starting with 0679d9549e1ddfb7a85c68566ca54117423cf3591e0382ab5220323f087f329d not found: ID does not exist" Nov 25 11:19:27 crc kubenswrapper[4821]: I1125 11:19:27.213588 4821 scope.go:117] "RemoveContainer" containerID="b3865bf837a4c4748798ff95bd40961f70e25b6aa6547621bc6514504cb27823" Nov 25 11:19:27 crc kubenswrapper[4821]: E1125 11:19:27.214133 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b3865bf837a4c4748798ff95bd40961f70e25b6aa6547621bc6514504cb27823\": container with ID starting with b3865bf837a4c4748798ff95bd40961f70e25b6aa6547621bc6514504cb27823 not found: ID does not exist" containerID="b3865bf837a4c4748798ff95bd40961f70e25b6aa6547621bc6514504cb27823" Nov 25 11:19:27 crc kubenswrapper[4821]: I1125 11:19:27.214191 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3865bf837a4c4748798ff95bd40961f70e25b6aa6547621bc6514504cb27823"} err="failed to get container status \"b3865bf837a4c4748798ff95bd40961f70e25b6aa6547621bc6514504cb27823\": rpc error: code = NotFound desc = could not find container \"b3865bf837a4c4748798ff95bd40961f70e25b6aa6547621bc6514504cb27823\": container with ID starting with b3865bf837a4c4748798ff95bd40961f70e25b6aa6547621bc6514504cb27823 not found: ID does not exist" Nov 25 11:19:27 crc kubenswrapper[4821]: I1125 11:19:27.214211 4821 scope.go:117] "RemoveContainer" containerID="e3847e40103f7a0d6475177ef9832dfef36e1ca85e4060559af9a8dd7d4329e2" Nov 25 11:19:27 crc kubenswrapper[4821]: E1125 11:19:27.214486 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3847e40103f7a0d6475177ef9832dfef36e1ca85e4060559af9a8dd7d4329e2\": container with ID starting with e3847e40103f7a0d6475177ef9832dfef36e1ca85e4060559af9a8dd7d4329e2 not found: ID does not exist" containerID="e3847e40103f7a0d6475177ef9832dfef36e1ca85e4060559af9a8dd7d4329e2" Nov 25 11:19:27 crc kubenswrapper[4821]: I1125 11:19:27.214528 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3847e40103f7a0d6475177ef9832dfef36e1ca85e4060559af9a8dd7d4329e2"} err="failed to get container status \"e3847e40103f7a0d6475177ef9832dfef36e1ca85e4060559af9a8dd7d4329e2\": rpc error: code = NotFound desc = could not find container \"e3847e40103f7a0d6475177ef9832dfef36e1ca85e4060559af9a8dd7d4329e2\": container with ID starting with e3847e40103f7a0d6475177ef9832dfef36e1ca85e4060559af9a8dd7d4329e2 not found: ID does not exist" Nov 25 11:19:28 crc kubenswrapper[4821]: I1125 11:19:28.124211 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7650a517-b31a-4b5e-acd5-01694c9e00b9" path="/var/lib/kubelet/pods/7650a517-b31a-4b5e-acd5-01694c9e00b9/volumes" Nov 25 11:19:34 crc kubenswrapper[4821]: I1125 11:19:34.714942 4821 patch_prober.go:28] interesting pod/machine-config-daemon-2krbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:19:34 crc kubenswrapper[4821]: I1125 11:19:34.715526 4821 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:19:34 crc kubenswrapper[4821]: I1125 11:19:34.715577 4821 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" Nov 25 11:19:34 crc kubenswrapper[4821]: I1125 11:19:34.716377 4821 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d5e77cf7ba2ff0993f9dc84aaf718cec52c804bab193ec24676471e97efd1f7b"} pod="openshift-machine-config-operator/machine-config-daemon-2krbf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 11:19:34 crc kubenswrapper[4821]: I1125 11:19:34.716433 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerName="machine-config-daemon" containerID="cri-o://d5e77cf7ba2ff0993f9dc84aaf718cec52c804bab193ec24676471e97efd1f7b" gracePeriod=600 Nov 25 11:19:35 crc kubenswrapper[4821]: I1125 11:19:35.195532 4821 generic.go:334] "Generic (PLEG): container finished" podID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerID="d5e77cf7ba2ff0993f9dc84aaf718cec52c804bab193ec24676471e97efd1f7b" exitCode=0 Nov 25 11:19:35 crc kubenswrapper[4821]: I1125 11:19:35.195599 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" event={"ID":"5f948b87-ac86-4de6-ad64-c2ef947f84d4","Type":"ContainerDied","Data":"d5e77cf7ba2ff0993f9dc84aaf718cec52c804bab193ec24676471e97efd1f7b"} Nov 25 11:19:35 crc kubenswrapper[4821]: I1125 11:19:35.195863 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" event={"ID":"5f948b87-ac86-4de6-ad64-c2ef947f84d4","Type":"ContainerStarted","Data":"df8d79f5b2e040490d30845f855d9fa73fed8e867a647684ba5fc0077d412c79"} Nov 25 11:19:35 crc kubenswrapper[4821]: I1125 11:19:35.195886 4821 scope.go:117] "RemoveContainer" containerID="61010995905bb7c40fff6417cf7e0b11d584b8a9a68f6fccf71aefe231f45291" Nov 25 11:22:04 crc kubenswrapper[4821]: I1125 11:22:04.714761 4821 patch_prober.go:28] interesting pod/machine-config-daemon-2krbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:22:04 crc kubenswrapper[4821]: I1125 11:22:04.716337 4821 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:22:34 crc kubenswrapper[4821]: I1125 11:22:34.715094 4821 patch_prober.go:28] interesting pod/machine-config-daemon-2krbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:22:34 crc kubenswrapper[4821]: I1125 11:22:34.715775 4821 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:23:04 crc kubenswrapper[4821]: I1125 11:23:04.714195 4821 patch_prober.go:28] interesting pod/machine-config-daemon-2krbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:23:04 crc kubenswrapper[4821]: I1125 11:23:04.714747 4821 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:23:04 crc kubenswrapper[4821]: I1125 11:23:04.714800 4821 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" Nov 25 11:23:04 crc kubenswrapper[4821]: I1125 11:23:04.715801 4821 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"df8d79f5b2e040490d30845f855d9fa73fed8e867a647684ba5fc0077d412c79"} pod="openshift-machine-config-operator/machine-config-daemon-2krbf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 11:23:04 crc kubenswrapper[4821]: I1125 11:23:04.715857 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerName="machine-config-daemon" containerID="cri-o://df8d79f5b2e040490d30845f855d9fa73fed8e867a647684ba5fc0077d412c79" gracePeriod=600 Nov 25 11:23:04 crc kubenswrapper[4821]: E1125 11:23:04.865202 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:23:05 crc kubenswrapper[4821]: I1125 11:23:05.032794 4821 generic.go:334] "Generic (PLEG): container finished" podID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerID="df8d79f5b2e040490d30845f855d9fa73fed8e867a647684ba5fc0077d412c79" exitCode=0 Nov 25 11:23:05 crc kubenswrapper[4821]: I1125 11:23:05.032832 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" event={"ID":"5f948b87-ac86-4de6-ad64-c2ef947f84d4","Type":"ContainerDied","Data":"df8d79f5b2e040490d30845f855d9fa73fed8e867a647684ba5fc0077d412c79"} Nov 25 11:23:05 crc kubenswrapper[4821]: I1125 11:23:05.032863 4821 scope.go:117] "RemoveContainer" containerID="d5e77cf7ba2ff0993f9dc84aaf718cec52c804bab193ec24676471e97efd1f7b" Nov 25 11:23:05 crc kubenswrapper[4821]: I1125 11:23:05.033426 4821 scope.go:117] "RemoveContainer" containerID="df8d79f5b2e040490d30845f855d9fa73fed8e867a647684ba5fc0077d412c79" Nov 25 11:23:05 crc kubenswrapper[4821]: E1125 11:23:05.033677 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:23:19 crc kubenswrapper[4821]: I1125 11:23:19.114200 4821 scope.go:117] "RemoveContainer" containerID="df8d79f5b2e040490d30845f855d9fa73fed8e867a647684ba5fc0077d412c79" Nov 25 11:23:19 crc kubenswrapper[4821]: E1125 11:23:19.115185 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:23:30 crc kubenswrapper[4821]: I1125 11:23:30.124260 4821 scope.go:117] "RemoveContainer" containerID="df8d79f5b2e040490d30845f855d9fa73fed8e867a647684ba5fc0077d412c79" Nov 25 11:23:30 crc kubenswrapper[4821]: E1125 11:23:30.124911 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:23:44 crc kubenswrapper[4821]: I1125 11:23:44.114759 4821 scope.go:117] "RemoveContainer" containerID="df8d79f5b2e040490d30845f855d9fa73fed8e867a647684ba5fc0077d412c79" Nov 25 11:23:44 crc kubenswrapper[4821]: E1125 11:23:44.115582 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:23:56 crc kubenswrapper[4821]: I1125 11:23:56.114691 4821 scope.go:117] "RemoveContainer" containerID="df8d79f5b2e040490d30845f855d9fa73fed8e867a647684ba5fc0077d412c79" Nov 25 11:23:56 crc kubenswrapper[4821]: E1125 11:23:56.115557 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:24:10 crc kubenswrapper[4821]: I1125 11:24:10.114188 4821 scope.go:117] "RemoveContainer" containerID="df8d79f5b2e040490d30845f855d9fa73fed8e867a647684ba5fc0077d412c79" Nov 25 11:24:10 crc kubenswrapper[4821]: E1125 11:24:10.115011 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:24:25 crc kubenswrapper[4821]: I1125 11:24:25.114732 4821 scope.go:117] "RemoveContainer" containerID="df8d79f5b2e040490d30845f855d9fa73fed8e867a647684ba5fc0077d412c79" Nov 25 11:24:25 crc kubenswrapper[4821]: E1125 11:24:25.115501 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:24:37 crc kubenswrapper[4821]: I1125 11:24:37.116053 4821 scope.go:117] "RemoveContainer" containerID="df8d79f5b2e040490d30845f855d9fa73fed8e867a647684ba5fc0077d412c79" Nov 25 11:24:37 crc kubenswrapper[4821]: E1125 11:24:37.116863 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:24:50 crc kubenswrapper[4821]: I1125 11:24:50.124102 4821 scope.go:117] "RemoveContainer" containerID="df8d79f5b2e040490d30845f855d9fa73fed8e867a647684ba5fc0077d412c79" Nov 25 11:24:50 crc kubenswrapper[4821]: E1125 11:24:50.126485 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:25:02 crc kubenswrapper[4821]: I1125 11:25:02.114739 4821 scope.go:117] "RemoveContainer" containerID="df8d79f5b2e040490d30845f855d9fa73fed8e867a647684ba5fc0077d412c79" Nov 25 11:25:02 crc kubenswrapper[4821]: E1125 11:25:02.115971 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:25:17 crc kubenswrapper[4821]: I1125 11:25:17.114375 4821 scope.go:117] "RemoveContainer" containerID="df8d79f5b2e040490d30845f855d9fa73fed8e867a647684ba5fc0077d412c79" Nov 25 11:25:17 crc kubenswrapper[4821]: E1125 11:25:17.115092 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:25:30 crc kubenswrapper[4821]: I1125 11:25:30.115223 4821 scope.go:117] "RemoveContainer" containerID="df8d79f5b2e040490d30845f855d9fa73fed8e867a647684ba5fc0077d412c79" Nov 25 11:25:30 crc kubenswrapper[4821]: E1125 11:25:30.116411 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:25:44 crc kubenswrapper[4821]: I1125 11:25:44.115184 4821 scope.go:117] "RemoveContainer" containerID="df8d79f5b2e040490d30845f855d9fa73fed8e867a647684ba5fc0077d412c79" Nov 25 11:25:44 crc kubenswrapper[4821]: E1125 11:25:44.115912 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:25:56 crc kubenswrapper[4821]: I1125 11:25:56.115190 4821 scope.go:117] "RemoveContainer" containerID="df8d79f5b2e040490d30845f855d9fa73fed8e867a647684ba5fc0077d412c79" Nov 25 11:25:56 crc kubenswrapper[4821]: E1125 11:25:56.116104 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:26:11 crc kubenswrapper[4821]: I1125 11:26:11.114078 4821 scope.go:117] "RemoveContainer" containerID="df8d79f5b2e040490d30845f855d9fa73fed8e867a647684ba5fc0077d412c79" Nov 25 11:26:11 crc kubenswrapper[4821]: E1125 11:26:11.114841 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:26:23 crc kubenswrapper[4821]: I1125 11:26:23.114061 4821 scope.go:117] "RemoveContainer" containerID="df8d79f5b2e040490d30845f855d9fa73fed8e867a647684ba5fc0077d412c79" Nov 25 11:26:23 crc kubenswrapper[4821]: E1125 11:26:23.114879 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:26:37 crc kubenswrapper[4821]: I1125 11:26:37.115184 4821 scope.go:117] "RemoveContainer" containerID="df8d79f5b2e040490d30845f855d9fa73fed8e867a647684ba5fc0077d412c79" Nov 25 11:26:37 crc kubenswrapper[4821]: E1125 11:26:37.115934 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:26:51 crc kubenswrapper[4821]: I1125 11:26:51.114277 4821 scope.go:117] "RemoveContainer" containerID="df8d79f5b2e040490d30845f855d9fa73fed8e867a647684ba5fc0077d412c79" Nov 25 11:26:51 crc kubenswrapper[4821]: E1125 11:26:51.116265 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:26:56 crc kubenswrapper[4821]: I1125 11:26:56.693032 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-kpp84"] Nov 25 11:26:56 crc kubenswrapper[4821]: E1125 11:26:56.694968 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7650a517-b31a-4b5e-acd5-01694c9e00b9" containerName="extract-content" Nov 25 11:26:56 crc kubenswrapper[4821]: I1125 11:26:56.694991 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="7650a517-b31a-4b5e-acd5-01694c9e00b9" containerName="extract-content" Nov 25 11:26:56 crc kubenswrapper[4821]: E1125 11:26:56.695010 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7650a517-b31a-4b5e-acd5-01694c9e00b9" containerName="extract-utilities" Nov 25 11:26:56 crc kubenswrapper[4821]: I1125 11:26:56.695024 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="7650a517-b31a-4b5e-acd5-01694c9e00b9" containerName="extract-utilities" Nov 25 11:26:56 crc kubenswrapper[4821]: E1125 11:26:56.695072 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7650a517-b31a-4b5e-acd5-01694c9e00b9" containerName="registry-server" Nov 25 11:26:56 crc kubenswrapper[4821]: I1125 11:26:56.695083 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="7650a517-b31a-4b5e-acd5-01694c9e00b9" containerName="registry-server" Nov 25 11:26:56 crc kubenswrapper[4821]: I1125 11:26:56.695376 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="7650a517-b31a-4b5e-acd5-01694c9e00b9" containerName="registry-server" Nov 25 11:26:56 crc kubenswrapper[4821]: I1125 11:26:56.697892 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kpp84" Nov 25 11:26:56 crc kubenswrapper[4821]: I1125 11:26:56.704664 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kpp84"] Nov 25 11:26:56 crc kubenswrapper[4821]: I1125 11:26:56.758987 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xsnlh\" (UniqueName: \"kubernetes.io/projected/33a5a62a-56ae-4412-a1f2-c5ed19a9430b-kube-api-access-xsnlh\") pod \"redhat-marketplace-kpp84\" (UID: \"33a5a62a-56ae-4412-a1f2-c5ed19a9430b\") " pod="openshift-marketplace/redhat-marketplace-kpp84" Nov 25 11:26:56 crc kubenswrapper[4821]: I1125 11:26:56.759072 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33a5a62a-56ae-4412-a1f2-c5ed19a9430b-utilities\") pod \"redhat-marketplace-kpp84\" (UID: \"33a5a62a-56ae-4412-a1f2-c5ed19a9430b\") " pod="openshift-marketplace/redhat-marketplace-kpp84" Nov 25 11:26:56 crc kubenswrapper[4821]: I1125 11:26:56.759180 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33a5a62a-56ae-4412-a1f2-c5ed19a9430b-catalog-content\") pod \"redhat-marketplace-kpp84\" (UID: \"33a5a62a-56ae-4412-a1f2-c5ed19a9430b\") " pod="openshift-marketplace/redhat-marketplace-kpp84" Nov 25 11:26:56 crc kubenswrapper[4821]: I1125 11:26:56.861573 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xsnlh\" (UniqueName: \"kubernetes.io/projected/33a5a62a-56ae-4412-a1f2-c5ed19a9430b-kube-api-access-xsnlh\") pod \"redhat-marketplace-kpp84\" (UID: \"33a5a62a-56ae-4412-a1f2-c5ed19a9430b\") " pod="openshift-marketplace/redhat-marketplace-kpp84" Nov 25 11:26:56 crc kubenswrapper[4821]: I1125 11:26:56.861679 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33a5a62a-56ae-4412-a1f2-c5ed19a9430b-utilities\") pod \"redhat-marketplace-kpp84\" (UID: \"33a5a62a-56ae-4412-a1f2-c5ed19a9430b\") " pod="openshift-marketplace/redhat-marketplace-kpp84" Nov 25 11:26:56 crc kubenswrapper[4821]: I1125 11:26:56.861739 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33a5a62a-56ae-4412-a1f2-c5ed19a9430b-catalog-content\") pod \"redhat-marketplace-kpp84\" (UID: \"33a5a62a-56ae-4412-a1f2-c5ed19a9430b\") " pod="openshift-marketplace/redhat-marketplace-kpp84" Nov 25 11:26:56 crc kubenswrapper[4821]: I1125 11:26:56.862336 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33a5a62a-56ae-4412-a1f2-c5ed19a9430b-catalog-content\") pod \"redhat-marketplace-kpp84\" (UID: \"33a5a62a-56ae-4412-a1f2-c5ed19a9430b\") " pod="openshift-marketplace/redhat-marketplace-kpp84" Nov 25 11:26:56 crc kubenswrapper[4821]: I1125 11:26:56.862595 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33a5a62a-56ae-4412-a1f2-c5ed19a9430b-utilities\") pod \"redhat-marketplace-kpp84\" (UID: \"33a5a62a-56ae-4412-a1f2-c5ed19a9430b\") " pod="openshift-marketplace/redhat-marketplace-kpp84" Nov 25 11:26:56 crc kubenswrapper[4821]: I1125 11:26:56.886316 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xsnlh\" (UniqueName: \"kubernetes.io/projected/33a5a62a-56ae-4412-a1f2-c5ed19a9430b-kube-api-access-xsnlh\") pod \"redhat-marketplace-kpp84\" (UID: \"33a5a62a-56ae-4412-a1f2-c5ed19a9430b\") " pod="openshift-marketplace/redhat-marketplace-kpp84" Nov 25 11:26:57 crc kubenswrapper[4821]: I1125 11:26:57.030018 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kpp84" Nov 25 11:26:57 crc kubenswrapper[4821]: I1125 11:26:57.496583 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kpp84"] Nov 25 11:26:58 crc kubenswrapper[4821]: I1125 11:26:58.074391 4821 generic.go:334] "Generic (PLEG): container finished" podID="33a5a62a-56ae-4412-a1f2-c5ed19a9430b" containerID="c8d6b8860c581859170b48cc93175506a7dcdac0968770e40c4bdfb8b2237c73" exitCode=0 Nov 25 11:26:58 crc kubenswrapper[4821]: I1125 11:26:58.074448 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kpp84" event={"ID":"33a5a62a-56ae-4412-a1f2-c5ed19a9430b","Type":"ContainerDied","Data":"c8d6b8860c581859170b48cc93175506a7dcdac0968770e40c4bdfb8b2237c73"} Nov 25 11:26:58 crc kubenswrapper[4821]: I1125 11:26:58.074652 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kpp84" event={"ID":"33a5a62a-56ae-4412-a1f2-c5ed19a9430b","Type":"ContainerStarted","Data":"f7416ea9c7b7e029b130d30c63ccbee530638fb4f479a3e1a634d662fc00b9cc"} Nov 25 11:26:58 crc kubenswrapper[4821]: I1125 11:26:58.076572 4821 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 11:26:59 crc kubenswrapper[4821]: I1125 11:26:59.088579 4821 generic.go:334] "Generic (PLEG): container finished" podID="33a5a62a-56ae-4412-a1f2-c5ed19a9430b" containerID="d503c0ac81ba020cbeca5504bd67b6ca1b04f2c9af1131239d2b867fd091240d" exitCode=0 Nov 25 11:26:59 crc kubenswrapper[4821]: I1125 11:26:59.088981 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kpp84" event={"ID":"33a5a62a-56ae-4412-a1f2-c5ed19a9430b","Type":"ContainerDied","Data":"d503c0ac81ba020cbeca5504bd67b6ca1b04f2c9af1131239d2b867fd091240d"} Nov 25 11:27:00 crc kubenswrapper[4821]: I1125 11:27:00.103137 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kpp84" event={"ID":"33a5a62a-56ae-4412-a1f2-c5ed19a9430b","Type":"ContainerStarted","Data":"8ff09f600dff958fd8194300ccbb09415916b88e7fc41a5508aab690e7006414"} Nov 25 11:27:00 crc kubenswrapper[4821]: I1125 11:27:00.133900 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-kpp84" podStartSLOduration=2.536620333 podStartE2EDuration="4.133879844s" podCreationTimestamp="2025-11-25 11:26:56 +0000 UTC" firstStartedPulling="2025-11-25 11:26:58.076323118 +0000 UTC m=+3288.612642965" lastFinishedPulling="2025-11-25 11:26:59.673582629 +0000 UTC m=+3290.209902476" observedRunningTime="2025-11-25 11:27:00.123938729 +0000 UTC m=+3290.660258586" watchObservedRunningTime="2025-11-25 11:27:00.133879844 +0000 UTC m=+3290.670199691" Nov 25 11:27:06 crc kubenswrapper[4821]: I1125 11:27:06.115007 4821 scope.go:117] "RemoveContainer" containerID="df8d79f5b2e040490d30845f855d9fa73fed8e867a647684ba5fc0077d412c79" Nov 25 11:27:06 crc kubenswrapper[4821]: E1125 11:27:06.115957 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:27:07 crc kubenswrapper[4821]: I1125 11:27:07.030515 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-kpp84" Nov 25 11:27:07 crc kubenswrapper[4821]: I1125 11:27:07.030836 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-kpp84" Nov 25 11:27:07 crc kubenswrapper[4821]: I1125 11:27:07.077605 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-kpp84" Nov 25 11:27:07 crc kubenswrapper[4821]: I1125 11:27:07.223839 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-kpp84" Nov 25 11:27:07 crc kubenswrapper[4821]: I1125 11:27:07.318027 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kpp84"] Nov 25 11:27:09 crc kubenswrapper[4821]: I1125 11:27:09.197915 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-kpp84" podUID="33a5a62a-56ae-4412-a1f2-c5ed19a9430b" containerName="registry-server" containerID="cri-o://8ff09f600dff958fd8194300ccbb09415916b88e7fc41a5508aab690e7006414" gracePeriod=2 Nov 25 11:27:09 crc kubenswrapper[4821]: I1125 11:27:09.704233 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kpp84" Nov 25 11:27:09 crc kubenswrapper[4821]: I1125 11:27:09.836460 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33a5a62a-56ae-4412-a1f2-c5ed19a9430b-catalog-content\") pod \"33a5a62a-56ae-4412-a1f2-c5ed19a9430b\" (UID: \"33a5a62a-56ae-4412-a1f2-c5ed19a9430b\") " Nov 25 11:27:09 crc kubenswrapper[4821]: I1125 11:27:09.836854 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xsnlh\" (UniqueName: \"kubernetes.io/projected/33a5a62a-56ae-4412-a1f2-c5ed19a9430b-kube-api-access-xsnlh\") pod \"33a5a62a-56ae-4412-a1f2-c5ed19a9430b\" (UID: \"33a5a62a-56ae-4412-a1f2-c5ed19a9430b\") " Nov 25 11:27:09 crc kubenswrapper[4821]: I1125 11:27:09.837001 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33a5a62a-56ae-4412-a1f2-c5ed19a9430b-utilities\") pod \"33a5a62a-56ae-4412-a1f2-c5ed19a9430b\" (UID: \"33a5a62a-56ae-4412-a1f2-c5ed19a9430b\") " Nov 25 11:27:09 crc kubenswrapper[4821]: I1125 11:27:09.838040 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33a5a62a-56ae-4412-a1f2-c5ed19a9430b-utilities" (OuterVolumeSpecName: "utilities") pod "33a5a62a-56ae-4412-a1f2-c5ed19a9430b" (UID: "33a5a62a-56ae-4412-a1f2-c5ed19a9430b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:27:09 crc kubenswrapper[4821]: I1125 11:27:09.843202 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33a5a62a-56ae-4412-a1f2-c5ed19a9430b-kube-api-access-xsnlh" (OuterVolumeSpecName: "kube-api-access-xsnlh") pod "33a5a62a-56ae-4412-a1f2-c5ed19a9430b" (UID: "33a5a62a-56ae-4412-a1f2-c5ed19a9430b"). InnerVolumeSpecName "kube-api-access-xsnlh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:27:09 crc kubenswrapper[4821]: I1125 11:27:09.886403 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33a5a62a-56ae-4412-a1f2-c5ed19a9430b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "33a5a62a-56ae-4412-a1f2-c5ed19a9430b" (UID: "33a5a62a-56ae-4412-a1f2-c5ed19a9430b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:27:09 crc kubenswrapper[4821]: I1125 11:27:09.939717 4821 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33a5a62a-56ae-4412-a1f2-c5ed19a9430b-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 11:27:09 crc kubenswrapper[4821]: I1125 11:27:09.939757 4821 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33a5a62a-56ae-4412-a1f2-c5ed19a9430b-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 11:27:09 crc kubenswrapper[4821]: I1125 11:27:09.939772 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xsnlh\" (UniqueName: \"kubernetes.io/projected/33a5a62a-56ae-4412-a1f2-c5ed19a9430b-kube-api-access-xsnlh\") on node \"crc\" DevicePath \"\"" Nov 25 11:27:10 crc kubenswrapper[4821]: I1125 11:27:10.210898 4821 generic.go:334] "Generic (PLEG): container finished" podID="33a5a62a-56ae-4412-a1f2-c5ed19a9430b" containerID="8ff09f600dff958fd8194300ccbb09415916b88e7fc41a5508aab690e7006414" exitCode=0 Nov 25 11:27:10 crc kubenswrapper[4821]: I1125 11:27:10.210940 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kpp84" event={"ID":"33a5a62a-56ae-4412-a1f2-c5ed19a9430b","Type":"ContainerDied","Data":"8ff09f600dff958fd8194300ccbb09415916b88e7fc41a5508aab690e7006414"} Nov 25 11:27:10 crc kubenswrapper[4821]: I1125 11:27:10.210968 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kpp84" event={"ID":"33a5a62a-56ae-4412-a1f2-c5ed19a9430b","Type":"ContainerDied","Data":"f7416ea9c7b7e029b130d30c63ccbee530638fb4f479a3e1a634d662fc00b9cc"} Nov 25 11:27:10 crc kubenswrapper[4821]: I1125 11:27:10.210985 4821 scope.go:117] "RemoveContainer" containerID="8ff09f600dff958fd8194300ccbb09415916b88e7fc41a5508aab690e7006414" Nov 25 11:27:10 crc kubenswrapper[4821]: I1125 11:27:10.210978 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kpp84" Nov 25 11:27:10 crc kubenswrapper[4821]: I1125 11:27:10.244774 4821 scope.go:117] "RemoveContainer" containerID="d503c0ac81ba020cbeca5504bd67b6ca1b04f2c9af1131239d2b867fd091240d" Nov 25 11:27:10 crc kubenswrapper[4821]: I1125 11:27:10.247306 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kpp84"] Nov 25 11:27:10 crc kubenswrapper[4821]: I1125 11:27:10.262321 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-kpp84"] Nov 25 11:27:10 crc kubenswrapper[4821]: I1125 11:27:10.265758 4821 scope.go:117] "RemoveContainer" containerID="c8d6b8860c581859170b48cc93175506a7dcdac0968770e40c4bdfb8b2237c73" Nov 25 11:27:10 crc kubenswrapper[4821]: I1125 11:27:10.309245 4821 scope.go:117] "RemoveContainer" containerID="8ff09f600dff958fd8194300ccbb09415916b88e7fc41a5508aab690e7006414" Nov 25 11:27:10 crc kubenswrapper[4821]: E1125 11:27:10.309653 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8ff09f600dff958fd8194300ccbb09415916b88e7fc41a5508aab690e7006414\": container with ID starting with 8ff09f600dff958fd8194300ccbb09415916b88e7fc41a5508aab690e7006414 not found: ID does not exist" containerID="8ff09f600dff958fd8194300ccbb09415916b88e7fc41a5508aab690e7006414" Nov 25 11:27:10 crc kubenswrapper[4821]: I1125 11:27:10.309698 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ff09f600dff958fd8194300ccbb09415916b88e7fc41a5508aab690e7006414"} err="failed to get container status \"8ff09f600dff958fd8194300ccbb09415916b88e7fc41a5508aab690e7006414\": rpc error: code = NotFound desc = could not find container \"8ff09f600dff958fd8194300ccbb09415916b88e7fc41a5508aab690e7006414\": container with ID starting with 8ff09f600dff958fd8194300ccbb09415916b88e7fc41a5508aab690e7006414 not found: ID does not exist" Nov 25 11:27:10 crc kubenswrapper[4821]: I1125 11:27:10.309732 4821 scope.go:117] "RemoveContainer" containerID="d503c0ac81ba020cbeca5504bd67b6ca1b04f2c9af1131239d2b867fd091240d" Nov 25 11:27:10 crc kubenswrapper[4821]: E1125 11:27:10.310020 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d503c0ac81ba020cbeca5504bd67b6ca1b04f2c9af1131239d2b867fd091240d\": container with ID starting with d503c0ac81ba020cbeca5504bd67b6ca1b04f2c9af1131239d2b867fd091240d not found: ID does not exist" containerID="d503c0ac81ba020cbeca5504bd67b6ca1b04f2c9af1131239d2b867fd091240d" Nov 25 11:27:10 crc kubenswrapper[4821]: I1125 11:27:10.310055 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d503c0ac81ba020cbeca5504bd67b6ca1b04f2c9af1131239d2b867fd091240d"} err="failed to get container status \"d503c0ac81ba020cbeca5504bd67b6ca1b04f2c9af1131239d2b867fd091240d\": rpc error: code = NotFound desc = could not find container \"d503c0ac81ba020cbeca5504bd67b6ca1b04f2c9af1131239d2b867fd091240d\": container with ID starting with d503c0ac81ba020cbeca5504bd67b6ca1b04f2c9af1131239d2b867fd091240d not found: ID does not exist" Nov 25 11:27:10 crc kubenswrapper[4821]: I1125 11:27:10.310078 4821 scope.go:117] "RemoveContainer" containerID="c8d6b8860c581859170b48cc93175506a7dcdac0968770e40c4bdfb8b2237c73" Nov 25 11:27:10 crc kubenswrapper[4821]: E1125 11:27:10.310726 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8d6b8860c581859170b48cc93175506a7dcdac0968770e40c4bdfb8b2237c73\": container with ID starting with c8d6b8860c581859170b48cc93175506a7dcdac0968770e40c4bdfb8b2237c73 not found: ID does not exist" containerID="c8d6b8860c581859170b48cc93175506a7dcdac0968770e40c4bdfb8b2237c73" Nov 25 11:27:10 crc kubenswrapper[4821]: I1125 11:27:10.310854 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8d6b8860c581859170b48cc93175506a7dcdac0968770e40c4bdfb8b2237c73"} err="failed to get container status \"c8d6b8860c581859170b48cc93175506a7dcdac0968770e40c4bdfb8b2237c73\": rpc error: code = NotFound desc = could not find container \"c8d6b8860c581859170b48cc93175506a7dcdac0968770e40c4bdfb8b2237c73\": container with ID starting with c8d6b8860c581859170b48cc93175506a7dcdac0968770e40c4bdfb8b2237c73 not found: ID does not exist" Nov 25 11:27:10 crc kubenswrapper[4821]: E1125 11:27:10.955857 4821 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod33a5a62a_56ae_4412_a1f2_c5ed19a9430b.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod33a5a62a_56ae_4412_a1f2_c5ed19a9430b.slice/crio-f7416ea9c7b7e029b130d30c63ccbee530638fb4f479a3e1a634d662fc00b9cc\": RecentStats: unable to find data in memory cache]" Nov 25 11:27:12 crc kubenswrapper[4821]: I1125 11:27:12.126658 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33a5a62a-56ae-4412-a1f2-c5ed19a9430b" path="/var/lib/kubelet/pods/33a5a62a-56ae-4412-a1f2-c5ed19a9430b/volumes" Nov 25 11:27:18 crc kubenswrapper[4821]: I1125 11:27:18.115397 4821 scope.go:117] "RemoveContainer" containerID="df8d79f5b2e040490d30845f855d9fa73fed8e867a647684ba5fc0077d412c79" Nov 25 11:27:18 crc kubenswrapper[4821]: E1125 11:27:18.116370 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:27:21 crc kubenswrapper[4821]: E1125 11:27:21.207390 4821 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod33a5a62a_56ae_4412_a1f2_c5ed19a9430b.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod33a5a62a_56ae_4412_a1f2_c5ed19a9430b.slice/crio-f7416ea9c7b7e029b130d30c63ccbee530638fb4f479a3e1a634d662fc00b9cc\": RecentStats: unable to find data in memory cache]" Nov 25 11:27:30 crc kubenswrapper[4821]: I1125 11:27:30.122436 4821 scope.go:117] "RemoveContainer" containerID="df8d79f5b2e040490d30845f855d9fa73fed8e867a647684ba5fc0077d412c79" Nov 25 11:27:30 crc kubenswrapper[4821]: E1125 11:27:30.124765 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:27:31 crc kubenswrapper[4821]: E1125 11:27:31.445909 4821 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod33a5a62a_56ae_4412_a1f2_c5ed19a9430b.slice/crio-f7416ea9c7b7e029b130d30c63ccbee530638fb4f479a3e1a634d662fc00b9cc\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod33a5a62a_56ae_4412_a1f2_c5ed19a9430b.slice\": RecentStats: unable to find data in memory cache]" Nov 25 11:27:41 crc kubenswrapper[4821]: E1125 11:27:41.676748 4821 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod33a5a62a_56ae_4412_a1f2_c5ed19a9430b.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod33a5a62a_56ae_4412_a1f2_c5ed19a9430b.slice/crio-f7416ea9c7b7e029b130d30c63ccbee530638fb4f479a3e1a634d662fc00b9cc\": RecentStats: unable to find data in memory cache]" Nov 25 11:27:42 crc kubenswrapper[4821]: I1125 11:27:42.115249 4821 scope.go:117] "RemoveContainer" containerID="df8d79f5b2e040490d30845f855d9fa73fed8e867a647684ba5fc0077d412c79" Nov 25 11:27:42 crc kubenswrapper[4821]: E1125 11:27:42.116043 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:27:51 crc kubenswrapper[4821]: E1125 11:27:51.945342 4821 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod33a5a62a_56ae_4412_a1f2_c5ed19a9430b.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod33a5a62a_56ae_4412_a1f2_c5ed19a9430b.slice/crio-f7416ea9c7b7e029b130d30c63ccbee530638fb4f479a3e1a634d662fc00b9cc\": RecentStats: unable to find data in memory cache]" Nov 25 11:27:54 crc kubenswrapper[4821]: I1125 11:27:54.114566 4821 scope.go:117] "RemoveContainer" containerID="df8d79f5b2e040490d30845f855d9fa73fed8e867a647684ba5fc0077d412c79" Nov 25 11:27:54 crc kubenswrapper[4821]: E1125 11:27:54.116495 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:27:58 crc kubenswrapper[4821]: I1125 11:27:58.883828 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-t2h46"] Nov 25 11:27:58 crc kubenswrapper[4821]: E1125 11:27:58.884993 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33a5a62a-56ae-4412-a1f2-c5ed19a9430b" containerName="extract-content" Nov 25 11:27:58 crc kubenswrapper[4821]: I1125 11:27:58.885010 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="33a5a62a-56ae-4412-a1f2-c5ed19a9430b" containerName="extract-content" Nov 25 11:27:58 crc kubenswrapper[4821]: E1125 11:27:58.885048 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33a5a62a-56ae-4412-a1f2-c5ed19a9430b" containerName="extract-utilities" Nov 25 11:27:58 crc kubenswrapper[4821]: I1125 11:27:58.885056 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="33a5a62a-56ae-4412-a1f2-c5ed19a9430b" containerName="extract-utilities" Nov 25 11:27:58 crc kubenswrapper[4821]: E1125 11:27:58.885071 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33a5a62a-56ae-4412-a1f2-c5ed19a9430b" containerName="registry-server" Nov 25 11:27:58 crc kubenswrapper[4821]: I1125 11:27:58.885077 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="33a5a62a-56ae-4412-a1f2-c5ed19a9430b" containerName="registry-server" Nov 25 11:27:58 crc kubenswrapper[4821]: I1125 11:27:58.885267 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="33a5a62a-56ae-4412-a1f2-c5ed19a9430b" containerName="registry-server" Nov 25 11:27:58 crc kubenswrapper[4821]: I1125 11:27:58.886815 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t2h46" Nov 25 11:27:58 crc kubenswrapper[4821]: I1125 11:27:58.892917 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-t2h46"] Nov 25 11:27:58 crc kubenswrapper[4821]: I1125 11:27:58.958254 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43f709d6-33c8-42bd-8dc4-c92a781c1e78-utilities\") pod \"redhat-operators-t2h46\" (UID: \"43f709d6-33c8-42bd-8dc4-c92a781c1e78\") " pod="openshift-marketplace/redhat-operators-t2h46" Nov 25 11:27:58 crc kubenswrapper[4821]: I1125 11:27:58.958605 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43f709d6-33c8-42bd-8dc4-c92a781c1e78-catalog-content\") pod \"redhat-operators-t2h46\" (UID: \"43f709d6-33c8-42bd-8dc4-c92a781c1e78\") " pod="openshift-marketplace/redhat-operators-t2h46" Nov 25 11:27:58 crc kubenswrapper[4821]: I1125 11:27:58.958632 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f6djt\" (UniqueName: \"kubernetes.io/projected/43f709d6-33c8-42bd-8dc4-c92a781c1e78-kube-api-access-f6djt\") pod \"redhat-operators-t2h46\" (UID: \"43f709d6-33c8-42bd-8dc4-c92a781c1e78\") " pod="openshift-marketplace/redhat-operators-t2h46" Nov 25 11:27:59 crc kubenswrapper[4821]: I1125 11:27:59.060386 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43f709d6-33c8-42bd-8dc4-c92a781c1e78-utilities\") pod \"redhat-operators-t2h46\" (UID: \"43f709d6-33c8-42bd-8dc4-c92a781c1e78\") " pod="openshift-marketplace/redhat-operators-t2h46" Nov 25 11:27:59 crc kubenswrapper[4821]: I1125 11:27:59.060483 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43f709d6-33c8-42bd-8dc4-c92a781c1e78-catalog-content\") pod \"redhat-operators-t2h46\" (UID: \"43f709d6-33c8-42bd-8dc4-c92a781c1e78\") " pod="openshift-marketplace/redhat-operators-t2h46" Nov 25 11:27:59 crc kubenswrapper[4821]: I1125 11:27:59.060519 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f6djt\" (UniqueName: \"kubernetes.io/projected/43f709d6-33c8-42bd-8dc4-c92a781c1e78-kube-api-access-f6djt\") pod \"redhat-operators-t2h46\" (UID: \"43f709d6-33c8-42bd-8dc4-c92a781c1e78\") " pod="openshift-marketplace/redhat-operators-t2h46" Nov 25 11:27:59 crc kubenswrapper[4821]: I1125 11:27:59.061034 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43f709d6-33c8-42bd-8dc4-c92a781c1e78-utilities\") pod \"redhat-operators-t2h46\" (UID: \"43f709d6-33c8-42bd-8dc4-c92a781c1e78\") " pod="openshift-marketplace/redhat-operators-t2h46" Nov 25 11:27:59 crc kubenswrapper[4821]: I1125 11:27:59.061087 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43f709d6-33c8-42bd-8dc4-c92a781c1e78-catalog-content\") pod \"redhat-operators-t2h46\" (UID: \"43f709d6-33c8-42bd-8dc4-c92a781c1e78\") " pod="openshift-marketplace/redhat-operators-t2h46" Nov 25 11:27:59 crc kubenswrapper[4821]: I1125 11:27:59.084841 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f6djt\" (UniqueName: \"kubernetes.io/projected/43f709d6-33c8-42bd-8dc4-c92a781c1e78-kube-api-access-f6djt\") pod \"redhat-operators-t2h46\" (UID: \"43f709d6-33c8-42bd-8dc4-c92a781c1e78\") " pod="openshift-marketplace/redhat-operators-t2h46" Nov 25 11:27:59 crc kubenswrapper[4821]: I1125 11:27:59.216521 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t2h46" Nov 25 11:27:59 crc kubenswrapper[4821]: I1125 11:27:59.685446 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-t2h46"] Nov 25 11:27:59 crc kubenswrapper[4821]: W1125 11:27:59.693901 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod43f709d6_33c8_42bd_8dc4_c92a781c1e78.slice/crio-708cce6ffc8a490d12fe3fde38091c1e26b2e738ab5ecc1eb76b4a8c6b33bd52 WatchSource:0}: Error finding container 708cce6ffc8a490d12fe3fde38091c1e26b2e738ab5ecc1eb76b4a8c6b33bd52: Status 404 returned error can't find the container with id 708cce6ffc8a490d12fe3fde38091c1e26b2e738ab5ecc1eb76b4a8c6b33bd52 Nov 25 11:28:00 crc kubenswrapper[4821]: I1125 11:28:00.644203 4821 generic.go:334] "Generic (PLEG): container finished" podID="43f709d6-33c8-42bd-8dc4-c92a781c1e78" containerID="ef57d227c9f79aefaf9c26b270a02c65861bb79ebe0f2835dc1f28b26c07a76e" exitCode=0 Nov 25 11:28:00 crc kubenswrapper[4821]: I1125 11:28:00.644302 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t2h46" event={"ID":"43f709d6-33c8-42bd-8dc4-c92a781c1e78","Type":"ContainerDied","Data":"ef57d227c9f79aefaf9c26b270a02c65861bb79ebe0f2835dc1f28b26c07a76e"} Nov 25 11:28:00 crc kubenswrapper[4821]: I1125 11:28:00.645527 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t2h46" event={"ID":"43f709d6-33c8-42bd-8dc4-c92a781c1e78","Type":"ContainerStarted","Data":"708cce6ffc8a490d12fe3fde38091c1e26b2e738ab5ecc1eb76b4a8c6b33bd52"} Nov 25 11:28:01 crc kubenswrapper[4821]: I1125 11:28:01.655238 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t2h46" event={"ID":"43f709d6-33c8-42bd-8dc4-c92a781c1e78","Type":"ContainerStarted","Data":"de163f58a985eae43706092ae42d42d28bba0555977a3307e163c65fcda1f996"} Nov 25 11:28:02 crc kubenswrapper[4821]: E1125 11:28:02.183543 4821 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod33a5a62a_56ae_4412_a1f2_c5ed19a9430b.slice/crio-f7416ea9c7b7e029b130d30c63ccbee530638fb4f479a3e1a634d662fc00b9cc\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod33a5a62a_56ae_4412_a1f2_c5ed19a9430b.slice\": RecentStats: unable to find data in memory cache]" Nov 25 11:28:02 crc kubenswrapper[4821]: I1125 11:28:02.666646 4821 generic.go:334] "Generic (PLEG): container finished" podID="43f709d6-33c8-42bd-8dc4-c92a781c1e78" containerID="de163f58a985eae43706092ae42d42d28bba0555977a3307e163c65fcda1f996" exitCode=0 Nov 25 11:28:02 crc kubenswrapper[4821]: I1125 11:28:02.666685 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t2h46" event={"ID":"43f709d6-33c8-42bd-8dc4-c92a781c1e78","Type":"ContainerDied","Data":"de163f58a985eae43706092ae42d42d28bba0555977a3307e163c65fcda1f996"} Nov 25 11:28:04 crc kubenswrapper[4821]: I1125 11:28:04.684247 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t2h46" event={"ID":"43f709d6-33c8-42bd-8dc4-c92a781c1e78","Type":"ContainerStarted","Data":"ee3efd40ae0589435ec2d7c71811e1afff1864c39066562f589269cdfee2b524"} Nov 25 11:28:04 crc kubenswrapper[4821]: I1125 11:28:04.713045 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-t2h46" podStartSLOduration=3.9317170580000003 podStartE2EDuration="6.713025973s" podCreationTimestamp="2025-11-25 11:27:58 +0000 UTC" firstStartedPulling="2025-11-25 11:28:00.646564407 +0000 UTC m=+3351.182884254" lastFinishedPulling="2025-11-25 11:28:03.427873322 +0000 UTC m=+3353.964193169" observedRunningTime="2025-11-25 11:28:04.708490083 +0000 UTC m=+3355.244809930" watchObservedRunningTime="2025-11-25 11:28:04.713025973 +0000 UTC m=+3355.249345840" Nov 25 11:28:09 crc kubenswrapper[4821]: I1125 11:28:09.115352 4821 scope.go:117] "RemoveContainer" containerID="df8d79f5b2e040490d30845f855d9fa73fed8e867a647684ba5fc0077d412c79" Nov 25 11:28:09 crc kubenswrapper[4821]: I1125 11:28:09.217050 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-t2h46" Nov 25 11:28:09 crc kubenswrapper[4821]: I1125 11:28:09.217129 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-t2h46" Nov 25 11:28:09 crc kubenswrapper[4821]: I1125 11:28:09.279945 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-t2h46" Nov 25 11:28:09 crc kubenswrapper[4821]: I1125 11:28:09.724510 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" event={"ID":"5f948b87-ac86-4de6-ad64-c2ef947f84d4","Type":"ContainerStarted","Data":"52d1a350addd03a9e1e92a65fe40ee197504bf6db7f5b0601ca8860f562470fb"} Nov 25 11:28:09 crc kubenswrapper[4821]: I1125 11:28:09.819030 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-t2h46" Nov 25 11:28:09 crc kubenswrapper[4821]: I1125 11:28:09.871337 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-t2h46"] Nov 25 11:28:10 crc kubenswrapper[4821]: E1125 11:28:10.165387 4821 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/ef44383c71bd886babff73e8c53f825be3b7c15103a89868cde56de91c05909d/diff" to get inode usage: stat /var/lib/containers/storage/overlay/ef44383c71bd886babff73e8c53f825be3b7c15103a89868cde56de91c05909d/diff: no such file or directory, extraDiskErr: Nov 25 11:28:11 crc kubenswrapper[4821]: I1125 11:28:11.741020 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-t2h46" podUID="43f709d6-33c8-42bd-8dc4-c92a781c1e78" containerName="registry-server" containerID="cri-o://ee3efd40ae0589435ec2d7c71811e1afff1864c39066562f589269cdfee2b524" gracePeriod=2 Nov 25 11:28:12 crc kubenswrapper[4821]: I1125 11:28:12.306741 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t2h46" Nov 25 11:28:12 crc kubenswrapper[4821]: I1125 11:28:12.422303 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43f709d6-33c8-42bd-8dc4-c92a781c1e78-utilities\") pod \"43f709d6-33c8-42bd-8dc4-c92a781c1e78\" (UID: \"43f709d6-33c8-42bd-8dc4-c92a781c1e78\") " Nov 25 11:28:12 crc kubenswrapper[4821]: I1125 11:28:12.422370 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f6djt\" (UniqueName: \"kubernetes.io/projected/43f709d6-33c8-42bd-8dc4-c92a781c1e78-kube-api-access-f6djt\") pod \"43f709d6-33c8-42bd-8dc4-c92a781c1e78\" (UID: \"43f709d6-33c8-42bd-8dc4-c92a781c1e78\") " Nov 25 11:28:12 crc kubenswrapper[4821]: I1125 11:28:12.422409 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43f709d6-33c8-42bd-8dc4-c92a781c1e78-catalog-content\") pod \"43f709d6-33c8-42bd-8dc4-c92a781c1e78\" (UID: \"43f709d6-33c8-42bd-8dc4-c92a781c1e78\") " Nov 25 11:28:12 crc kubenswrapper[4821]: I1125 11:28:12.423140 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/43f709d6-33c8-42bd-8dc4-c92a781c1e78-utilities" (OuterVolumeSpecName: "utilities") pod "43f709d6-33c8-42bd-8dc4-c92a781c1e78" (UID: "43f709d6-33c8-42bd-8dc4-c92a781c1e78"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:28:12 crc kubenswrapper[4821]: I1125 11:28:12.428716 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43f709d6-33c8-42bd-8dc4-c92a781c1e78-kube-api-access-f6djt" (OuterVolumeSpecName: "kube-api-access-f6djt") pod "43f709d6-33c8-42bd-8dc4-c92a781c1e78" (UID: "43f709d6-33c8-42bd-8dc4-c92a781c1e78"). InnerVolumeSpecName "kube-api-access-f6djt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:28:12 crc kubenswrapper[4821]: I1125 11:28:12.519230 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/43f709d6-33c8-42bd-8dc4-c92a781c1e78-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "43f709d6-33c8-42bd-8dc4-c92a781c1e78" (UID: "43f709d6-33c8-42bd-8dc4-c92a781c1e78"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:28:12 crc kubenswrapper[4821]: I1125 11:28:12.524440 4821 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43f709d6-33c8-42bd-8dc4-c92a781c1e78-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 11:28:12 crc kubenswrapper[4821]: I1125 11:28:12.524808 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f6djt\" (UniqueName: \"kubernetes.io/projected/43f709d6-33c8-42bd-8dc4-c92a781c1e78-kube-api-access-f6djt\") on node \"crc\" DevicePath \"\"" Nov 25 11:28:12 crc kubenswrapper[4821]: I1125 11:28:12.524826 4821 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43f709d6-33c8-42bd-8dc4-c92a781c1e78-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 11:28:12 crc kubenswrapper[4821]: I1125 11:28:12.751251 4821 generic.go:334] "Generic (PLEG): container finished" podID="43f709d6-33c8-42bd-8dc4-c92a781c1e78" containerID="ee3efd40ae0589435ec2d7c71811e1afff1864c39066562f589269cdfee2b524" exitCode=0 Nov 25 11:28:12 crc kubenswrapper[4821]: I1125 11:28:12.751307 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t2h46" event={"ID":"43f709d6-33c8-42bd-8dc4-c92a781c1e78","Type":"ContainerDied","Data":"ee3efd40ae0589435ec2d7c71811e1afff1864c39066562f589269cdfee2b524"} Nov 25 11:28:12 crc kubenswrapper[4821]: I1125 11:28:12.751339 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t2h46" event={"ID":"43f709d6-33c8-42bd-8dc4-c92a781c1e78","Type":"ContainerDied","Data":"708cce6ffc8a490d12fe3fde38091c1e26b2e738ab5ecc1eb76b4a8c6b33bd52"} Nov 25 11:28:12 crc kubenswrapper[4821]: I1125 11:28:12.751361 4821 scope.go:117] "RemoveContainer" containerID="ee3efd40ae0589435ec2d7c71811e1afff1864c39066562f589269cdfee2b524" Nov 25 11:28:12 crc kubenswrapper[4821]: I1125 11:28:12.751470 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t2h46" Nov 25 11:28:12 crc kubenswrapper[4821]: I1125 11:28:12.785611 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-t2h46"] Nov 25 11:28:12 crc kubenswrapper[4821]: I1125 11:28:12.797409 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-t2h46"] Nov 25 11:28:12 crc kubenswrapper[4821]: I1125 11:28:12.797536 4821 scope.go:117] "RemoveContainer" containerID="de163f58a985eae43706092ae42d42d28bba0555977a3307e163c65fcda1f996" Nov 25 11:28:12 crc kubenswrapper[4821]: I1125 11:28:12.818956 4821 scope.go:117] "RemoveContainer" containerID="ef57d227c9f79aefaf9c26b270a02c65861bb79ebe0f2835dc1f28b26c07a76e" Nov 25 11:28:13 crc kubenswrapper[4821]: I1125 11:28:13.527979 4821 scope.go:117] "RemoveContainer" containerID="ee3efd40ae0589435ec2d7c71811e1afff1864c39066562f589269cdfee2b524" Nov 25 11:28:13 crc kubenswrapper[4821]: E1125 11:28:13.529652 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee3efd40ae0589435ec2d7c71811e1afff1864c39066562f589269cdfee2b524\": container with ID starting with ee3efd40ae0589435ec2d7c71811e1afff1864c39066562f589269cdfee2b524 not found: ID does not exist" containerID="ee3efd40ae0589435ec2d7c71811e1afff1864c39066562f589269cdfee2b524" Nov 25 11:28:13 crc kubenswrapper[4821]: I1125 11:28:13.529706 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee3efd40ae0589435ec2d7c71811e1afff1864c39066562f589269cdfee2b524"} err="failed to get container status \"ee3efd40ae0589435ec2d7c71811e1afff1864c39066562f589269cdfee2b524\": rpc error: code = NotFound desc = could not find container \"ee3efd40ae0589435ec2d7c71811e1afff1864c39066562f589269cdfee2b524\": container with ID starting with ee3efd40ae0589435ec2d7c71811e1afff1864c39066562f589269cdfee2b524 not found: ID does not exist" Nov 25 11:28:13 crc kubenswrapper[4821]: I1125 11:28:13.529746 4821 scope.go:117] "RemoveContainer" containerID="de163f58a985eae43706092ae42d42d28bba0555977a3307e163c65fcda1f996" Nov 25 11:28:13 crc kubenswrapper[4821]: E1125 11:28:13.530398 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de163f58a985eae43706092ae42d42d28bba0555977a3307e163c65fcda1f996\": container with ID starting with de163f58a985eae43706092ae42d42d28bba0555977a3307e163c65fcda1f996 not found: ID does not exist" containerID="de163f58a985eae43706092ae42d42d28bba0555977a3307e163c65fcda1f996" Nov 25 11:28:13 crc kubenswrapper[4821]: I1125 11:28:13.530470 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de163f58a985eae43706092ae42d42d28bba0555977a3307e163c65fcda1f996"} err="failed to get container status \"de163f58a985eae43706092ae42d42d28bba0555977a3307e163c65fcda1f996\": rpc error: code = NotFound desc = could not find container \"de163f58a985eae43706092ae42d42d28bba0555977a3307e163c65fcda1f996\": container with ID starting with de163f58a985eae43706092ae42d42d28bba0555977a3307e163c65fcda1f996 not found: ID does not exist" Nov 25 11:28:13 crc kubenswrapper[4821]: I1125 11:28:13.530518 4821 scope.go:117] "RemoveContainer" containerID="ef57d227c9f79aefaf9c26b270a02c65861bb79ebe0f2835dc1f28b26c07a76e" Nov 25 11:28:13 crc kubenswrapper[4821]: E1125 11:28:13.531176 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef57d227c9f79aefaf9c26b270a02c65861bb79ebe0f2835dc1f28b26c07a76e\": container with ID starting with ef57d227c9f79aefaf9c26b270a02c65861bb79ebe0f2835dc1f28b26c07a76e not found: ID does not exist" containerID="ef57d227c9f79aefaf9c26b270a02c65861bb79ebe0f2835dc1f28b26c07a76e" Nov 25 11:28:13 crc kubenswrapper[4821]: I1125 11:28:13.531270 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef57d227c9f79aefaf9c26b270a02c65861bb79ebe0f2835dc1f28b26c07a76e"} err="failed to get container status \"ef57d227c9f79aefaf9c26b270a02c65861bb79ebe0f2835dc1f28b26c07a76e\": rpc error: code = NotFound desc = could not find container \"ef57d227c9f79aefaf9c26b270a02c65861bb79ebe0f2835dc1f28b26c07a76e\": container with ID starting with ef57d227c9f79aefaf9c26b270a02c65861bb79ebe0f2835dc1f28b26c07a76e not found: ID does not exist" Nov 25 11:28:14 crc kubenswrapper[4821]: I1125 11:28:14.124762 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43f709d6-33c8-42bd-8dc4-c92a781c1e78" path="/var/lib/kubelet/pods/43f709d6-33c8-42bd-8dc4-c92a781c1e78/volumes" Nov 25 11:28:54 crc kubenswrapper[4821]: I1125 11:28:54.626859 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-s8dl4"] Nov 25 11:28:54 crc kubenswrapper[4821]: E1125 11:28:54.628771 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43f709d6-33c8-42bd-8dc4-c92a781c1e78" containerName="extract-content" Nov 25 11:28:54 crc kubenswrapper[4821]: I1125 11:28:54.628791 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="43f709d6-33c8-42bd-8dc4-c92a781c1e78" containerName="extract-content" Nov 25 11:28:54 crc kubenswrapper[4821]: E1125 11:28:54.628832 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43f709d6-33c8-42bd-8dc4-c92a781c1e78" containerName="extract-utilities" Nov 25 11:28:54 crc kubenswrapper[4821]: I1125 11:28:54.628838 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="43f709d6-33c8-42bd-8dc4-c92a781c1e78" containerName="extract-utilities" Nov 25 11:28:54 crc kubenswrapper[4821]: E1125 11:28:54.628866 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43f709d6-33c8-42bd-8dc4-c92a781c1e78" containerName="registry-server" Nov 25 11:28:54 crc kubenswrapper[4821]: I1125 11:28:54.628875 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="43f709d6-33c8-42bd-8dc4-c92a781c1e78" containerName="registry-server" Nov 25 11:28:54 crc kubenswrapper[4821]: I1125 11:28:54.629105 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="43f709d6-33c8-42bd-8dc4-c92a781c1e78" containerName="registry-server" Nov 25 11:28:54 crc kubenswrapper[4821]: I1125 11:28:54.632563 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s8dl4" Nov 25 11:28:54 crc kubenswrapper[4821]: I1125 11:28:54.646970 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-s8dl4"] Nov 25 11:28:54 crc kubenswrapper[4821]: I1125 11:28:54.700104 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xgt6m\" (UniqueName: \"kubernetes.io/projected/9eca4516-73c8-4bb9-ae67-5047b083f7f1-kube-api-access-xgt6m\") pod \"community-operators-s8dl4\" (UID: \"9eca4516-73c8-4bb9-ae67-5047b083f7f1\") " pod="openshift-marketplace/community-operators-s8dl4" Nov 25 11:28:54 crc kubenswrapper[4821]: I1125 11:28:54.700274 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9eca4516-73c8-4bb9-ae67-5047b083f7f1-catalog-content\") pod \"community-operators-s8dl4\" (UID: \"9eca4516-73c8-4bb9-ae67-5047b083f7f1\") " pod="openshift-marketplace/community-operators-s8dl4" Nov 25 11:28:54 crc kubenswrapper[4821]: I1125 11:28:54.700319 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9eca4516-73c8-4bb9-ae67-5047b083f7f1-utilities\") pod \"community-operators-s8dl4\" (UID: \"9eca4516-73c8-4bb9-ae67-5047b083f7f1\") " pod="openshift-marketplace/community-operators-s8dl4" Nov 25 11:28:54 crc kubenswrapper[4821]: I1125 11:28:54.802055 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9eca4516-73c8-4bb9-ae67-5047b083f7f1-catalog-content\") pod \"community-operators-s8dl4\" (UID: \"9eca4516-73c8-4bb9-ae67-5047b083f7f1\") " pod="openshift-marketplace/community-operators-s8dl4" Nov 25 11:28:54 crc kubenswrapper[4821]: I1125 11:28:54.802121 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9eca4516-73c8-4bb9-ae67-5047b083f7f1-utilities\") pod \"community-operators-s8dl4\" (UID: \"9eca4516-73c8-4bb9-ae67-5047b083f7f1\") " pod="openshift-marketplace/community-operators-s8dl4" Nov 25 11:28:54 crc kubenswrapper[4821]: I1125 11:28:54.802267 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xgt6m\" (UniqueName: \"kubernetes.io/projected/9eca4516-73c8-4bb9-ae67-5047b083f7f1-kube-api-access-xgt6m\") pod \"community-operators-s8dl4\" (UID: \"9eca4516-73c8-4bb9-ae67-5047b083f7f1\") " pod="openshift-marketplace/community-operators-s8dl4" Nov 25 11:28:54 crc kubenswrapper[4821]: I1125 11:28:54.802605 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9eca4516-73c8-4bb9-ae67-5047b083f7f1-utilities\") pod \"community-operators-s8dl4\" (UID: \"9eca4516-73c8-4bb9-ae67-5047b083f7f1\") " pod="openshift-marketplace/community-operators-s8dl4" Nov 25 11:28:54 crc kubenswrapper[4821]: I1125 11:28:54.802666 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9eca4516-73c8-4bb9-ae67-5047b083f7f1-catalog-content\") pod \"community-operators-s8dl4\" (UID: \"9eca4516-73c8-4bb9-ae67-5047b083f7f1\") " pod="openshift-marketplace/community-operators-s8dl4" Nov 25 11:28:54 crc kubenswrapper[4821]: I1125 11:28:54.828953 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xgt6m\" (UniqueName: \"kubernetes.io/projected/9eca4516-73c8-4bb9-ae67-5047b083f7f1-kube-api-access-xgt6m\") pod \"community-operators-s8dl4\" (UID: \"9eca4516-73c8-4bb9-ae67-5047b083f7f1\") " pod="openshift-marketplace/community-operators-s8dl4" Nov 25 11:28:54 crc kubenswrapper[4821]: I1125 11:28:54.962094 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s8dl4" Nov 25 11:28:55 crc kubenswrapper[4821]: I1125 11:28:55.625914 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-s8dl4"] Nov 25 11:28:56 crc kubenswrapper[4821]: I1125 11:28:56.133780 4821 generic.go:334] "Generic (PLEG): container finished" podID="9eca4516-73c8-4bb9-ae67-5047b083f7f1" containerID="91c68c19f1ca972b7886a01115f5c570ced1cd2b9a8c9da2b8824d32562d224c" exitCode=0 Nov 25 11:28:56 crc kubenswrapper[4821]: I1125 11:28:56.133875 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s8dl4" event={"ID":"9eca4516-73c8-4bb9-ae67-5047b083f7f1","Type":"ContainerDied","Data":"91c68c19f1ca972b7886a01115f5c570ced1cd2b9a8c9da2b8824d32562d224c"} Nov 25 11:28:56 crc kubenswrapper[4821]: I1125 11:28:56.134065 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s8dl4" event={"ID":"9eca4516-73c8-4bb9-ae67-5047b083f7f1","Type":"ContainerStarted","Data":"b346725d3173041adb7204ce743ac8efb67463ab99e7cc78b84b1166a0fe47b2"} Nov 25 11:28:57 crc kubenswrapper[4821]: I1125 11:28:57.145361 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s8dl4" event={"ID":"9eca4516-73c8-4bb9-ae67-5047b083f7f1","Type":"ContainerStarted","Data":"430af09e5275a5f79306c5a0e8bf15b49533798f8c4ce0ee8696059d6e1ab501"} Nov 25 11:28:58 crc kubenswrapper[4821]: I1125 11:28:58.154977 4821 generic.go:334] "Generic (PLEG): container finished" podID="9eca4516-73c8-4bb9-ae67-5047b083f7f1" containerID="430af09e5275a5f79306c5a0e8bf15b49533798f8c4ce0ee8696059d6e1ab501" exitCode=0 Nov 25 11:28:58 crc kubenswrapper[4821]: I1125 11:28:58.155024 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s8dl4" event={"ID":"9eca4516-73c8-4bb9-ae67-5047b083f7f1","Type":"ContainerDied","Data":"430af09e5275a5f79306c5a0e8bf15b49533798f8c4ce0ee8696059d6e1ab501"} Nov 25 11:28:59 crc kubenswrapper[4821]: I1125 11:28:59.168353 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s8dl4" event={"ID":"9eca4516-73c8-4bb9-ae67-5047b083f7f1","Type":"ContainerStarted","Data":"3f13be7d06646ba0a45a3ba81fd54e563dbcc43e4770f536954c3b950481588e"} Nov 25 11:28:59 crc kubenswrapper[4821]: I1125 11:28:59.195230 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-s8dl4" podStartSLOduration=2.791316793 podStartE2EDuration="5.195211134s" podCreationTimestamp="2025-11-25 11:28:54 +0000 UTC" firstStartedPulling="2025-11-25 11:28:56.135892553 +0000 UTC m=+3406.672212400" lastFinishedPulling="2025-11-25 11:28:58.539786894 +0000 UTC m=+3409.076106741" observedRunningTime="2025-11-25 11:28:59.187639428 +0000 UTC m=+3409.723959275" watchObservedRunningTime="2025-11-25 11:28:59.195211134 +0000 UTC m=+3409.731530981" Nov 25 11:29:04 crc kubenswrapper[4821]: I1125 11:29:04.963178 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-s8dl4" Nov 25 11:29:04 crc kubenswrapper[4821]: I1125 11:29:04.963701 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-s8dl4" Nov 25 11:29:05 crc kubenswrapper[4821]: I1125 11:29:05.028493 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-s8dl4" Nov 25 11:29:05 crc kubenswrapper[4821]: I1125 11:29:05.259475 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-s8dl4" Nov 25 11:29:05 crc kubenswrapper[4821]: I1125 11:29:05.312351 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-s8dl4"] Nov 25 11:29:07 crc kubenswrapper[4821]: I1125 11:29:07.234053 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-s8dl4" podUID="9eca4516-73c8-4bb9-ae67-5047b083f7f1" containerName="registry-server" containerID="cri-o://3f13be7d06646ba0a45a3ba81fd54e563dbcc43e4770f536954c3b950481588e" gracePeriod=2 Nov 25 11:29:08 crc kubenswrapper[4821]: I1125 11:29:08.253575 4821 generic.go:334] "Generic (PLEG): container finished" podID="9eca4516-73c8-4bb9-ae67-5047b083f7f1" containerID="3f13be7d06646ba0a45a3ba81fd54e563dbcc43e4770f536954c3b950481588e" exitCode=0 Nov 25 11:29:08 crc kubenswrapper[4821]: I1125 11:29:08.253625 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s8dl4" event={"ID":"9eca4516-73c8-4bb9-ae67-5047b083f7f1","Type":"ContainerDied","Data":"3f13be7d06646ba0a45a3ba81fd54e563dbcc43e4770f536954c3b950481588e"} Nov 25 11:29:08 crc kubenswrapper[4821]: I1125 11:29:08.897832 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s8dl4" Nov 25 11:29:08 crc kubenswrapper[4821]: I1125 11:29:08.990969 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xgt6m\" (UniqueName: \"kubernetes.io/projected/9eca4516-73c8-4bb9-ae67-5047b083f7f1-kube-api-access-xgt6m\") pod \"9eca4516-73c8-4bb9-ae67-5047b083f7f1\" (UID: \"9eca4516-73c8-4bb9-ae67-5047b083f7f1\") " Nov 25 11:29:08 crc kubenswrapper[4821]: I1125 11:29:08.991086 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9eca4516-73c8-4bb9-ae67-5047b083f7f1-catalog-content\") pod \"9eca4516-73c8-4bb9-ae67-5047b083f7f1\" (UID: \"9eca4516-73c8-4bb9-ae67-5047b083f7f1\") " Nov 25 11:29:08 crc kubenswrapper[4821]: I1125 11:29:08.991182 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9eca4516-73c8-4bb9-ae67-5047b083f7f1-utilities\") pod \"9eca4516-73c8-4bb9-ae67-5047b083f7f1\" (UID: \"9eca4516-73c8-4bb9-ae67-5047b083f7f1\") " Nov 25 11:29:08 crc kubenswrapper[4821]: I1125 11:29:08.992312 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9eca4516-73c8-4bb9-ae67-5047b083f7f1-utilities" (OuterVolumeSpecName: "utilities") pod "9eca4516-73c8-4bb9-ae67-5047b083f7f1" (UID: "9eca4516-73c8-4bb9-ae67-5047b083f7f1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:29:08 crc kubenswrapper[4821]: I1125 11:29:08.997097 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9eca4516-73c8-4bb9-ae67-5047b083f7f1-kube-api-access-xgt6m" (OuterVolumeSpecName: "kube-api-access-xgt6m") pod "9eca4516-73c8-4bb9-ae67-5047b083f7f1" (UID: "9eca4516-73c8-4bb9-ae67-5047b083f7f1"). InnerVolumeSpecName "kube-api-access-xgt6m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:29:09 crc kubenswrapper[4821]: I1125 11:29:09.045472 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9eca4516-73c8-4bb9-ae67-5047b083f7f1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9eca4516-73c8-4bb9-ae67-5047b083f7f1" (UID: "9eca4516-73c8-4bb9-ae67-5047b083f7f1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:29:09 crc kubenswrapper[4821]: I1125 11:29:09.093027 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xgt6m\" (UniqueName: \"kubernetes.io/projected/9eca4516-73c8-4bb9-ae67-5047b083f7f1-kube-api-access-xgt6m\") on node \"crc\" DevicePath \"\"" Nov 25 11:29:09 crc kubenswrapper[4821]: I1125 11:29:09.093066 4821 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9eca4516-73c8-4bb9-ae67-5047b083f7f1-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 11:29:09 crc kubenswrapper[4821]: I1125 11:29:09.093075 4821 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9eca4516-73c8-4bb9-ae67-5047b083f7f1-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 11:29:09 crc kubenswrapper[4821]: I1125 11:29:09.266964 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s8dl4" event={"ID":"9eca4516-73c8-4bb9-ae67-5047b083f7f1","Type":"ContainerDied","Data":"b346725d3173041adb7204ce743ac8efb67463ab99e7cc78b84b1166a0fe47b2"} Nov 25 11:29:09 crc kubenswrapper[4821]: I1125 11:29:09.267015 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s8dl4" Nov 25 11:29:09 crc kubenswrapper[4821]: I1125 11:29:09.267023 4821 scope.go:117] "RemoveContainer" containerID="3f13be7d06646ba0a45a3ba81fd54e563dbcc43e4770f536954c3b950481588e" Nov 25 11:29:09 crc kubenswrapper[4821]: I1125 11:29:09.291199 4821 scope.go:117] "RemoveContainer" containerID="430af09e5275a5f79306c5a0e8bf15b49533798f8c4ce0ee8696059d6e1ab501" Nov 25 11:29:09 crc kubenswrapper[4821]: I1125 11:29:09.307996 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-s8dl4"] Nov 25 11:29:09 crc kubenswrapper[4821]: I1125 11:29:09.314839 4821 scope.go:117] "RemoveContainer" containerID="91c68c19f1ca972b7886a01115f5c570ced1cd2b9a8c9da2b8824d32562d224c" Nov 25 11:29:09 crc kubenswrapper[4821]: I1125 11:29:09.317155 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-s8dl4"] Nov 25 11:29:10 crc kubenswrapper[4821]: I1125 11:29:10.124822 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9eca4516-73c8-4bb9-ae67-5047b083f7f1" path="/var/lib/kubelet/pods/9eca4516-73c8-4bb9-ae67-5047b083f7f1/volumes" Nov 25 11:29:41 crc kubenswrapper[4821]: I1125 11:29:41.574450 4821 generic.go:334] "Generic (PLEG): container finished" podID="d8aea755-bf2f-4443-8196-46936ce3e2fd" containerID="2e8913f0f951c82e4bd46360606b86de76140c25a93d213dd39ebaae984f2756" exitCode=0 Nov 25 11:29:41 crc kubenswrapper[4821]: I1125 11:29:41.574536 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"d8aea755-bf2f-4443-8196-46936ce3e2fd","Type":"ContainerDied","Data":"2e8913f0f951c82e4bd46360606b86de76140c25a93d213dd39ebaae984f2756"} Nov 25 11:29:42 crc kubenswrapper[4821]: I1125 11:29:42.922388 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 25 11:29:43 crc kubenswrapper[4821]: I1125 11:29:43.041836 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d8aea755-bf2f-4443-8196-46936ce3e2fd-openstack-config\") pod \"d8aea755-bf2f-4443-8196-46936ce3e2fd\" (UID: \"d8aea755-bf2f-4443-8196-46936ce3e2fd\") " Nov 25 11:29:43 crc kubenswrapper[4821]: I1125 11:29:43.042352 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/d8aea755-bf2f-4443-8196-46936ce3e2fd-test-operator-ephemeral-temporary\") pod \"d8aea755-bf2f-4443-8196-46936ce3e2fd\" (UID: \"d8aea755-bf2f-4443-8196-46936ce3e2fd\") " Nov 25 11:29:43 crc kubenswrapper[4821]: I1125 11:29:43.042505 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zz66w\" (UniqueName: \"kubernetes.io/projected/d8aea755-bf2f-4443-8196-46936ce3e2fd-kube-api-access-zz66w\") pod \"d8aea755-bf2f-4443-8196-46936ce3e2fd\" (UID: \"d8aea755-bf2f-4443-8196-46936ce3e2fd\") " Nov 25 11:29:43 crc kubenswrapper[4821]: I1125 11:29:43.042652 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d8aea755-bf2f-4443-8196-46936ce3e2fd-openstack-config-secret\") pod \"d8aea755-bf2f-4443-8196-46936ce3e2fd\" (UID: \"d8aea755-bf2f-4443-8196-46936ce3e2fd\") " Nov 25 11:29:43 crc kubenswrapper[4821]: I1125 11:29:43.042767 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"d8aea755-bf2f-4443-8196-46936ce3e2fd\" (UID: \"d8aea755-bf2f-4443-8196-46936ce3e2fd\") " Nov 25 11:29:43 crc kubenswrapper[4821]: I1125 11:29:43.042887 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/d8aea755-bf2f-4443-8196-46936ce3e2fd-test-operator-ephemeral-workdir\") pod \"d8aea755-bf2f-4443-8196-46936ce3e2fd\" (UID: \"d8aea755-bf2f-4443-8196-46936ce3e2fd\") " Nov 25 11:29:43 crc kubenswrapper[4821]: I1125 11:29:43.042969 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/d8aea755-bf2f-4443-8196-46936ce3e2fd-ca-certs\") pod \"d8aea755-bf2f-4443-8196-46936ce3e2fd\" (UID: \"d8aea755-bf2f-4443-8196-46936ce3e2fd\") " Nov 25 11:29:43 crc kubenswrapper[4821]: I1125 11:29:43.043066 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d8aea755-bf2f-4443-8196-46936ce3e2fd-ssh-key\") pod \"d8aea755-bf2f-4443-8196-46936ce3e2fd\" (UID: \"d8aea755-bf2f-4443-8196-46936ce3e2fd\") " Nov 25 11:29:43 crc kubenswrapper[4821]: I1125 11:29:43.043159 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d8aea755-bf2f-4443-8196-46936ce3e2fd-config-data\") pod \"d8aea755-bf2f-4443-8196-46936ce3e2fd\" (UID: \"d8aea755-bf2f-4443-8196-46936ce3e2fd\") " Nov 25 11:29:43 crc kubenswrapper[4821]: I1125 11:29:43.044552 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d8aea755-bf2f-4443-8196-46936ce3e2fd-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "d8aea755-bf2f-4443-8196-46936ce3e2fd" (UID: "d8aea755-bf2f-4443-8196-46936ce3e2fd"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:29:43 crc kubenswrapper[4821]: I1125 11:29:43.049834 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8aea755-bf2f-4443-8196-46936ce3e2fd-kube-api-access-zz66w" (OuterVolumeSpecName: "kube-api-access-zz66w") pod "d8aea755-bf2f-4443-8196-46936ce3e2fd" (UID: "d8aea755-bf2f-4443-8196-46936ce3e2fd"). InnerVolumeSpecName "kube-api-access-zz66w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:29:43 crc kubenswrapper[4821]: I1125 11:29:43.050182 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d8aea755-bf2f-4443-8196-46936ce3e2fd-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "d8aea755-bf2f-4443-8196-46936ce3e2fd" (UID: "d8aea755-bf2f-4443-8196-46936ce3e2fd"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:29:43 crc kubenswrapper[4821]: I1125 11:29:43.050710 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8aea755-bf2f-4443-8196-46936ce3e2fd-config-data" (OuterVolumeSpecName: "config-data") pod "d8aea755-bf2f-4443-8196-46936ce3e2fd" (UID: "d8aea755-bf2f-4443-8196-46936ce3e2fd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:29:43 crc kubenswrapper[4821]: I1125 11:29:43.053773 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "test-operator-logs") pod "d8aea755-bf2f-4443-8196-46936ce3e2fd" (UID: "d8aea755-bf2f-4443-8196-46936ce3e2fd"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 11:29:43 crc kubenswrapper[4821]: I1125 11:29:43.073241 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8aea755-bf2f-4443-8196-46936ce3e2fd-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "d8aea755-bf2f-4443-8196-46936ce3e2fd" (UID: "d8aea755-bf2f-4443-8196-46936ce3e2fd"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:29:43 crc kubenswrapper[4821]: I1125 11:29:43.075316 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8aea755-bf2f-4443-8196-46936ce3e2fd-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "d8aea755-bf2f-4443-8196-46936ce3e2fd" (UID: "d8aea755-bf2f-4443-8196-46936ce3e2fd"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:29:43 crc kubenswrapper[4821]: I1125 11:29:43.085854 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8aea755-bf2f-4443-8196-46936ce3e2fd-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d8aea755-bf2f-4443-8196-46936ce3e2fd" (UID: "d8aea755-bf2f-4443-8196-46936ce3e2fd"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:29:43 crc kubenswrapper[4821]: I1125 11:29:43.125895 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8aea755-bf2f-4443-8196-46936ce3e2fd-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "d8aea755-bf2f-4443-8196-46936ce3e2fd" (UID: "d8aea755-bf2f-4443-8196-46936ce3e2fd"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:29:43 crc kubenswrapper[4821]: I1125 11:29:43.145533 4821 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d8aea755-bf2f-4443-8196-46936ce3e2fd-openstack-config\") on node \"crc\" DevicePath \"\"" Nov 25 11:29:43 crc kubenswrapper[4821]: I1125 11:29:43.145573 4821 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/d8aea755-bf2f-4443-8196-46936ce3e2fd-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Nov 25 11:29:43 crc kubenswrapper[4821]: I1125 11:29:43.145586 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zz66w\" (UniqueName: \"kubernetes.io/projected/d8aea755-bf2f-4443-8196-46936ce3e2fd-kube-api-access-zz66w\") on node \"crc\" DevicePath \"\"" Nov 25 11:29:43 crc kubenswrapper[4821]: I1125 11:29:43.145596 4821 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d8aea755-bf2f-4443-8196-46936ce3e2fd-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 25 11:29:43 crc kubenswrapper[4821]: I1125 11:29:43.145628 4821 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Nov 25 11:29:43 crc kubenswrapper[4821]: I1125 11:29:43.145638 4821 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/d8aea755-bf2f-4443-8196-46936ce3e2fd-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Nov 25 11:29:43 crc kubenswrapper[4821]: I1125 11:29:43.145646 4821 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/d8aea755-bf2f-4443-8196-46936ce3e2fd-ca-certs\") on node \"crc\" DevicePath \"\"" Nov 25 11:29:43 crc kubenswrapper[4821]: I1125 11:29:43.145658 4821 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d8aea755-bf2f-4443-8196-46936ce3e2fd-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 11:29:43 crc kubenswrapper[4821]: I1125 11:29:43.145666 4821 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d8aea755-bf2f-4443-8196-46936ce3e2fd-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 11:29:43 crc kubenswrapper[4821]: I1125 11:29:43.166259 4821 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Nov 25 11:29:43 crc kubenswrapper[4821]: I1125 11:29:43.247460 4821 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Nov 25 11:29:43 crc kubenswrapper[4821]: I1125 11:29:43.593326 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"d8aea755-bf2f-4443-8196-46936ce3e2fd","Type":"ContainerDied","Data":"2dc6c27588da93a980af00c534f445a80ec1a579e213d38d1d64998bc6cebf78"} Nov 25 11:29:43 crc kubenswrapper[4821]: I1125 11:29:43.593370 4821 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2dc6c27588da93a980af00c534f445a80ec1a579e213d38d1d64998bc6cebf78" Nov 25 11:29:43 crc kubenswrapper[4821]: I1125 11:29:43.593395 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 25 11:29:52 crc kubenswrapper[4821]: I1125 11:29:52.063003 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Nov 25 11:29:52 crc kubenswrapper[4821]: E1125 11:29:52.063956 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8aea755-bf2f-4443-8196-46936ce3e2fd" containerName="tempest-tests-tempest-tests-runner" Nov 25 11:29:52 crc kubenswrapper[4821]: I1125 11:29:52.063971 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8aea755-bf2f-4443-8196-46936ce3e2fd" containerName="tempest-tests-tempest-tests-runner" Nov 25 11:29:52 crc kubenswrapper[4821]: E1125 11:29:52.063995 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9eca4516-73c8-4bb9-ae67-5047b083f7f1" containerName="extract-content" Nov 25 11:29:52 crc kubenswrapper[4821]: I1125 11:29:52.064001 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="9eca4516-73c8-4bb9-ae67-5047b083f7f1" containerName="extract-content" Nov 25 11:29:52 crc kubenswrapper[4821]: E1125 11:29:52.064012 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9eca4516-73c8-4bb9-ae67-5047b083f7f1" containerName="registry-server" Nov 25 11:29:52 crc kubenswrapper[4821]: I1125 11:29:52.064019 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="9eca4516-73c8-4bb9-ae67-5047b083f7f1" containerName="registry-server" Nov 25 11:29:52 crc kubenswrapper[4821]: E1125 11:29:52.064043 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9eca4516-73c8-4bb9-ae67-5047b083f7f1" containerName="extract-utilities" Nov 25 11:29:52 crc kubenswrapper[4821]: I1125 11:29:52.064049 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="9eca4516-73c8-4bb9-ae67-5047b083f7f1" containerName="extract-utilities" Nov 25 11:29:52 crc kubenswrapper[4821]: I1125 11:29:52.064398 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="9eca4516-73c8-4bb9-ae67-5047b083f7f1" containerName="registry-server" Nov 25 11:29:52 crc kubenswrapper[4821]: I1125 11:29:52.064414 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8aea755-bf2f-4443-8196-46936ce3e2fd" containerName="tempest-tests-tempest-tests-runner" Nov 25 11:29:52 crc kubenswrapper[4821]: I1125 11:29:52.065185 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 25 11:29:52 crc kubenswrapper[4821]: I1125 11:29:52.067710 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-qrs9n" Nov 25 11:29:52 crc kubenswrapper[4821]: I1125 11:29:52.074240 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Nov 25 11:29:52 crc kubenswrapper[4821]: I1125 11:29:52.117588 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s7jpx\" (UniqueName: \"kubernetes.io/projected/adcf552f-2365-451b-b4e2-12b58631195e-kube-api-access-s7jpx\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"adcf552f-2365-451b-b4e2-12b58631195e\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 25 11:29:52 crc kubenswrapper[4821]: I1125 11:29:52.117679 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"adcf552f-2365-451b-b4e2-12b58631195e\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 25 11:29:52 crc kubenswrapper[4821]: I1125 11:29:52.220423 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s7jpx\" (UniqueName: \"kubernetes.io/projected/adcf552f-2365-451b-b4e2-12b58631195e-kube-api-access-s7jpx\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"adcf552f-2365-451b-b4e2-12b58631195e\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 25 11:29:52 crc kubenswrapper[4821]: I1125 11:29:52.220567 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"adcf552f-2365-451b-b4e2-12b58631195e\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 25 11:29:52 crc kubenswrapper[4821]: I1125 11:29:52.221644 4821 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"adcf552f-2365-451b-b4e2-12b58631195e\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 25 11:29:52 crc kubenswrapper[4821]: I1125 11:29:52.243729 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s7jpx\" (UniqueName: \"kubernetes.io/projected/adcf552f-2365-451b-b4e2-12b58631195e-kube-api-access-s7jpx\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"adcf552f-2365-451b-b4e2-12b58631195e\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 25 11:29:52 crc kubenswrapper[4821]: I1125 11:29:52.252529 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"adcf552f-2365-451b-b4e2-12b58631195e\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 25 11:29:52 crc kubenswrapper[4821]: I1125 11:29:52.400694 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 25 11:29:52 crc kubenswrapper[4821]: I1125 11:29:52.921754 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Nov 25 11:29:53 crc kubenswrapper[4821]: I1125 11:29:53.698383 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"adcf552f-2365-451b-b4e2-12b58631195e","Type":"ContainerStarted","Data":"bac7c235d492cf8d2863b50329c7dbaceeda18050ec6eb8fb50b746f5d89ff14"} Nov 25 11:29:55 crc kubenswrapper[4821]: I1125 11:29:55.716842 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"adcf552f-2365-451b-b4e2-12b58631195e","Type":"ContainerStarted","Data":"8ce563bf6e4691ccbd26b91ef365201c39c526efba2dffdc93e0ce20401bb8ab"} Nov 25 11:29:55 crc kubenswrapper[4821]: I1125 11:29:55.739585 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=1.35149314 podStartE2EDuration="3.739567255s" podCreationTimestamp="2025-11-25 11:29:52 +0000 UTC" firstStartedPulling="2025-11-25 11:29:52.922544875 +0000 UTC m=+3463.458864712" lastFinishedPulling="2025-11-25 11:29:55.31061898 +0000 UTC m=+3465.846938827" observedRunningTime="2025-11-25 11:29:55.734946062 +0000 UTC m=+3466.271265909" watchObservedRunningTime="2025-11-25 11:29:55.739567255 +0000 UTC m=+3466.275887102" Nov 25 11:29:59 crc kubenswrapper[4821]: I1125 11:29:59.803047 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-529qb"] Nov 25 11:29:59 crc kubenswrapper[4821]: I1125 11:29:59.806741 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-529qb" Nov 25 11:29:59 crc kubenswrapper[4821]: I1125 11:29:59.817325 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-529qb"] Nov 25 11:29:59 crc kubenswrapper[4821]: I1125 11:29:59.975311 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/410373d3-842e-43b3-a2d8-2293c543a4ac-catalog-content\") pod \"certified-operators-529qb\" (UID: \"410373d3-842e-43b3-a2d8-2293c543a4ac\") " pod="openshift-marketplace/certified-operators-529qb" Nov 25 11:29:59 crc kubenswrapper[4821]: I1125 11:29:59.975559 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/410373d3-842e-43b3-a2d8-2293c543a4ac-utilities\") pod \"certified-operators-529qb\" (UID: \"410373d3-842e-43b3-a2d8-2293c543a4ac\") " pod="openshift-marketplace/certified-operators-529qb" Nov 25 11:29:59 crc kubenswrapper[4821]: I1125 11:29:59.975704 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kpjrm\" (UniqueName: \"kubernetes.io/projected/410373d3-842e-43b3-a2d8-2293c543a4ac-kube-api-access-kpjrm\") pod \"certified-operators-529qb\" (UID: \"410373d3-842e-43b3-a2d8-2293c543a4ac\") " pod="openshift-marketplace/certified-operators-529qb" Nov 25 11:30:00 crc kubenswrapper[4821]: I1125 11:30:00.077238 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/410373d3-842e-43b3-a2d8-2293c543a4ac-catalog-content\") pod \"certified-operators-529qb\" (UID: \"410373d3-842e-43b3-a2d8-2293c543a4ac\") " pod="openshift-marketplace/certified-operators-529qb" Nov 25 11:30:00 crc kubenswrapper[4821]: I1125 11:30:00.077604 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/410373d3-842e-43b3-a2d8-2293c543a4ac-utilities\") pod \"certified-operators-529qb\" (UID: \"410373d3-842e-43b3-a2d8-2293c543a4ac\") " pod="openshift-marketplace/certified-operators-529qb" Nov 25 11:30:00 crc kubenswrapper[4821]: I1125 11:30:00.077749 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kpjrm\" (UniqueName: \"kubernetes.io/projected/410373d3-842e-43b3-a2d8-2293c543a4ac-kube-api-access-kpjrm\") pod \"certified-operators-529qb\" (UID: \"410373d3-842e-43b3-a2d8-2293c543a4ac\") " pod="openshift-marketplace/certified-operators-529qb" Nov 25 11:30:00 crc kubenswrapper[4821]: I1125 11:30:00.078005 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/410373d3-842e-43b3-a2d8-2293c543a4ac-utilities\") pod \"certified-operators-529qb\" (UID: \"410373d3-842e-43b3-a2d8-2293c543a4ac\") " pod="openshift-marketplace/certified-operators-529qb" Nov 25 11:30:00 crc kubenswrapper[4821]: I1125 11:30:00.077782 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/410373d3-842e-43b3-a2d8-2293c543a4ac-catalog-content\") pod \"certified-operators-529qb\" (UID: \"410373d3-842e-43b3-a2d8-2293c543a4ac\") " pod="openshift-marketplace/certified-operators-529qb" Nov 25 11:30:00 crc kubenswrapper[4821]: I1125 11:30:00.105339 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kpjrm\" (UniqueName: \"kubernetes.io/projected/410373d3-842e-43b3-a2d8-2293c543a4ac-kube-api-access-kpjrm\") pod \"certified-operators-529qb\" (UID: \"410373d3-842e-43b3-a2d8-2293c543a4ac\") " pod="openshift-marketplace/certified-operators-529qb" Nov 25 11:30:00 crc kubenswrapper[4821]: I1125 11:30:00.145713 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-529qb" Nov 25 11:30:00 crc kubenswrapper[4821]: I1125 11:30:00.191892 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401170-gwq95"] Nov 25 11:30:00 crc kubenswrapper[4821]: I1125 11:30:00.193180 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401170-gwq95" Nov 25 11:30:00 crc kubenswrapper[4821]: I1125 11:30:00.195520 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 25 11:30:00 crc kubenswrapper[4821]: I1125 11:30:00.195719 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 25 11:30:00 crc kubenswrapper[4821]: I1125 11:30:00.209695 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401170-gwq95"] Nov 25 11:30:00 crc kubenswrapper[4821]: I1125 11:30:00.281682 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/47cd3f86-f784-494c-b959-8125ea191612-secret-volume\") pod \"collect-profiles-29401170-gwq95\" (UID: \"47cd3f86-f784-494c-b959-8125ea191612\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401170-gwq95" Nov 25 11:30:00 crc kubenswrapper[4821]: I1125 11:30:00.281820 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/47cd3f86-f784-494c-b959-8125ea191612-config-volume\") pod \"collect-profiles-29401170-gwq95\" (UID: \"47cd3f86-f784-494c-b959-8125ea191612\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401170-gwq95" Nov 25 11:30:00 crc kubenswrapper[4821]: I1125 11:30:00.281916 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fs2vw\" (UniqueName: \"kubernetes.io/projected/47cd3f86-f784-494c-b959-8125ea191612-kube-api-access-fs2vw\") pod \"collect-profiles-29401170-gwq95\" (UID: \"47cd3f86-f784-494c-b959-8125ea191612\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401170-gwq95" Nov 25 11:30:00 crc kubenswrapper[4821]: I1125 11:30:00.390303 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fs2vw\" (UniqueName: \"kubernetes.io/projected/47cd3f86-f784-494c-b959-8125ea191612-kube-api-access-fs2vw\") pod \"collect-profiles-29401170-gwq95\" (UID: \"47cd3f86-f784-494c-b959-8125ea191612\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401170-gwq95" Nov 25 11:30:00 crc kubenswrapper[4821]: I1125 11:30:00.391254 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/47cd3f86-f784-494c-b959-8125ea191612-secret-volume\") pod \"collect-profiles-29401170-gwq95\" (UID: \"47cd3f86-f784-494c-b959-8125ea191612\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401170-gwq95" Nov 25 11:30:00 crc kubenswrapper[4821]: I1125 11:30:00.391418 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/47cd3f86-f784-494c-b959-8125ea191612-config-volume\") pod \"collect-profiles-29401170-gwq95\" (UID: \"47cd3f86-f784-494c-b959-8125ea191612\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401170-gwq95" Nov 25 11:30:00 crc kubenswrapper[4821]: I1125 11:30:00.392954 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/47cd3f86-f784-494c-b959-8125ea191612-config-volume\") pod \"collect-profiles-29401170-gwq95\" (UID: \"47cd3f86-f784-494c-b959-8125ea191612\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401170-gwq95" Nov 25 11:30:00 crc kubenswrapper[4821]: I1125 11:30:00.396835 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/47cd3f86-f784-494c-b959-8125ea191612-secret-volume\") pod \"collect-profiles-29401170-gwq95\" (UID: \"47cd3f86-f784-494c-b959-8125ea191612\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401170-gwq95" Nov 25 11:30:00 crc kubenswrapper[4821]: I1125 11:30:00.415018 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fs2vw\" (UniqueName: \"kubernetes.io/projected/47cd3f86-f784-494c-b959-8125ea191612-kube-api-access-fs2vw\") pod \"collect-profiles-29401170-gwq95\" (UID: \"47cd3f86-f784-494c-b959-8125ea191612\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401170-gwq95" Nov 25 11:30:00 crc kubenswrapper[4821]: I1125 11:30:00.589643 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401170-gwq95" Nov 25 11:30:00 crc kubenswrapper[4821]: I1125 11:30:00.702588 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-529qb"] Nov 25 11:30:00 crc kubenswrapper[4821]: I1125 11:30:00.772279 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-529qb" event={"ID":"410373d3-842e-43b3-a2d8-2293c543a4ac","Type":"ContainerStarted","Data":"c3c780eecba350cbc2ec7ea637e1cf6255d26a8a2d132e1fa0077d731ed554db"} Nov 25 11:30:01 crc kubenswrapper[4821]: I1125 11:30:01.030585 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401170-gwq95"] Nov 25 11:30:01 crc kubenswrapper[4821]: I1125 11:30:01.780677 4821 generic.go:334] "Generic (PLEG): container finished" podID="47cd3f86-f784-494c-b959-8125ea191612" containerID="def1e2fed0737fd5a5bf47a5dec6f60751da6ce415c2e483c291785874f4e04f" exitCode=0 Nov 25 11:30:01 crc kubenswrapper[4821]: I1125 11:30:01.780736 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401170-gwq95" event={"ID":"47cd3f86-f784-494c-b959-8125ea191612","Type":"ContainerDied","Data":"def1e2fed0737fd5a5bf47a5dec6f60751da6ce415c2e483c291785874f4e04f"} Nov 25 11:30:01 crc kubenswrapper[4821]: I1125 11:30:01.781024 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401170-gwq95" event={"ID":"47cd3f86-f784-494c-b959-8125ea191612","Type":"ContainerStarted","Data":"8ca707933bce8ec850551f310f7c21919fa41228ed189c6f61e1f1dedd371f9e"} Nov 25 11:30:01 crc kubenswrapper[4821]: I1125 11:30:01.782311 4821 generic.go:334] "Generic (PLEG): container finished" podID="410373d3-842e-43b3-a2d8-2293c543a4ac" containerID="2ea5263999efc0044a91e5354fe595c0fd765326e91add89922a68e24c9d9674" exitCode=0 Nov 25 11:30:01 crc kubenswrapper[4821]: I1125 11:30:01.782342 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-529qb" event={"ID":"410373d3-842e-43b3-a2d8-2293c543a4ac","Type":"ContainerDied","Data":"2ea5263999efc0044a91e5354fe595c0fd765326e91add89922a68e24c9d9674"} Nov 25 11:30:03 crc kubenswrapper[4821]: I1125 11:30:03.112767 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401170-gwq95" Nov 25 11:30:03 crc kubenswrapper[4821]: I1125 11:30:03.246520 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/47cd3f86-f784-494c-b959-8125ea191612-secret-volume\") pod \"47cd3f86-f784-494c-b959-8125ea191612\" (UID: \"47cd3f86-f784-494c-b959-8125ea191612\") " Nov 25 11:30:03 crc kubenswrapper[4821]: I1125 11:30:03.246880 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fs2vw\" (UniqueName: \"kubernetes.io/projected/47cd3f86-f784-494c-b959-8125ea191612-kube-api-access-fs2vw\") pod \"47cd3f86-f784-494c-b959-8125ea191612\" (UID: \"47cd3f86-f784-494c-b959-8125ea191612\") " Nov 25 11:30:03 crc kubenswrapper[4821]: I1125 11:30:03.246951 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/47cd3f86-f784-494c-b959-8125ea191612-config-volume\") pod \"47cd3f86-f784-494c-b959-8125ea191612\" (UID: \"47cd3f86-f784-494c-b959-8125ea191612\") " Nov 25 11:30:03 crc kubenswrapper[4821]: I1125 11:30:03.247642 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/47cd3f86-f784-494c-b959-8125ea191612-config-volume" (OuterVolumeSpecName: "config-volume") pod "47cd3f86-f784-494c-b959-8125ea191612" (UID: "47cd3f86-f784-494c-b959-8125ea191612"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:30:03 crc kubenswrapper[4821]: I1125 11:30:03.247928 4821 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/47cd3f86-f784-494c-b959-8125ea191612-config-volume\") on node \"crc\" DevicePath \"\"" Nov 25 11:30:03 crc kubenswrapper[4821]: I1125 11:30:03.252299 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47cd3f86-f784-494c-b959-8125ea191612-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "47cd3f86-f784-494c-b959-8125ea191612" (UID: "47cd3f86-f784-494c-b959-8125ea191612"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:30:03 crc kubenswrapper[4821]: I1125 11:30:03.252525 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47cd3f86-f784-494c-b959-8125ea191612-kube-api-access-fs2vw" (OuterVolumeSpecName: "kube-api-access-fs2vw") pod "47cd3f86-f784-494c-b959-8125ea191612" (UID: "47cd3f86-f784-494c-b959-8125ea191612"). InnerVolumeSpecName "kube-api-access-fs2vw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:30:03 crc kubenswrapper[4821]: I1125 11:30:03.350023 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fs2vw\" (UniqueName: \"kubernetes.io/projected/47cd3f86-f784-494c-b959-8125ea191612-kube-api-access-fs2vw\") on node \"crc\" DevicePath \"\"" Nov 25 11:30:03 crc kubenswrapper[4821]: I1125 11:30:03.350265 4821 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/47cd3f86-f784-494c-b959-8125ea191612-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 25 11:30:03 crc kubenswrapper[4821]: I1125 11:30:03.800032 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401170-gwq95" event={"ID":"47cd3f86-f784-494c-b959-8125ea191612","Type":"ContainerDied","Data":"8ca707933bce8ec850551f310f7c21919fa41228ed189c6f61e1f1dedd371f9e"} Nov 25 11:30:03 crc kubenswrapper[4821]: I1125 11:30:03.800075 4821 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8ca707933bce8ec850551f310f7c21919fa41228ed189c6f61e1f1dedd371f9e" Nov 25 11:30:03 crc kubenswrapper[4821]: I1125 11:30:03.800084 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401170-gwq95" Nov 25 11:30:04 crc kubenswrapper[4821]: I1125 11:30:04.184018 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401125-5p66d"] Nov 25 11:30:04 crc kubenswrapper[4821]: I1125 11:30:04.192850 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401125-5p66d"] Nov 25 11:30:06 crc kubenswrapper[4821]: I1125 11:30:06.316229 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="52eea54a-9fe6-40ec-9354-8937bcc5a924" path="/var/lib/kubelet/pods/52eea54a-9fe6-40ec-9354-8937bcc5a924/volumes" Nov 25 11:30:07 crc kubenswrapper[4821]: I1125 11:30:07.834912 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-529qb" event={"ID":"410373d3-842e-43b3-a2d8-2293c543a4ac","Type":"ContainerStarted","Data":"6068d18680dbf361de7600b3c81dd361b7999f417a344cf6c6c43dada141bb7e"} Nov 25 11:30:08 crc kubenswrapper[4821]: I1125 11:30:08.846137 4821 generic.go:334] "Generic (PLEG): container finished" podID="410373d3-842e-43b3-a2d8-2293c543a4ac" containerID="6068d18680dbf361de7600b3c81dd361b7999f417a344cf6c6c43dada141bb7e" exitCode=0 Nov 25 11:30:08 crc kubenswrapper[4821]: I1125 11:30:08.846224 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-529qb" event={"ID":"410373d3-842e-43b3-a2d8-2293c543a4ac","Type":"ContainerDied","Data":"6068d18680dbf361de7600b3c81dd361b7999f417a344cf6c6c43dada141bb7e"} Nov 25 11:30:09 crc kubenswrapper[4821]: I1125 11:30:09.856204 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-529qb" event={"ID":"410373d3-842e-43b3-a2d8-2293c543a4ac","Type":"ContainerStarted","Data":"a0edfe30d3655726b065064a52f5a11347187aff07964cded993dc4fec1d7fdc"} Nov 25 11:30:09 crc kubenswrapper[4821]: I1125 11:30:09.878414 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-529qb" podStartSLOduration=3.412498005 podStartE2EDuration="10.87839418s" podCreationTimestamp="2025-11-25 11:29:59 +0000 UTC" firstStartedPulling="2025-11-25 11:30:01.783747504 +0000 UTC m=+3472.320067351" lastFinishedPulling="2025-11-25 11:30:09.249643689 +0000 UTC m=+3479.785963526" observedRunningTime="2025-11-25 11:30:09.870917778 +0000 UTC m=+3480.407237625" watchObservedRunningTime="2025-11-25 11:30:09.87839418 +0000 UTC m=+3480.414714017" Nov 25 11:30:10 crc kubenswrapper[4821]: I1125 11:30:10.147076 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-529qb" Nov 25 11:30:10 crc kubenswrapper[4821]: I1125 11:30:10.147786 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-529qb" Nov 25 11:30:11 crc kubenswrapper[4821]: I1125 11:30:11.195480 4821 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-529qb" podUID="410373d3-842e-43b3-a2d8-2293c543a4ac" containerName="registry-server" probeResult="failure" output=< Nov 25 11:30:11 crc kubenswrapper[4821]: timeout: failed to connect service ":50051" within 1s Nov 25 11:30:11 crc kubenswrapper[4821]: > Nov 25 11:30:17 crc kubenswrapper[4821]: I1125 11:30:17.375214 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-9zwfm/must-gather-kmj2h"] Nov 25 11:30:17 crc kubenswrapper[4821]: E1125 11:30:17.376078 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47cd3f86-f784-494c-b959-8125ea191612" containerName="collect-profiles" Nov 25 11:30:17 crc kubenswrapper[4821]: I1125 11:30:17.376091 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="47cd3f86-f784-494c-b959-8125ea191612" containerName="collect-profiles" Nov 25 11:30:17 crc kubenswrapper[4821]: I1125 11:30:17.376287 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="47cd3f86-f784-494c-b959-8125ea191612" containerName="collect-profiles" Nov 25 11:30:17 crc kubenswrapper[4821]: I1125 11:30:17.377248 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9zwfm/must-gather-kmj2h" Nov 25 11:30:17 crc kubenswrapper[4821]: I1125 11:30:17.381132 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-9zwfm"/"kube-root-ca.crt" Nov 25 11:30:17 crc kubenswrapper[4821]: I1125 11:30:17.381281 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-9zwfm"/"default-dockercfg-vrc6q" Nov 25 11:30:17 crc kubenswrapper[4821]: I1125 11:30:17.383849 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-9zwfm"/"openshift-service-ca.crt" Nov 25 11:30:17 crc kubenswrapper[4821]: I1125 11:30:17.393393 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-9zwfm/must-gather-kmj2h"] Nov 25 11:30:17 crc kubenswrapper[4821]: I1125 11:30:17.520994 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wghn6\" (UniqueName: \"kubernetes.io/projected/d93b17f1-e742-45a7-a5ec-141e52a382c2-kube-api-access-wghn6\") pod \"must-gather-kmj2h\" (UID: \"d93b17f1-e742-45a7-a5ec-141e52a382c2\") " pod="openshift-must-gather-9zwfm/must-gather-kmj2h" Nov 25 11:30:17 crc kubenswrapper[4821]: I1125 11:30:17.521083 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d93b17f1-e742-45a7-a5ec-141e52a382c2-must-gather-output\") pod \"must-gather-kmj2h\" (UID: \"d93b17f1-e742-45a7-a5ec-141e52a382c2\") " pod="openshift-must-gather-9zwfm/must-gather-kmj2h" Nov 25 11:30:17 crc kubenswrapper[4821]: I1125 11:30:17.622724 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d93b17f1-e742-45a7-a5ec-141e52a382c2-must-gather-output\") pod \"must-gather-kmj2h\" (UID: \"d93b17f1-e742-45a7-a5ec-141e52a382c2\") " pod="openshift-must-gather-9zwfm/must-gather-kmj2h" Nov 25 11:30:17 crc kubenswrapper[4821]: I1125 11:30:17.622896 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wghn6\" (UniqueName: \"kubernetes.io/projected/d93b17f1-e742-45a7-a5ec-141e52a382c2-kube-api-access-wghn6\") pod \"must-gather-kmj2h\" (UID: \"d93b17f1-e742-45a7-a5ec-141e52a382c2\") " pod="openshift-must-gather-9zwfm/must-gather-kmj2h" Nov 25 11:30:17 crc kubenswrapper[4821]: I1125 11:30:17.623623 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d93b17f1-e742-45a7-a5ec-141e52a382c2-must-gather-output\") pod \"must-gather-kmj2h\" (UID: \"d93b17f1-e742-45a7-a5ec-141e52a382c2\") " pod="openshift-must-gather-9zwfm/must-gather-kmj2h" Nov 25 11:30:17 crc kubenswrapper[4821]: I1125 11:30:17.646141 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wghn6\" (UniqueName: \"kubernetes.io/projected/d93b17f1-e742-45a7-a5ec-141e52a382c2-kube-api-access-wghn6\") pod \"must-gather-kmj2h\" (UID: \"d93b17f1-e742-45a7-a5ec-141e52a382c2\") " pod="openshift-must-gather-9zwfm/must-gather-kmj2h" Nov 25 11:30:17 crc kubenswrapper[4821]: I1125 11:30:17.707474 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9zwfm/must-gather-kmj2h" Nov 25 11:30:18 crc kubenswrapper[4821]: I1125 11:30:18.275848 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-9zwfm/must-gather-kmj2h"] Nov 25 11:30:18 crc kubenswrapper[4821]: I1125 11:30:18.957190 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9zwfm/must-gather-kmj2h" event={"ID":"d93b17f1-e742-45a7-a5ec-141e52a382c2","Type":"ContainerStarted","Data":"1e5fb6421076343d0db34dc4fe5f0f37437958c671c425a8e8f0ccf4e1eb0740"} Nov 25 11:30:20 crc kubenswrapper[4821]: I1125 11:30:20.203078 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-529qb" Nov 25 11:30:20 crc kubenswrapper[4821]: I1125 11:30:20.256592 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-529qb" Nov 25 11:30:20 crc kubenswrapper[4821]: I1125 11:30:20.364739 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-529qb"] Nov 25 11:30:20 crc kubenswrapper[4821]: I1125 11:30:20.468284 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-g4r76"] Nov 25 11:30:20 crc kubenswrapper[4821]: I1125 11:30:20.468580 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-g4r76" podUID="268c8582-b9a0-4e22-ac84-e47a6fa3f284" containerName="registry-server" containerID="cri-o://d62769432b7b327ab2021a111e56df775eb219b3e265c3391fa560d009de6411" gracePeriod=2 Nov 25 11:30:21 crc kubenswrapper[4821]: I1125 11:30:20.999833 4821 generic.go:334] "Generic (PLEG): container finished" podID="268c8582-b9a0-4e22-ac84-e47a6fa3f284" containerID="d62769432b7b327ab2021a111e56df775eb219b3e265c3391fa560d009de6411" exitCode=0 Nov 25 11:30:21 crc kubenswrapper[4821]: I1125 11:30:20.999921 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g4r76" event={"ID":"268c8582-b9a0-4e22-ac84-e47a6fa3f284","Type":"ContainerDied","Data":"d62769432b7b327ab2021a111e56df775eb219b3e265c3391fa560d009de6411"} Nov 25 11:30:21 crc kubenswrapper[4821]: I1125 11:30:21.567258 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g4r76" Nov 25 11:30:21 crc kubenswrapper[4821]: I1125 11:30:21.614953 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/268c8582-b9a0-4e22-ac84-e47a6fa3f284-utilities\") pod \"268c8582-b9a0-4e22-ac84-e47a6fa3f284\" (UID: \"268c8582-b9a0-4e22-ac84-e47a6fa3f284\") " Nov 25 11:30:21 crc kubenswrapper[4821]: I1125 11:30:21.615148 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/268c8582-b9a0-4e22-ac84-e47a6fa3f284-catalog-content\") pod \"268c8582-b9a0-4e22-ac84-e47a6fa3f284\" (UID: \"268c8582-b9a0-4e22-ac84-e47a6fa3f284\") " Nov 25 11:30:21 crc kubenswrapper[4821]: I1125 11:30:21.615363 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g7lnv\" (UniqueName: \"kubernetes.io/projected/268c8582-b9a0-4e22-ac84-e47a6fa3f284-kube-api-access-g7lnv\") pod \"268c8582-b9a0-4e22-ac84-e47a6fa3f284\" (UID: \"268c8582-b9a0-4e22-ac84-e47a6fa3f284\") " Nov 25 11:30:21 crc kubenswrapper[4821]: I1125 11:30:21.618332 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/268c8582-b9a0-4e22-ac84-e47a6fa3f284-utilities" (OuterVolumeSpecName: "utilities") pod "268c8582-b9a0-4e22-ac84-e47a6fa3f284" (UID: "268c8582-b9a0-4e22-ac84-e47a6fa3f284"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:30:21 crc kubenswrapper[4821]: I1125 11:30:21.644516 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/268c8582-b9a0-4e22-ac84-e47a6fa3f284-kube-api-access-g7lnv" (OuterVolumeSpecName: "kube-api-access-g7lnv") pod "268c8582-b9a0-4e22-ac84-e47a6fa3f284" (UID: "268c8582-b9a0-4e22-ac84-e47a6fa3f284"). InnerVolumeSpecName "kube-api-access-g7lnv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:30:21 crc kubenswrapper[4821]: I1125 11:30:21.717720 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g7lnv\" (UniqueName: \"kubernetes.io/projected/268c8582-b9a0-4e22-ac84-e47a6fa3f284-kube-api-access-g7lnv\") on node \"crc\" DevicePath \"\"" Nov 25 11:30:21 crc kubenswrapper[4821]: I1125 11:30:21.717757 4821 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/268c8582-b9a0-4e22-ac84-e47a6fa3f284-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 11:30:21 crc kubenswrapper[4821]: I1125 11:30:21.734832 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/268c8582-b9a0-4e22-ac84-e47a6fa3f284-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "268c8582-b9a0-4e22-ac84-e47a6fa3f284" (UID: "268c8582-b9a0-4e22-ac84-e47a6fa3f284"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:30:21 crc kubenswrapper[4821]: I1125 11:30:21.819013 4821 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/268c8582-b9a0-4e22-ac84-e47a6fa3f284-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 11:30:22 crc kubenswrapper[4821]: I1125 11:30:22.019568 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g4r76" Nov 25 11:30:22 crc kubenswrapper[4821]: I1125 11:30:22.022246 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g4r76" event={"ID":"268c8582-b9a0-4e22-ac84-e47a6fa3f284","Type":"ContainerDied","Data":"d3ce50fc235bad9afb3ca250067a103239ab31c884315d7903686e5831ae6e04"} Nov 25 11:30:22 crc kubenswrapper[4821]: I1125 11:30:22.022327 4821 scope.go:117] "RemoveContainer" containerID="d62769432b7b327ab2021a111e56df775eb219b3e265c3391fa560d009de6411" Nov 25 11:30:22 crc kubenswrapper[4821]: I1125 11:30:22.066660 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-g4r76"] Nov 25 11:30:22 crc kubenswrapper[4821]: I1125 11:30:22.080284 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-g4r76"] Nov 25 11:30:22 crc kubenswrapper[4821]: I1125 11:30:22.131935 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="268c8582-b9a0-4e22-ac84-e47a6fa3f284" path="/var/lib/kubelet/pods/268c8582-b9a0-4e22-ac84-e47a6fa3f284/volumes" Nov 25 11:30:24 crc kubenswrapper[4821]: I1125 11:30:24.447358 4821 scope.go:117] "RemoveContainer" containerID="e9d2dc3ae125b4c3a0a7f0a47570f0351d7988bd73e7f68b65fddb861667f227" Nov 25 11:30:24 crc kubenswrapper[4821]: I1125 11:30:24.483599 4821 scope.go:117] "RemoveContainer" containerID="1cc00a4c480de98e11c9fcbe09399ddb90d9a63af828f33af5ae299c987ddc6a" Nov 25 11:30:25 crc kubenswrapper[4821]: I1125 11:30:25.049049 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9zwfm/must-gather-kmj2h" event={"ID":"d93b17f1-e742-45a7-a5ec-141e52a382c2","Type":"ContainerStarted","Data":"81bc3cd2f14c92f9509d803c926854123e6f0f440a6479ebd4c24da154176cae"} Nov 25 11:30:25 crc kubenswrapper[4821]: I1125 11:30:25.784416 4821 scope.go:117] "RemoveContainer" containerID="e6513bd7ff68841deae35d9fd9ce21fe69b3ac3a094be4c9a379dbe841e020dd" Nov 25 11:30:26 crc kubenswrapper[4821]: I1125 11:30:26.063027 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9zwfm/must-gather-kmj2h" event={"ID":"d93b17f1-e742-45a7-a5ec-141e52a382c2","Type":"ContainerStarted","Data":"3e1b6937222640bbbef56eb7665a33f444c4d41c3ebe1880942f498bf2797891"} Nov 25 11:30:26 crc kubenswrapper[4821]: I1125 11:30:26.082881 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-9zwfm/must-gather-kmj2h" podStartSLOduration=2.697317614 podStartE2EDuration="9.082863152s" podCreationTimestamp="2025-11-25 11:30:17 +0000 UTC" firstStartedPulling="2025-11-25 11:30:18.290509529 +0000 UTC m=+3488.826829376" lastFinishedPulling="2025-11-25 11:30:24.676055067 +0000 UTC m=+3495.212374914" observedRunningTime="2025-11-25 11:30:26.077991744 +0000 UTC m=+3496.614311581" watchObservedRunningTime="2025-11-25 11:30:26.082863152 +0000 UTC m=+3496.619182999" Nov 25 11:30:28 crc kubenswrapper[4821]: I1125 11:30:28.571451 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-9zwfm/crc-debug-gl6hc"] Nov 25 11:30:28 crc kubenswrapper[4821]: E1125 11:30:28.572385 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="268c8582-b9a0-4e22-ac84-e47a6fa3f284" containerName="registry-server" Nov 25 11:30:28 crc kubenswrapper[4821]: I1125 11:30:28.572402 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="268c8582-b9a0-4e22-ac84-e47a6fa3f284" containerName="registry-server" Nov 25 11:30:28 crc kubenswrapper[4821]: E1125 11:30:28.572422 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="268c8582-b9a0-4e22-ac84-e47a6fa3f284" containerName="extract-utilities" Nov 25 11:30:28 crc kubenswrapper[4821]: I1125 11:30:28.572430 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="268c8582-b9a0-4e22-ac84-e47a6fa3f284" containerName="extract-utilities" Nov 25 11:30:28 crc kubenswrapper[4821]: E1125 11:30:28.572473 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="268c8582-b9a0-4e22-ac84-e47a6fa3f284" containerName="extract-content" Nov 25 11:30:28 crc kubenswrapper[4821]: I1125 11:30:28.572482 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="268c8582-b9a0-4e22-ac84-e47a6fa3f284" containerName="extract-content" Nov 25 11:30:28 crc kubenswrapper[4821]: I1125 11:30:28.572726 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="268c8582-b9a0-4e22-ac84-e47a6fa3f284" containerName="registry-server" Nov 25 11:30:28 crc kubenswrapper[4821]: I1125 11:30:28.573563 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9zwfm/crc-debug-gl6hc" Nov 25 11:30:28 crc kubenswrapper[4821]: I1125 11:30:28.667819 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jk8f4\" (UniqueName: \"kubernetes.io/projected/1027f991-c500-4846-8756-4b423215ba44-kube-api-access-jk8f4\") pod \"crc-debug-gl6hc\" (UID: \"1027f991-c500-4846-8756-4b423215ba44\") " pod="openshift-must-gather-9zwfm/crc-debug-gl6hc" Nov 25 11:30:28 crc kubenswrapper[4821]: I1125 11:30:28.667982 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1027f991-c500-4846-8756-4b423215ba44-host\") pod \"crc-debug-gl6hc\" (UID: \"1027f991-c500-4846-8756-4b423215ba44\") " pod="openshift-must-gather-9zwfm/crc-debug-gl6hc" Nov 25 11:30:28 crc kubenswrapper[4821]: I1125 11:30:28.769417 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jk8f4\" (UniqueName: \"kubernetes.io/projected/1027f991-c500-4846-8756-4b423215ba44-kube-api-access-jk8f4\") pod \"crc-debug-gl6hc\" (UID: \"1027f991-c500-4846-8756-4b423215ba44\") " pod="openshift-must-gather-9zwfm/crc-debug-gl6hc" Nov 25 11:30:28 crc kubenswrapper[4821]: I1125 11:30:28.769485 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1027f991-c500-4846-8756-4b423215ba44-host\") pod \"crc-debug-gl6hc\" (UID: \"1027f991-c500-4846-8756-4b423215ba44\") " pod="openshift-must-gather-9zwfm/crc-debug-gl6hc" Nov 25 11:30:28 crc kubenswrapper[4821]: I1125 11:30:28.769584 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1027f991-c500-4846-8756-4b423215ba44-host\") pod \"crc-debug-gl6hc\" (UID: \"1027f991-c500-4846-8756-4b423215ba44\") " pod="openshift-must-gather-9zwfm/crc-debug-gl6hc" Nov 25 11:30:28 crc kubenswrapper[4821]: I1125 11:30:28.788691 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jk8f4\" (UniqueName: \"kubernetes.io/projected/1027f991-c500-4846-8756-4b423215ba44-kube-api-access-jk8f4\") pod \"crc-debug-gl6hc\" (UID: \"1027f991-c500-4846-8756-4b423215ba44\") " pod="openshift-must-gather-9zwfm/crc-debug-gl6hc" Nov 25 11:30:28 crc kubenswrapper[4821]: I1125 11:30:28.893685 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9zwfm/crc-debug-gl6hc" Nov 25 11:30:29 crc kubenswrapper[4821]: I1125 11:30:29.090791 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9zwfm/crc-debug-gl6hc" event={"ID":"1027f991-c500-4846-8756-4b423215ba44","Type":"ContainerStarted","Data":"68bc4451cfb9e937d51939da8fc8396674934f55b44301190fc9b590d04512da"} Nov 25 11:30:34 crc kubenswrapper[4821]: I1125 11:30:34.715470 4821 patch_prober.go:28] interesting pod/machine-config-daemon-2krbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:30:34 crc kubenswrapper[4821]: I1125 11:30:34.716579 4821 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:30:44 crc kubenswrapper[4821]: I1125 11:30:44.228182 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9zwfm/crc-debug-gl6hc" event={"ID":"1027f991-c500-4846-8756-4b423215ba44","Type":"ContainerStarted","Data":"3f75c1323e9159aa66657b856a5905d399ef8b6c10e67b18c77e54c8f6ef5ceb"} Nov 25 11:30:44 crc kubenswrapper[4821]: I1125 11:30:44.245594 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-9zwfm/crc-debug-gl6hc" podStartSLOduration=2.14649977 podStartE2EDuration="16.245561341s" podCreationTimestamp="2025-11-25 11:30:28 +0000 UTC" firstStartedPulling="2025-11-25 11:30:28.931729215 +0000 UTC m=+3499.468049062" lastFinishedPulling="2025-11-25 11:30:43.030790786 +0000 UTC m=+3513.567110633" observedRunningTime="2025-11-25 11:30:44.240214572 +0000 UTC m=+3514.776534439" watchObservedRunningTime="2025-11-25 11:30:44.245561341 +0000 UTC m=+3514.781881188" Nov 25 11:31:04 crc kubenswrapper[4821]: I1125 11:31:04.715284 4821 patch_prober.go:28] interesting pod/machine-config-daemon-2krbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:31:04 crc kubenswrapper[4821]: I1125 11:31:04.715919 4821 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:31:25 crc kubenswrapper[4821]: I1125 11:31:25.587149 4821 generic.go:334] "Generic (PLEG): container finished" podID="1027f991-c500-4846-8756-4b423215ba44" containerID="3f75c1323e9159aa66657b856a5905d399ef8b6c10e67b18c77e54c8f6ef5ceb" exitCode=0 Nov 25 11:31:25 crc kubenswrapper[4821]: I1125 11:31:25.587387 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9zwfm/crc-debug-gl6hc" event={"ID":"1027f991-c500-4846-8756-4b423215ba44","Type":"ContainerDied","Data":"3f75c1323e9159aa66657b856a5905d399ef8b6c10e67b18c77e54c8f6ef5ceb"} Nov 25 11:31:26 crc kubenswrapper[4821]: I1125 11:31:26.700051 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9zwfm/crc-debug-gl6hc" Nov 25 11:31:26 crc kubenswrapper[4821]: I1125 11:31:26.742673 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-9zwfm/crc-debug-gl6hc"] Nov 25 11:31:26 crc kubenswrapper[4821]: I1125 11:31:26.754647 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-9zwfm/crc-debug-gl6hc"] Nov 25 11:31:26 crc kubenswrapper[4821]: I1125 11:31:26.893935 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jk8f4\" (UniqueName: \"kubernetes.io/projected/1027f991-c500-4846-8756-4b423215ba44-kube-api-access-jk8f4\") pod \"1027f991-c500-4846-8756-4b423215ba44\" (UID: \"1027f991-c500-4846-8756-4b423215ba44\") " Nov 25 11:31:26 crc kubenswrapper[4821]: I1125 11:31:26.894094 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1027f991-c500-4846-8756-4b423215ba44-host\") pod \"1027f991-c500-4846-8756-4b423215ba44\" (UID: \"1027f991-c500-4846-8756-4b423215ba44\") " Nov 25 11:31:26 crc kubenswrapper[4821]: I1125 11:31:26.894404 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1027f991-c500-4846-8756-4b423215ba44-host" (OuterVolumeSpecName: "host") pod "1027f991-c500-4846-8756-4b423215ba44" (UID: "1027f991-c500-4846-8756-4b423215ba44"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 11:31:26 crc kubenswrapper[4821]: I1125 11:31:26.894821 4821 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1027f991-c500-4846-8756-4b423215ba44-host\") on node \"crc\" DevicePath \"\"" Nov 25 11:31:26 crc kubenswrapper[4821]: I1125 11:31:26.903399 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1027f991-c500-4846-8756-4b423215ba44-kube-api-access-jk8f4" (OuterVolumeSpecName: "kube-api-access-jk8f4") pod "1027f991-c500-4846-8756-4b423215ba44" (UID: "1027f991-c500-4846-8756-4b423215ba44"). InnerVolumeSpecName "kube-api-access-jk8f4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:31:26 crc kubenswrapper[4821]: I1125 11:31:26.995686 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jk8f4\" (UniqueName: \"kubernetes.io/projected/1027f991-c500-4846-8756-4b423215ba44-kube-api-access-jk8f4\") on node \"crc\" DevicePath \"\"" Nov 25 11:31:27 crc kubenswrapper[4821]: I1125 11:31:27.605481 4821 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="68bc4451cfb9e937d51939da8fc8396674934f55b44301190fc9b590d04512da" Nov 25 11:31:27 crc kubenswrapper[4821]: I1125 11:31:27.605582 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9zwfm/crc-debug-gl6hc" Nov 25 11:31:27 crc kubenswrapper[4821]: I1125 11:31:27.934516 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-9zwfm/crc-debug-7qvhn"] Nov 25 11:31:27 crc kubenswrapper[4821]: E1125 11:31:27.935252 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1027f991-c500-4846-8756-4b423215ba44" containerName="container-00" Nov 25 11:31:27 crc kubenswrapper[4821]: I1125 11:31:27.935268 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="1027f991-c500-4846-8756-4b423215ba44" containerName="container-00" Nov 25 11:31:27 crc kubenswrapper[4821]: I1125 11:31:27.935530 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="1027f991-c500-4846-8756-4b423215ba44" containerName="container-00" Nov 25 11:31:27 crc kubenswrapper[4821]: I1125 11:31:27.936125 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9zwfm/crc-debug-7qvhn" Nov 25 11:31:28 crc kubenswrapper[4821]: I1125 11:31:28.013081 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s85h4\" (UniqueName: \"kubernetes.io/projected/84ae1948-050c-4d83-88be-4b93040c7e06-kube-api-access-s85h4\") pod \"crc-debug-7qvhn\" (UID: \"84ae1948-050c-4d83-88be-4b93040c7e06\") " pod="openshift-must-gather-9zwfm/crc-debug-7qvhn" Nov 25 11:31:28 crc kubenswrapper[4821]: I1125 11:31:28.013154 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/84ae1948-050c-4d83-88be-4b93040c7e06-host\") pod \"crc-debug-7qvhn\" (UID: \"84ae1948-050c-4d83-88be-4b93040c7e06\") " pod="openshift-must-gather-9zwfm/crc-debug-7qvhn" Nov 25 11:31:28 crc kubenswrapper[4821]: I1125 11:31:28.114961 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s85h4\" (UniqueName: \"kubernetes.io/projected/84ae1948-050c-4d83-88be-4b93040c7e06-kube-api-access-s85h4\") pod \"crc-debug-7qvhn\" (UID: \"84ae1948-050c-4d83-88be-4b93040c7e06\") " pod="openshift-must-gather-9zwfm/crc-debug-7qvhn" Nov 25 11:31:28 crc kubenswrapper[4821]: I1125 11:31:28.115058 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/84ae1948-050c-4d83-88be-4b93040c7e06-host\") pod \"crc-debug-7qvhn\" (UID: \"84ae1948-050c-4d83-88be-4b93040c7e06\") " pod="openshift-must-gather-9zwfm/crc-debug-7qvhn" Nov 25 11:31:28 crc kubenswrapper[4821]: I1125 11:31:28.115251 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/84ae1948-050c-4d83-88be-4b93040c7e06-host\") pod \"crc-debug-7qvhn\" (UID: \"84ae1948-050c-4d83-88be-4b93040c7e06\") " pod="openshift-must-gather-9zwfm/crc-debug-7qvhn" Nov 25 11:31:28 crc kubenswrapper[4821]: I1125 11:31:28.125671 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1027f991-c500-4846-8756-4b423215ba44" path="/var/lib/kubelet/pods/1027f991-c500-4846-8756-4b423215ba44/volumes" Nov 25 11:31:28 crc kubenswrapper[4821]: I1125 11:31:28.146044 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s85h4\" (UniqueName: \"kubernetes.io/projected/84ae1948-050c-4d83-88be-4b93040c7e06-kube-api-access-s85h4\") pod \"crc-debug-7qvhn\" (UID: \"84ae1948-050c-4d83-88be-4b93040c7e06\") " pod="openshift-must-gather-9zwfm/crc-debug-7qvhn" Nov 25 11:31:28 crc kubenswrapper[4821]: I1125 11:31:28.255935 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9zwfm/crc-debug-7qvhn" Nov 25 11:31:28 crc kubenswrapper[4821]: I1125 11:31:28.615374 4821 generic.go:334] "Generic (PLEG): container finished" podID="84ae1948-050c-4d83-88be-4b93040c7e06" containerID="8cb2388580979ba0200b7f59b0fa9d5a5e87f9809e5c4b05c768c72baa961abb" exitCode=0 Nov 25 11:31:28 crc kubenswrapper[4821]: I1125 11:31:28.615506 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9zwfm/crc-debug-7qvhn" event={"ID":"84ae1948-050c-4d83-88be-4b93040c7e06","Type":"ContainerDied","Data":"8cb2388580979ba0200b7f59b0fa9d5a5e87f9809e5c4b05c768c72baa961abb"} Nov 25 11:31:28 crc kubenswrapper[4821]: I1125 11:31:28.615775 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9zwfm/crc-debug-7qvhn" event={"ID":"84ae1948-050c-4d83-88be-4b93040c7e06","Type":"ContainerStarted","Data":"5937971489ac417ee87834ac6c3eaf4ffa176b29d3aac49eccf13c7b55814331"} Nov 25 11:31:29 crc kubenswrapper[4821]: I1125 11:31:29.073257 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-9zwfm/crc-debug-7qvhn"] Nov 25 11:31:29 crc kubenswrapper[4821]: I1125 11:31:29.081214 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-9zwfm/crc-debug-7qvhn"] Nov 25 11:31:29 crc kubenswrapper[4821]: I1125 11:31:29.737732 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9zwfm/crc-debug-7qvhn" Nov 25 11:31:29 crc kubenswrapper[4821]: I1125 11:31:29.848735 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s85h4\" (UniqueName: \"kubernetes.io/projected/84ae1948-050c-4d83-88be-4b93040c7e06-kube-api-access-s85h4\") pod \"84ae1948-050c-4d83-88be-4b93040c7e06\" (UID: \"84ae1948-050c-4d83-88be-4b93040c7e06\") " Nov 25 11:31:29 crc kubenswrapper[4821]: I1125 11:31:29.848991 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/84ae1948-050c-4d83-88be-4b93040c7e06-host\") pod \"84ae1948-050c-4d83-88be-4b93040c7e06\" (UID: \"84ae1948-050c-4d83-88be-4b93040c7e06\") " Nov 25 11:31:29 crc kubenswrapper[4821]: I1125 11:31:29.849645 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/84ae1948-050c-4d83-88be-4b93040c7e06-host" (OuterVolumeSpecName: "host") pod "84ae1948-050c-4d83-88be-4b93040c7e06" (UID: "84ae1948-050c-4d83-88be-4b93040c7e06"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 11:31:29 crc kubenswrapper[4821]: I1125 11:31:29.855118 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84ae1948-050c-4d83-88be-4b93040c7e06-kube-api-access-s85h4" (OuterVolumeSpecName: "kube-api-access-s85h4") pod "84ae1948-050c-4d83-88be-4b93040c7e06" (UID: "84ae1948-050c-4d83-88be-4b93040c7e06"). InnerVolumeSpecName "kube-api-access-s85h4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:31:29 crc kubenswrapper[4821]: I1125 11:31:29.951340 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s85h4\" (UniqueName: \"kubernetes.io/projected/84ae1948-050c-4d83-88be-4b93040c7e06-kube-api-access-s85h4\") on node \"crc\" DevicePath \"\"" Nov 25 11:31:29 crc kubenswrapper[4821]: I1125 11:31:29.951385 4821 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/84ae1948-050c-4d83-88be-4b93040c7e06-host\") on node \"crc\" DevicePath \"\"" Nov 25 11:31:30 crc kubenswrapper[4821]: I1125 11:31:30.127811 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84ae1948-050c-4d83-88be-4b93040c7e06" path="/var/lib/kubelet/pods/84ae1948-050c-4d83-88be-4b93040c7e06/volumes" Nov 25 11:31:30 crc kubenswrapper[4821]: I1125 11:31:30.260944 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-9zwfm/crc-debug-jpcmw"] Nov 25 11:31:30 crc kubenswrapper[4821]: E1125 11:31:30.261681 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84ae1948-050c-4d83-88be-4b93040c7e06" containerName="container-00" Nov 25 11:31:30 crc kubenswrapper[4821]: I1125 11:31:30.261719 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="84ae1948-050c-4d83-88be-4b93040c7e06" containerName="container-00" Nov 25 11:31:30 crc kubenswrapper[4821]: I1125 11:31:30.261954 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="84ae1948-050c-4d83-88be-4b93040c7e06" containerName="container-00" Nov 25 11:31:30 crc kubenswrapper[4821]: I1125 11:31:30.262834 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9zwfm/crc-debug-jpcmw" Nov 25 11:31:30 crc kubenswrapper[4821]: I1125 11:31:30.461016 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bhp97\" (UniqueName: \"kubernetes.io/projected/cdd5fc1c-bf87-4bb1-9834-93e2bef9e773-kube-api-access-bhp97\") pod \"crc-debug-jpcmw\" (UID: \"cdd5fc1c-bf87-4bb1-9834-93e2bef9e773\") " pod="openshift-must-gather-9zwfm/crc-debug-jpcmw" Nov 25 11:31:30 crc kubenswrapper[4821]: I1125 11:31:30.461358 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cdd5fc1c-bf87-4bb1-9834-93e2bef9e773-host\") pod \"crc-debug-jpcmw\" (UID: \"cdd5fc1c-bf87-4bb1-9834-93e2bef9e773\") " pod="openshift-must-gather-9zwfm/crc-debug-jpcmw" Nov 25 11:31:30 crc kubenswrapper[4821]: I1125 11:31:30.564093 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bhp97\" (UniqueName: \"kubernetes.io/projected/cdd5fc1c-bf87-4bb1-9834-93e2bef9e773-kube-api-access-bhp97\") pod \"crc-debug-jpcmw\" (UID: \"cdd5fc1c-bf87-4bb1-9834-93e2bef9e773\") " pod="openshift-must-gather-9zwfm/crc-debug-jpcmw" Nov 25 11:31:30 crc kubenswrapper[4821]: I1125 11:31:30.564190 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cdd5fc1c-bf87-4bb1-9834-93e2bef9e773-host\") pod \"crc-debug-jpcmw\" (UID: \"cdd5fc1c-bf87-4bb1-9834-93e2bef9e773\") " pod="openshift-must-gather-9zwfm/crc-debug-jpcmw" Nov 25 11:31:30 crc kubenswrapper[4821]: I1125 11:31:30.564402 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cdd5fc1c-bf87-4bb1-9834-93e2bef9e773-host\") pod \"crc-debug-jpcmw\" (UID: \"cdd5fc1c-bf87-4bb1-9834-93e2bef9e773\") " pod="openshift-must-gather-9zwfm/crc-debug-jpcmw" Nov 25 11:31:30 crc kubenswrapper[4821]: I1125 11:31:30.584931 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bhp97\" (UniqueName: \"kubernetes.io/projected/cdd5fc1c-bf87-4bb1-9834-93e2bef9e773-kube-api-access-bhp97\") pod \"crc-debug-jpcmw\" (UID: \"cdd5fc1c-bf87-4bb1-9834-93e2bef9e773\") " pod="openshift-must-gather-9zwfm/crc-debug-jpcmw" Nov 25 11:31:30 crc kubenswrapper[4821]: I1125 11:31:30.636478 4821 scope.go:117] "RemoveContainer" containerID="8cb2388580979ba0200b7f59b0fa9d5a5e87f9809e5c4b05c768c72baa961abb" Nov 25 11:31:30 crc kubenswrapper[4821]: I1125 11:31:30.636514 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9zwfm/crc-debug-7qvhn" Nov 25 11:31:30 crc kubenswrapper[4821]: I1125 11:31:30.877473 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9zwfm/crc-debug-jpcmw" Nov 25 11:31:30 crc kubenswrapper[4821]: W1125 11:31:30.900258 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcdd5fc1c_bf87_4bb1_9834_93e2bef9e773.slice/crio-c8dad9614a25625cdc16b57aca56480de2483e99de9a1dbc16874a3d342d5a21 WatchSource:0}: Error finding container c8dad9614a25625cdc16b57aca56480de2483e99de9a1dbc16874a3d342d5a21: Status 404 returned error can't find the container with id c8dad9614a25625cdc16b57aca56480de2483e99de9a1dbc16874a3d342d5a21 Nov 25 11:31:31 crc kubenswrapper[4821]: I1125 11:31:31.646116 4821 generic.go:334] "Generic (PLEG): container finished" podID="cdd5fc1c-bf87-4bb1-9834-93e2bef9e773" containerID="cabe4e56553938dcc980b87e64c0cb9a68c9a0e3fa0b0d47fb7a0db653ae4d92" exitCode=0 Nov 25 11:31:31 crc kubenswrapper[4821]: I1125 11:31:31.646306 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9zwfm/crc-debug-jpcmw" event={"ID":"cdd5fc1c-bf87-4bb1-9834-93e2bef9e773","Type":"ContainerDied","Data":"cabe4e56553938dcc980b87e64c0cb9a68c9a0e3fa0b0d47fb7a0db653ae4d92"} Nov 25 11:31:31 crc kubenswrapper[4821]: I1125 11:31:31.646705 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9zwfm/crc-debug-jpcmw" event={"ID":"cdd5fc1c-bf87-4bb1-9834-93e2bef9e773","Type":"ContainerStarted","Data":"c8dad9614a25625cdc16b57aca56480de2483e99de9a1dbc16874a3d342d5a21"} Nov 25 11:31:31 crc kubenswrapper[4821]: I1125 11:31:31.684198 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-9zwfm/crc-debug-jpcmw"] Nov 25 11:31:31 crc kubenswrapper[4821]: I1125 11:31:31.691275 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-9zwfm/crc-debug-jpcmw"] Nov 25 11:31:32 crc kubenswrapper[4821]: I1125 11:31:32.768262 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9zwfm/crc-debug-jpcmw" Nov 25 11:31:32 crc kubenswrapper[4821]: I1125 11:31:32.905080 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bhp97\" (UniqueName: \"kubernetes.io/projected/cdd5fc1c-bf87-4bb1-9834-93e2bef9e773-kube-api-access-bhp97\") pod \"cdd5fc1c-bf87-4bb1-9834-93e2bef9e773\" (UID: \"cdd5fc1c-bf87-4bb1-9834-93e2bef9e773\") " Nov 25 11:31:32 crc kubenswrapper[4821]: I1125 11:31:32.905227 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cdd5fc1c-bf87-4bb1-9834-93e2bef9e773-host\") pod \"cdd5fc1c-bf87-4bb1-9834-93e2bef9e773\" (UID: \"cdd5fc1c-bf87-4bb1-9834-93e2bef9e773\") " Nov 25 11:31:32 crc kubenswrapper[4821]: I1125 11:31:32.905351 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cdd5fc1c-bf87-4bb1-9834-93e2bef9e773-host" (OuterVolumeSpecName: "host") pod "cdd5fc1c-bf87-4bb1-9834-93e2bef9e773" (UID: "cdd5fc1c-bf87-4bb1-9834-93e2bef9e773"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 11:31:32 crc kubenswrapper[4821]: I1125 11:31:32.912366 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cdd5fc1c-bf87-4bb1-9834-93e2bef9e773-kube-api-access-bhp97" (OuterVolumeSpecName: "kube-api-access-bhp97") pod "cdd5fc1c-bf87-4bb1-9834-93e2bef9e773" (UID: "cdd5fc1c-bf87-4bb1-9834-93e2bef9e773"). InnerVolumeSpecName "kube-api-access-bhp97". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:31:32 crc kubenswrapper[4821]: I1125 11:31:32.919002 4821 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cdd5fc1c-bf87-4bb1-9834-93e2bef9e773-host\") on node \"crc\" DevicePath \"\"" Nov 25 11:31:33 crc kubenswrapper[4821]: I1125 11:31:33.020617 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bhp97\" (UniqueName: \"kubernetes.io/projected/cdd5fc1c-bf87-4bb1-9834-93e2bef9e773-kube-api-access-bhp97\") on node \"crc\" DevicePath \"\"" Nov 25 11:31:33 crc kubenswrapper[4821]: I1125 11:31:33.667568 4821 scope.go:117] "RemoveContainer" containerID="cabe4e56553938dcc980b87e64c0cb9a68c9a0e3fa0b0d47fb7a0db653ae4d92" Nov 25 11:31:33 crc kubenswrapper[4821]: I1125 11:31:33.667741 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9zwfm/crc-debug-jpcmw" Nov 25 11:31:34 crc kubenswrapper[4821]: I1125 11:31:34.137534 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cdd5fc1c-bf87-4bb1-9834-93e2bef9e773" path="/var/lib/kubelet/pods/cdd5fc1c-bf87-4bb1-9834-93e2bef9e773/volumes" Nov 25 11:31:34 crc kubenswrapper[4821]: I1125 11:31:34.715132 4821 patch_prober.go:28] interesting pod/machine-config-daemon-2krbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:31:34 crc kubenswrapper[4821]: I1125 11:31:34.715724 4821 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:31:34 crc kubenswrapper[4821]: I1125 11:31:34.715844 4821 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" Nov 25 11:31:34 crc kubenswrapper[4821]: I1125 11:31:34.716675 4821 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"52d1a350addd03a9e1e92a65fe40ee197504bf6db7f5b0601ca8860f562470fb"} pod="openshift-machine-config-operator/machine-config-daemon-2krbf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 11:31:34 crc kubenswrapper[4821]: I1125 11:31:34.716809 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerName="machine-config-daemon" containerID="cri-o://52d1a350addd03a9e1e92a65fe40ee197504bf6db7f5b0601ca8860f562470fb" gracePeriod=600 Nov 25 11:31:35 crc kubenswrapper[4821]: I1125 11:31:35.685983 4821 generic.go:334] "Generic (PLEG): container finished" podID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerID="52d1a350addd03a9e1e92a65fe40ee197504bf6db7f5b0601ca8860f562470fb" exitCode=0 Nov 25 11:31:35 crc kubenswrapper[4821]: I1125 11:31:35.686063 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" event={"ID":"5f948b87-ac86-4de6-ad64-c2ef947f84d4","Type":"ContainerDied","Data":"52d1a350addd03a9e1e92a65fe40ee197504bf6db7f5b0601ca8860f562470fb"} Nov 25 11:31:35 crc kubenswrapper[4821]: I1125 11:31:35.686642 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" event={"ID":"5f948b87-ac86-4de6-ad64-c2ef947f84d4","Type":"ContainerStarted","Data":"f357cf6563c37c0c7cd86102a3e1eea3ba9146a525631fc115bab1d95f4015f3"} Nov 25 11:31:35 crc kubenswrapper[4821]: I1125 11:31:35.686680 4821 scope.go:117] "RemoveContainer" containerID="df8d79f5b2e040490d30845f855d9fa73fed8e867a647684ba5fc0077d412c79" Nov 25 11:31:46 crc kubenswrapper[4821]: I1125 11:31:46.768014 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-67b899d6f8-pscf9_b5bf27f5-8585-42d9-9d58-3c6e7f3a34d5/barbican-api/0.log" Nov 25 11:31:46 crc kubenswrapper[4821]: I1125 11:31:46.943033 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-67b899d6f8-pscf9_b5bf27f5-8585-42d9-9d58-3c6e7f3a34d5/barbican-api-log/0.log" Nov 25 11:31:47 crc kubenswrapper[4821]: I1125 11:31:47.000955 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-f75f66468-2k6lr_1e6fee50-9d84-4965-a222-a55f6dcfc4cc/barbican-keystone-listener/0.log" Nov 25 11:31:47 crc kubenswrapper[4821]: I1125 11:31:47.055989 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-f75f66468-2k6lr_1e6fee50-9d84-4965-a222-a55f6dcfc4cc/barbican-keystone-listener-log/0.log" Nov 25 11:31:47 crc kubenswrapper[4821]: I1125 11:31:47.450475 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5c9c8c87dc-mqk8l_62fc74a6-526c-4953-8439-b4af8ce5f39b/barbican-worker/0.log" Nov 25 11:31:47 crc kubenswrapper[4821]: I1125 11:31:47.492751 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5c9c8c87dc-mqk8l_62fc74a6-526c-4953-8439-b4af8ce5f39b/barbican-worker-log/0.log" Nov 25 11:31:47 crc kubenswrapper[4821]: I1125 11:31:47.661586 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-pkx5r_0b93320f-a41f-4c0e-874f-b4f945f14c90/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 11:31:47 crc kubenswrapper[4821]: I1125 11:31:47.739443 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_08ee9d9a-c488-4289-9d8f-a87d338992d6/ceilometer-central-agent/0.log" Nov 25 11:31:47 crc kubenswrapper[4821]: I1125 11:31:47.809909 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_08ee9d9a-c488-4289-9d8f-a87d338992d6/ceilometer-notification-agent/0.log" Nov 25 11:31:47 crc kubenswrapper[4821]: I1125 11:31:47.899745 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_08ee9d9a-c488-4289-9d8f-a87d338992d6/proxy-httpd/0.log" Nov 25 11:31:47 crc kubenswrapper[4821]: I1125 11:31:47.966919 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_08ee9d9a-c488-4289-9d8f-a87d338992d6/sg-core/0.log" Nov 25 11:31:48 crc kubenswrapper[4821]: I1125 11:31:48.060142 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_4fd0ccca-05dd-43e3-80b1-1c312d130056/cinder-api/0.log" Nov 25 11:31:48 crc kubenswrapper[4821]: I1125 11:31:48.161131 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_4fd0ccca-05dd-43e3-80b1-1c312d130056/cinder-api-log/0.log" Nov 25 11:31:48 crc kubenswrapper[4821]: I1125 11:31:48.326841 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_7b0c571f-6107-4ecd-8ffb-cd7b992d05a5/cinder-scheduler/0.log" Nov 25 11:31:48 crc kubenswrapper[4821]: I1125 11:31:48.434278 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_7b0c571f-6107-4ecd-8ffb-cd7b992d05a5/probe/0.log" Nov 25 11:31:48 crc kubenswrapper[4821]: I1125 11:31:48.570629 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-sg8cc_5a15c7fc-4220-46ae-8af3-490c28f8b4ff/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 11:31:48 crc kubenswrapper[4821]: I1125 11:31:48.654605 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-d8lr8_c20e6196-b767-4fbb-9d59-a88080a3aa11/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 11:31:48 crc kubenswrapper[4821]: I1125 11:31:48.822097 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-78c64bc9c5-8r5jj_bda8a37b-363b-446a-8cbb-a708cb3f716f/init/0.log" Nov 25 11:31:48 crc kubenswrapper[4821]: I1125 11:31:48.992074 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-78c64bc9c5-8r5jj_bda8a37b-363b-446a-8cbb-a708cb3f716f/init/0.log" Nov 25 11:31:49 crc kubenswrapper[4821]: I1125 11:31:49.063842 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-tfcn9_fcf6a791-0724-4ac2-8881-c7f8804f6540/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 11:31:49 crc kubenswrapper[4821]: I1125 11:31:49.065859 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-78c64bc9c5-8r5jj_bda8a37b-363b-446a-8cbb-a708cb3f716f/dnsmasq-dns/0.log" Nov 25 11:31:49 crc kubenswrapper[4821]: I1125 11:31:49.290345 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_fada5a81-0157-452d-9ade-dd6ce6303918/glance-log/0.log" Nov 25 11:31:49 crc kubenswrapper[4821]: I1125 11:31:49.312070 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_fada5a81-0157-452d-9ade-dd6ce6303918/glance-httpd/0.log" Nov 25 11:31:49 crc kubenswrapper[4821]: I1125 11:31:49.488295 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_1816fd90-5883-4de2-9124-64e326d1743f/glance-httpd/0.log" Nov 25 11:31:49 crc kubenswrapper[4821]: I1125 11:31:49.501563 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_1816fd90-5883-4de2-9124-64e326d1743f/glance-log/0.log" Nov 25 11:31:49 crc kubenswrapper[4821]: I1125 11:31:49.623478 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-67f765d856-ncc6t_93b4abfa-c14a-4df9-a6d1-9cdeaec918dd/horizon/0.log" Nov 25 11:31:49 crc kubenswrapper[4821]: I1125 11:31:49.797246 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-48x26_98d0e490-2704-4aa6-baa5-32ec701a5b68/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 11:31:50 crc kubenswrapper[4821]: I1125 11:31:50.022476 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-67f765d856-ncc6t_93b4abfa-c14a-4df9-a6d1-9cdeaec918dd/horizon-log/0.log" Nov 25 11:31:50 crc kubenswrapper[4821]: I1125 11:31:50.080306 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-x5qj5_9418fa73-031c-4c0a-bfbd-ffb206714075/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 11:31:50 crc kubenswrapper[4821]: I1125 11:31:50.392799 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-5fb467464c-hlrl9_eabb0cb5-e9c3-474f-80ba-df6ecf373c3f/keystone-api/0.log" Nov 25 11:31:50 crc kubenswrapper[4821]: I1125 11:31:50.487019 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29401141-j95bv_0f34ab84-5b6a-423d-b622-9b90c288695e/keystone-cron/0.log" Nov 25 11:31:50 crc kubenswrapper[4821]: I1125 11:31:50.690892 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_7c286dad-cd09-4085-a4e0-4b274f30d38d/kube-state-metrics/0.log" Nov 25 11:31:50 crc kubenswrapper[4821]: I1125 11:31:50.783566 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-9w8v8_0ee4d7f9-162b-4882-b189-13c53bcdc274/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 11:31:51 crc kubenswrapper[4821]: I1125 11:31:51.161013 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-55575b4d57-6t4kz_a8419a57-35ff-4340-b2d7-fd11724948fe/neutron-api/0.log" Nov 25 11:31:51 crc kubenswrapper[4821]: I1125 11:31:51.246961 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-55575b4d57-6t4kz_a8419a57-35ff-4340-b2d7-fd11724948fe/neutron-httpd/0.log" Nov 25 11:31:51 crc kubenswrapper[4821]: I1125 11:31:51.305348 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-225z5_0f26a3b0-ba05-4484-8281-215ff2f09d25/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 11:31:51 crc kubenswrapper[4821]: I1125 11:31:51.933842 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_2bd244d2-4836-4caf-a6f6-1a00aac5dbb3/nova-api-log/0.log" Nov 25 11:31:51 crc kubenswrapper[4821]: I1125 11:31:51.951315 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_e5dfcd15-af77-4be7-93db-e26f3da1d7d2/nova-cell0-conductor-conductor/0.log" Nov 25 11:31:52 crc kubenswrapper[4821]: I1125 11:31:52.201656 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_fb540939-5683-495a-b8df-d7a000fedaa2/nova-cell1-conductor-conductor/0.log" Nov 25 11:31:52 crc kubenswrapper[4821]: I1125 11:31:52.206572 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_2bd244d2-4836-4caf-a6f6-1a00aac5dbb3/nova-api-api/0.log" Nov 25 11:31:52 crc kubenswrapper[4821]: I1125 11:31:52.305678 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_5b8402cb-db96-4ea4-8f95-a2ea84f202f5/nova-cell1-novncproxy-novncproxy/0.log" Nov 25 11:31:52 crc kubenswrapper[4821]: I1125 11:31:52.498724 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-mt9vk_40682cac-f449-4465-9ad3-3bfa28b3800d/nova-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 11:31:52 crc kubenswrapper[4821]: I1125 11:31:52.681911 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_39dff25e-5059-4eaf-bf98-93aea0eeb258/nova-metadata-log/0.log" Nov 25 11:31:52 crc kubenswrapper[4821]: I1125 11:31:52.992822 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_3bdff849-753c-42b5-9e06-2e37dfe42666/nova-scheduler-scheduler/0.log" Nov 25 11:31:53 crc kubenswrapper[4821]: I1125 11:31:53.047969 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_41c8c049-b466-4c19-944e-50a639491190/mysql-bootstrap/0.log" Nov 25 11:31:53 crc kubenswrapper[4821]: I1125 11:31:53.187738 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_41c8c049-b466-4c19-944e-50a639491190/mysql-bootstrap/0.log" Nov 25 11:31:53 crc kubenswrapper[4821]: I1125 11:31:53.293721 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_41c8c049-b466-4c19-944e-50a639491190/galera/0.log" Nov 25 11:31:53 crc kubenswrapper[4821]: I1125 11:31:53.444049 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_ecb5eaf2-4178-43ec-bc21-0e05de8fc237/mysql-bootstrap/0.log" Nov 25 11:31:53 crc kubenswrapper[4821]: I1125 11:31:53.626117 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_ecb5eaf2-4178-43ec-bc21-0e05de8fc237/mysql-bootstrap/0.log" Nov 25 11:31:53 crc kubenswrapper[4821]: I1125 11:31:53.679900 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_ecb5eaf2-4178-43ec-bc21-0e05de8fc237/galera/0.log" Nov 25 11:31:53 crc kubenswrapper[4821]: I1125 11:31:53.831592 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_d969bbb6-564d-4563-98c6-5e2ec965c2dc/openstackclient/0.log" Nov 25 11:31:53 crc kubenswrapper[4821]: I1125 11:31:53.985231 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_39dff25e-5059-4eaf-bf98-93aea0eeb258/nova-metadata-metadata/0.log" Nov 25 11:31:54 crc kubenswrapper[4821]: I1125 11:31:54.023956 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ltxc6_47eaf58d-190b-4e75-9aac-11d7e51df259/ovn-controller/0.log" Nov 25 11:31:54 crc kubenswrapper[4821]: I1125 11:31:54.384645 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-kkntf_832f5849-e58a-407b-9f45-55c52cfbcd02/openstack-network-exporter/0.log" Nov 25 11:31:54 crc kubenswrapper[4821]: I1125 11:31:54.431743 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-6mbr2_72b307b1-f36e-424f-91ef-f7402b6d7535/ovsdb-server-init/0.log" Nov 25 11:31:54 crc kubenswrapper[4821]: I1125 11:31:54.743882 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-6mbr2_72b307b1-f36e-424f-91ef-f7402b6d7535/ovsdb-server-init/0.log" Nov 25 11:31:54 crc kubenswrapper[4821]: I1125 11:31:54.763486 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-6mbr2_72b307b1-f36e-424f-91ef-f7402b6d7535/ovsdb-server/0.log" Nov 25 11:31:54 crc kubenswrapper[4821]: I1125 11:31:54.795301 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-6mbr2_72b307b1-f36e-424f-91ef-f7402b6d7535/ovs-vswitchd/0.log" Nov 25 11:31:55 crc kubenswrapper[4821]: I1125 11:31:55.010279 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-9vb7k_df834005-0241-4897-9112-05186899309e/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 11:31:55 crc kubenswrapper[4821]: I1125 11:31:55.071845 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_006cceec-8fb3-44a5-b3e1-50a01bc4b334/openstack-network-exporter/0.log" Nov 25 11:31:55 crc kubenswrapper[4821]: I1125 11:31:55.116104 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_006cceec-8fb3-44a5-b3e1-50a01bc4b334/ovn-northd/0.log" Nov 25 11:31:55 crc kubenswrapper[4821]: I1125 11:31:55.246316 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_30500d7b-4e4b-4839-8a9d-0b05a0ebd1fb/openstack-network-exporter/0.log" Nov 25 11:31:55 crc kubenswrapper[4821]: I1125 11:31:55.287580 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_30500d7b-4e4b-4839-8a9d-0b05a0ebd1fb/ovsdbserver-nb/0.log" Nov 25 11:31:55 crc kubenswrapper[4821]: I1125 11:31:55.533440 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_9aed573a-ad60-479a-ba27-8666d7dc4d9e/openstack-network-exporter/0.log" Nov 25 11:31:55 crc kubenswrapper[4821]: I1125 11:31:55.574267 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_9aed573a-ad60-479a-ba27-8666d7dc4d9e/ovsdbserver-sb/0.log" Nov 25 11:31:55 crc kubenswrapper[4821]: I1125 11:31:55.740087 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-f8667547d-9v5xh_b755bd74-3de7-49ed-857a-a5f23cf84f30/placement-api/0.log" Nov 25 11:31:55 crc kubenswrapper[4821]: I1125 11:31:55.881993 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_3556cc6b-7d73-4cbc-9c2e-378c6ba606e4/setup-container/0.log" Nov 25 11:31:55 crc kubenswrapper[4821]: I1125 11:31:55.901869 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-f8667547d-9v5xh_b755bd74-3de7-49ed-857a-a5f23cf84f30/placement-log/0.log" Nov 25 11:31:56 crc kubenswrapper[4821]: I1125 11:31:56.070947 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_3556cc6b-7d73-4cbc-9c2e-378c6ba606e4/rabbitmq/0.log" Nov 25 11:31:56 crc kubenswrapper[4821]: I1125 11:31:56.074913 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_3556cc6b-7d73-4cbc-9c2e-378c6ba606e4/setup-container/0.log" Nov 25 11:31:56 crc kubenswrapper[4821]: I1125 11:31:56.191851 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_50c27d16-2f67-4f90-83f0-83e16161de50/setup-container/0.log" Nov 25 11:31:56 crc kubenswrapper[4821]: I1125 11:31:56.374220 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_50c27d16-2f67-4f90-83f0-83e16161de50/setup-container/0.log" Nov 25 11:31:56 crc kubenswrapper[4821]: I1125 11:31:56.396103 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_50c27d16-2f67-4f90-83f0-83e16161de50/rabbitmq/0.log" Nov 25 11:31:56 crc kubenswrapper[4821]: I1125 11:31:56.468386 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-98mzz_0d9e3bea-dfb1-479d-a1b2-a17b2334cd86/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 11:31:56 crc kubenswrapper[4821]: I1125 11:31:56.618631 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-9tzcw_74c5a130-3a40-4920-8e84-e90e9ee4b39f/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 11:31:56 crc kubenswrapper[4821]: I1125 11:31:56.767887 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-rdcx6_a15db0db-c211-40aa-bb3e-8f222669b005/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 11:31:56 crc kubenswrapper[4821]: I1125 11:31:56.984287 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-j8sp8_3ad8bc38-c088-4632-8864-00d472c78fbf/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 11:31:57 crc kubenswrapper[4821]: I1125 11:31:57.092851 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-dxdpz_e1247bbd-54fc-45ce-8a90-0b8a5df02c47/ssh-known-hosts-edpm-deployment/0.log" Nov 25 11:31:57 crc kubenswrapper[4821]: I1125 11:31:57.271596 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-ff8c8c76c-ctf8b_db085977-6941-483f-a800-411ff71612af/proxy-server/0.log" Nov 25 11:31:57 crc kubenswrapper[4821]: I1125 11:31:57.354111 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-gmrpw_1aa8233d-1fe4-44ec-acc6-0d5548c70d59/swift-ring-rebalance/0.log" Nov 25 11:31:57 crc kubenswrapper[4821]: I1125 11:31:57.362215 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-ff8c8c76c-ctf8b_db085977-6941-483f-a800-411ff71612af/proxy-httpd/0.log" Nov 25 11:31:57 crc kubenswrapper[4821]: I1125 11:31:57.527858 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0ffb0da2-affb-4a95-87f3-eac442dc754d/account-auditor/0.log" Nov 25 11:31:57 crc kubenswrapper[4821]: I1125 11:31:57.643319 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0ffb0da2-affb-4a95-87f3-eac442dc754d/account-replicator/0.log" Nov 25 11:31:57 crc kubenswrapper[4821]: I1125 11:31:57.655647 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0ffb0da2-affb-4a95-87f3-eac442dc754d/account-reaper/0.log" Nov 25 11:31:57 crc kubenswrapper[4821]: I1125 11:31:57.941973 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0ffb0da2-affb-4a95-87f3-eac442dc754d/account-server/0.log" Nov 25 11:31:58 crc kubenswrapper[4821]: I1125 11:31:58.073207 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0ffb0da2-affb-4a95-87f3-eac442dc754d/container-auditor/0.log" Nov 25 11:31:58 crc kubenswrapper[4821]: I1125 11:31:58.087106 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0ffb0da2-affb-4a95-87f3-eac442dc754d/container-server/0.log" Nov 25 11:31:58 crc kubenswrapper[4821]: I1125 11:31:58.091491 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0ffb0da2-affb-4a95-87f3-eac442dc754d/container-replicator/0.log" Nov 25 11:31:58 crc kubenswrapper[4821]: I1125 11:31:58.219686 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0ffb0da2-affb-4a95-87f3-eac442dc754d/container-updater/0.log" Nov 25 11:31:58 crc kubenswrapper[4821]: I1125 11:31:58.352448 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0ffb0da2-affb-4a95-87f3-eac442dc754d/object-auditor/0.log" Nov 25 11:31:58 crc kubenswrapper[4821]: I1125 11:31:58.359880 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0ffb0da2-affb-4a95-87f3-eac442dc754d/object-expirer/0.log" Nov 25 11:31:58 crc kubenswrapper[4821]: I1125 11:31:58.409674 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0ffb0da2-affb-4a95-87f3-eac442dc754d/object-replicator/0.log" Nov 25 11:31:58 crc kubenswrapper[4821]: I1125 11:31:58.437561 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0ffb0da2-affb-4a95-87f3-eac442dc754d/object-server/0.log" Nov 25 11:31:58 crc kubenswrapper[4821]: I1125 11:31:58.591470 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0ffb0da2-affb-4a95-87f3-eac442dc754d/rsync/0.log" Nov 25 11:31:58 crc kubenswrapper[4821]: I1125 11:31:58.639005 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0ffb0da2-affb-4a95-87f3-eac442dc754d/swift-recon-cron/0.log" Nov 25 11:31:58 crc kubenswrapper[4821]: I1125 11:31:58.642644 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0ffb0da2-affb-4a95-87f3-eac442dc754d/object-updater/0.log" Nov 25 11:31:58 crc kubenswrapper[4821]: I1125 11:31:58.866811 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-w75cf_814ef062-0cc8-40bf-ac77-c87010466f7d/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 11:31:58 crc kubenswrapper[4821]: I1125 11:31:58.889344 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_d8aea755-bf2f-4443-8196-46936ce3e2fd/tempest-tests-tempest-tests-runner/0.log" Nov 25 11:31:59 crc kubenswrapper[4821]: I1125 11:31:59.121699 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_adcf552f-2365-451b-b4e2-12b58631195e/test-operator-logs-container/0.log" Nov 25 11:31:59 crc kubenswrapper[4821]: I1125 11:31:59.214988 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-x5tw6_d77aa250-b203-4596-913c-7878a4d8b517/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 11:32:05 crc kubenswrapper[4821]: I1125 11:32:05.352364 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_8b9e8e7c-b055-47d3-874a-227d5f02432a/memcached/0.log" Nov 25 11:32:21 crc kubenswrapper[4821]: I1125 11:32:21.321174 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-86dc4d89c8-pnq4j_db28ec64-d8be-4b56-a427-e4cceb256fd0/kube-rbac-proxy/0.log" Nov 25 11:32:21 crc kubenswrapper[4821]: I1125 11:32:21.398284 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-86dc4d89c8-pnq4j_db28ec64-d8be-4b56-a427-e4cceb256fd0/manager/0.log" Nov 25 11:32:21 crc kubenswrapper[4821]: I1125 11:32:21.565215 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-79856dc55c-p4bwc_8f16d64f-ea14-4853-b30b-2df294ce6cc3/kube-rbac-proxy/0.log" Nov 25 11:32:21 crc kubenswrapper[4821]: I1125 11:32:21.600788 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-79856dc55c-p4bwc_8f16d64f-ea14-4853-b30b-2df294ce6cc3/manager/0.log" Nov 25 11:32:21 crc kubenswrapper[4821]: I1125 11:32:21.703506 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-7d695c9b56-7sx9m_2236607e-2a7b-4858-b5b9-c43e47c9375d/kube-rbac-proxy/0.log" Nov 25 11:32:21 crc kubenswrapper[4821]: I1125 11:32:21.739935 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-7d695c9b56-7sx9m_2236607e-2a7b-4858-b5b9-c43e47c9375d/manager/0.log" Nov 25 11:32:21 crc kubenswrapper[4821]: I1125 11:32:21.772642 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f381f3655dcf4f93a6a93fe2921d660ff39f878b90de9b88a6ab8f46514g6tz_bd4ce1de-05e0-49f4-ad3a-7baae72f3e09/util/0.log" Nov 25 11:32:22 crc kubenswrapper[4821]: I1125 11:32:22.005014 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f381f3655dcf4f93a6a93fe2921d660ff39f878b90de9b88a6ab8f46514g6tz_bd4ce1de-05e0-49f4-ad3a-7baae72f3e09/util/0.log" Nov 25 11:32:22 crc kubenswrapper[4821]: I1125 11:32:22.010698 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f381f3655dcf4f93a6a93fe2921d660ff39f878b90de9b88a6ab8f46514g6tz_bd4ce1de-05e0-49f4-ad3a-7baae72f3e09/pull/0.log" Nov 25 11:32:22 crc kubenswrapper[4821]: I1125 11:32:22.021143 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f381f3655dcf4f93a6a93fe2921d660ff39f878b90de9b88a6ab8f46514g6tz_bd4ce1de-05e0-49f4-ad3a-7baae72f3e09/pull/0.log" Nov 25 11:32:22 crc kubenswrapper[4821]: I1125 11:32:22.188771 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f381f3655dcf4f93a6a93fe2921d660ff39f878b90de9b88a6ab8f46514g6tz_bd4ce1de-05e0-49f4-ad3a-7baae72f3e09/pull/0.log" Nov 25 11:32:22 crc kubenswrapper[4821]: I1125 11:32:22.209525 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f381f3655dcf4f93a6a93fe2921d660ff39f878b90de9b88a6ab8f46514g6tz_bd4ce1de-05e0-49f4-ad3a-7baae72f3e09/extract/0.log" Nov 25 11:32:22 crc kubenswrapper[4821]: I1125 11:32:22.216391 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f381f3655dcf4f93a6a93fe2921d660ff39f878b90de9b88a6ab8f46514g6tz_bd4ce1de-05e0-49f4-ad3a-7baae72f3e09/util/0.log" Nov 25 11:32:22 crc kubenswrapper[4821]: I1125 11:32:22.367670 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-68b95954c9-qgs92_a3dc5167-4997-44bf-ace3-706829fd0f17/kube-rbac-proxy/0.log" Nov 25 11:32:22 crc kubenswrapper[4821]: I1125 11:32:22.446897 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-68b95954c9-qgs92_a3dc5167-4997-44bf-ace3-706829fd0f17/manager/0.log" Nov 25 11:32:22 crc kubenswrapper[4821]: I1125 11:32:22.454045 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-774b86978c-s49f6_b35e42b0-2a6f-454f-b8e2-4f7cd9b9217f/kube-rbac-proxy/0.log" Nov 25 11:32:22 crc kubenswrapper[4821]: I1125 11:32:22.577223 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-774b86978c-s49f6_b35e42b0-2a6f-454f-b8e2-4f7cd9b9217f/manager/0.log" Nov 25 11:32:22 crc kubenswrapper[4821]: I1125 11:32:22.658232 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c9694994-2qzrg_b7895321-27ec-4c75-aa5e-581fe46e0789/kube-rbac-proxy/0.log" Nov 25 11:32:22 crc kubenswrapper[4821]: I1125 11:32:22.685543 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c9694994-2qzrg_b7895321-27ec-4c75-aa5e-581fe46e0789/manager/0.log" Nov 25 11:32:22 crc kubenswrapper[4821]: I1125 11:32:22.804510 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-d5cc86f4b-8cqfp_9db84cac-a8df-4359-8051-ebc064a97b4d/kube-rbac-proxy/0.log" Nov 25 11:32:23 crc kubenswrapper[4821]: I1125 11:32:23.029241 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5bfcdc958c-fqfwh_1976151f-2b7a-4cde-95d3-9020b17d938c/kube-rbac-proxy/0.log" Nov 25 11:32:23 crc kubenswrapper[4821]: I1125 11:32:23.042133 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-d5cc86f4b-8cqfp_9db84cac-a8df-4359-8051-ebc064a97b4d/manager/0.log" Nov 25 11:32:23 crc kubenswrapper[4821]: I1125 11:32:23.069100 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5bfcdc958c-fqfwh_1976151f-2b7a-4cde-95d3-9020b17d938c/manager/0.log" Nov 25 11:32:23 crc kubenswrapper[4821]: I1125 11:32:23.289765 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-748dc6576f-f8tvv_7982734e-b887-4cea-9724-3a37f9369ac5/kube-rbac-proxy/0.log" Nov 25 11:32:23 crc kubenswrapper[4821]: I1125 11:32:23.303353 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-748dc6576f-f8tvv_7982734e-b887-4cea-9724-3a37f9369ac5/manager/0.log" Nov 25 11:32:23 crc kubenswrapper[4821]: I1125 11:32:23.343717 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-58bb8d67cc-vnf9x_5fefd5f2-18cb-4c63-848f-2fc9d63bb828/kube-rbac-proxy/0.log" Nov 25 11:32:23 crc kubenswrapper[4821]: I1125 11:32:23.443502 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-58bb8d67cc-vnf9x_5fefd5f2-18cb-4c63-848f-2fc9d63bb828/manager/0.log" Nov 25 11:32:23 crc kubenswrapper[4821]: I1125 11:32:23.487030 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-cb6c4fdb7-fw96v_d42d0991-9e6f-46d0-9236-00e85c26a603/kube-rbac-proxy/0.log" Nov 25 11:32:23 crc kubenswrapper[4821]: I1125 11:32:23.543258 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-cb6c4fdb7-fw96v_d42d0991-9e6f-46d0-9236-00e85c26a603/manager/0.log" Nov 25 11:32:23 crc kubenswrapper[4821]: I1125 11:32:23.678438 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-7c57c8bbc4-hwp4x_c0376a5f-ed42-4399-8f41-2d2157a654fd/kube-rbac-proxy/0.log" Nov 25 11:32:23 crc kubenswrapper[4821]: I1125 11:32:23.735560 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-7c57c8bbc4-hwp4x_c0376a5f-ed42-4399-8f41-2d2157a654fd/manager/0.log" Nov 25 11:32:23 crc kubenswrapper[4821]: I1125 11:32:23.920430 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-79556f57fc-4ftbp_1f69797f-4204-4989-a00b-d9722d44d3c6/kube-rbac-proxy/0.log" Nov 25 11:32:23 crc kubenswrapper[4821]: I1125 11:32:23.946424 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-79556f57fc-4ftbp_1f69797f-4204-4989-a00b-d9722d44d3c6/manager/0.log" Nov 25 11:32:23 crc kubenswrapper[4821]: I1125 11:32:23.956076 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-fd75fd47d-5ncqq_db1d55a6-1285-429d-b9dd-b2e100e602c6/kube-rbac-proxy/0.log" Nov 25 11:32:24 crc kubenswrapper[4821]: I1125 11:32:24.115473 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-fd75fd47d-5ncqq_db1d55a6-1285-429d-b9dd-b2e100e602c6/manager/0.log" Nov 25 11:32:24 crc kubenswrapper[4821]: I1125 11:32:24.159222 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-544b9bb9-lwp9z_14385dd9-9b06-473b-919d-9d2f9207ffe1/kube-rbac-proxy/0.log" Nov 25 11:32:24 crc kubenswrapper[4821]: I1125 11:32:24.186893 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-544b9bb9-lwp9z_14385dd9-9b06-473b-919d-9d2f9207ffe1/manager/0.log" Nov 25 11:32:24 crc kubenswrapper[4821]: I1125 11:32:24.660356 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-7f8f598fc5-8gsxh_1875cb8b-6a5c-48dc-bf01-07eef1d129fe/operator/0.log" Nov 25 11:32:24 crc kubenswrapper[4821]: I1125 11:32:24.727620 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-tzjrl_afdcf8ff-544a-4625-b72e-d1a880434014/registry-server/0.log" Nov 25 11:32:25 crc kubenswrapper[4821]: I1125 11:32:25.066545 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-66cf5c67ff-qxgcp_2b44040a-e224-4468-a528-9f96a98e3891/kube-rbac-proxy/0.log" Nov 25 11:32:25 crc kubenswrapper[4821]: I1125 11:32:25.193332 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-66cf5c67ff-qxgcp_2b44040a-e224-4468-a528-9f96a98e3891/manager/0.log" Nov 25 11:32:25 crc kubenswrapper[4821]: I1125 11:32:25.269794 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5db546f9d9-66kf4_876c088c-7ee5-4b61-a798-eb36ac1e7e21/kube-rbac-proxy/0.log" Nov 25 11:32:25 crc kubenswrapper[4821]: I1125 11:32:25.416265 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5db546f9d9-66kf4_876c088c-7ee5-4b61-a798-eb36ac1e7e21/manager/0.log" Nov 25 11:32:25 crc kubenswrapper[4821]: I1125 11:32:25.521135 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-fprdh_5cea2a2e-a19f-42fc-ac84-aadfa677ec4c/operator/0.log" Nov 25 11:32:25 crc kubenswrapper[4821]: I1125 11:32:25.537935 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-64b4f4bf8b-vzb6r_40e2ab14-9d05-4569-8a3a-69a34586c504/manager/0.log" Nov 25 11:32:25 crc kubenswrapper[4821]: I1125 11:32:25.634558 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6fdc4fcf86-75q4l_c77cb0f4-1dff-466c-a8bf-6a987bc87ced/kube-rbac-proxy/0.log" Nov 25 11:32:25 crc kubenswrapper[4821]: I1125 11:32:25.697956 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6fdc4fcf86-75q4l_c77cb0f4-1dff-466c-a8bf-6a987bc87ced/manager/0.log" Nov 25 11:32:25 crc kubenswrapper[4821]: I1125 11:32:25.701429 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-8dbc5685d-gcl85_a64e88ce-fe1a-4cb1-984a-c6d8b5e5ffd9/kube-rbac-proxy/0.log" Nov 25 11:32:25 crc kubenswrapper[4821]: I1125 11:32:25.794822 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-8dbc5685d-gcl85_a64e88ce-fe1a-4cb1-984a-c6d8b5e5ffd9/manager/0.log" Nov 25 11:32:25 crc kubenswrapper[4821]: I1125 11:32:25.877440 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cb74df96-lppg4_8fef5c2e-676a-49b6-a10c-94abe226600a/kube-rbac-proxy/0.log" Nov 25 11:32:25 crc kubenswrapper[4821]: I1125 11:32:25.906914 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cb74df96-lppg4_8fef5c2e-676a-49b6-a10c-94abe226600a/manager/0.log" Nov 25 11:32:25 crc kubenswrapper[4821]: I1125 11:32:25.995051 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-864885998-vnmtq_505faacd-65f5-4078-9867-9ae648652b50/kube-rbac-proxy/0.log" Nov 25 11:32:26 crc kubenswrapper[4821]: I1125 11:32:26.059716 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-864885998-vnmtq_505faacd-65f5-4078-9867-9ae648652b50/manager/0.log" Nov 25 11:32:41 crc kubenswrapper[4821]: I1125 11:32:41.622808 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-2stv6_4a74ce9e-b145-4e40-ad4e-67ecdf722195/control-plane-machine-set-operator/0.log" Nov 25 11:32:41 crc kubenswrapper[4821]: I1125 11:32:41.770504 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-zj4s2_24708c13-efe2-435b-acde-b29b48871f16/machine-api-operator/0.log" Nov 25 11:32:41 crc kubenswrapper[4821]: I1125 11:32:41.804475 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-zj4s2_24708c13-efe2-435b-acde-b29b48871f16/kube-rbac-proxy/0.log" Nov 25 11:32:53 crc kubenswrapper[4821]: I1125 11:32:53.379754 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-crm7d_c7f95e74-b964-455f-bacd-60e1fceb87fb/cert-manager-controller/0.log" Nov 25 11:32:53 crc kubenswrapper[4821]: I1125 11:32:53.503478 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-g7gr9_baa81eb8-c73e-41d1-bd01-9958477aa031/cert-manager-cainjector/0.log" Nov 25 11:32:53 crc kubenswrapper[4821]: I1125 11:32:53.534743 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-4qh5c_4096527e-0fc1-4991-ad0c-f47a84cd6e84/cert-manager-webhook/0.log" Nov 25 11:33:05 crc kubenswrapper[4821]: I1125 11:33:05.533680 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-5874bd7bc5-7mnfd_dff35e42-91ef-41f1-81b9-c12191388f31/nmstate-console-plugin/0.log" Nov 25 11:33:05 crc kubenswrapper[4821]: I1125 11:33:05.703932 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-89r7d_e29f6151-8a55-4b3e-8a15-8f8ec4b5ac66/nmstate-handler/0.log" Nov 25 11:33:05 crc kubenswrapper[4821]: I1125 11:33:05.817723 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-79cx4_221a1f74-bcbb-4788-83d1-4c5c92d3fedd/kube-rbac-proxy/0.log" Nov 25 11:33:05 crc kubenswrapper[4821]: I1125 11:33:05.848629 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-79cx4_221a1f74-bcbb-4788-83d1-4c5c92d3fedd/nmstate-metrics/0.log" Nov 25 11:33:06 crc kubenswrapper[4821]: I1125 11:33:06.002574 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-557fdffb88-d8wjl_7218bb0e-8630-47d7-a259-3d4f79f1a470/nmstate-operator/0.log" Nov 25 11:33:06 crc kubenswrapper[4821]: I1125 11:33:06.027308 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6b89b748d8-c9bsx_de41e739-6f86-44cc-80a2-99bb874d259f/nmstate-webhook/0.log" Nov 25 11:33:19 crc kubenswrapper[4821]: I1125 11:33:19.845028 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-5knnl_902ced61-7b2c-4a77-b8fc-ea9a226e05ba/kube-rbac-proxy/0.log" Nov 25 11:33:19 crc kubenswrapper[4821]: I1125 11:33:19.862244 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-5knnl_902ced61-7b2c-4a77-b8fc-ea9a226e05ba/controller/0.log" Nov 25 11:33:20 crc kubenswrapper[4821]: I1125 11:33:20.040147 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7knkm_0a268005-d45e-4cb1-b04a-5bedb97ea5ee/cp-frr-files/0.log" Nov 25 11:33:20 crc kubenswrapper[4821]: I1125 11:33:20.248419 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7knkm_0a268005-d45e-4cb1-b04a-5bedb97ea5ee/cp-reloader/0.log" Nov 25 11:33:20 crc kubenswrapper[4821]: I1125 11:33:20.257503 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7knkm_0a268005-d45e-4cb1-b04a-5bedb97ea5ee/cp-reloader/0.log" Nov 25 11:33:20 crc kubenswrapper[4821]: I1125 11:33:20.265483 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7knkm_0a268005-d45e-4cb1-b04a-5bedb97ea5ee/cp-metrics/0.log" Nov 25 11:33:20 crc kubenswrapper[4821]: I1125 11:33:20.279738 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7knkm_0a268005-d45e-4cb1-b04a-5bedb97ea5ee/cp-frr-files/0.log" Nov 25 11:33:20 crc kubenswrapper[4821]: I1125 11:33:20.517541 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7knkm_0a268005-d45e-4cb1-b04a-5bedb97ea5ee/cp-metrics/0.log" Nov 25 11:33:20 crc kubenswrapper[4821]: I1125 11:33:20.521819 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7knkm_0a268005-d45e-4cb1-b04a-5bedb97ea5ee/cp-reloader/0.log" Nov 25 11:33:20 crc kubenswrapper[4821]: I1125 11:33:20.538032 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7knkm_0a268005-d45e-4cb1-b04a-5bedb97ea5ee/cp-frr-files/0.log" Nov 25 11:33:20 crc kubenswrapper[4821]: I1125 11:33:20.575930 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7knkm_0a268005-d45e-4cb1-b04a-5bedb97ea5ee/cp-metrics/0.log" Nov 25 11:33:20 crc kubenswrapper[4821]: I1125 11:33:20.691612 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7knkm_0a268005-d45e-4cb1-b04a-5bedb97ea5ee/cp-frr-files/0.log" Nov 25 11:33:20 crc kubenswrapper[4821]: I1125 11:33:20.721437 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7knkm_0a268005-d45e-4cb1-b04a-5bedb97ea5ee/cp-reloader/0.log" Nov 25 11:33:20 crc kubenswrapper[4821]: I1125 11:33:20.759906 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7knkm_0a268005-d45e-4cb1-b04a-5bedb97ea5ee/cp-metrics/0.log" Nov 25 11:33:20 crc kubenswrapper[4821]: I1125 11:33:20.797552 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7knkm_0a268005-d45e-4cb1-b04a-5bedb97ea5ee/controller/0.log" Nov 25 11:33:20 crc kubenswrapper[4821]: I1125 11:33:20.896873 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7knkm_0a268005-d45e-4cb1-b04a-5bedb97ea5ee/frr-metrics/0.log" Nov 25 11:33:20 crc kubenswrapper[4821]: I1125 11:33:20.941473 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7knkm_0a268005-d45e-4cb1-b04a-5bedb97ea5ee/kube-rbac-proxy/0.log" Nov 25 11:33:20 crc kubenswrapper[4821]: I1125 11:33:20.983492 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7knkm_0a268005-d45e-4cb1-b04a-5bedb97ea5ee/kube-rbac-proxy-frr/0.log" Nov 25 11:33:21 crc kubenswrapper[4821]: I1125 11:33:21.137300 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7knkm_0a268005-d45e-4cb1-b04a-5bedb97ea5ee/reloader/0.log" Nov 25 11:33:21 crc kubenswrapper[4821]: I1125 11:33:21.202086 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-6998585d5-bclxv_1144ecf2-7246-4c05-a7ca-f7b7213092fd/frr-k8s-webhook-server/0.log" Nov 25 11:33:21 crc kubenswrapper[4821]: I1125 11:33:21.420157 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-644d88f8f5-gtq55_b1969fcd-bedc-42dc-966d-9cd5b8770fd9/manager/0.log" Nov 25 11:33:21 crc kubenswrapper[4821]: I1125 11:33:21.546638 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-666cb5dd59-l9nfz_83c8e5d6-48b9-47ff-8c0a-aad92bbe511a/webhook-server/0.log" Nov 25 11:33:21 crc kubenswrapper[4821]: I1125 11:33:21.711007 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-5qbb4_c8b12b33-c9ca-48dd-9f7f-96b5f1c40f86/kube-rbac-proxy/0.log" Nov 25 11:33:22 crc kubenswrapper[4821]: I1125 11:33:22.184622 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-5qbb4_c8b12b33-c9ca-48dd-9f7f-96b5f1c40f86/speaker/0.log" Nov 25 11:33:22 crc kubenswrapper[4821]: I1125 11:33:22.344128 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7knkm_0a268005-d45e-4cb1-b04a-5bedb97ea5ee/frr/0.log" Nov 25 11:33:33 crc kubenswrapper[4821]: I1125 11:33:33.315384 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6jtmt_faf2c05e-02b4-4b92-b614-d3de8e443677/util/0.log" Nov 25 11:33:33 crc kubenswrapper[4821]: I1125 11:33:33.501513 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6jtmt_faf2c05e-02b4-4b92-b614-d3de8e443677/pull/0.log" Nov 25 11:33:33 crc kubenswrapper[4821]: I1125 11:33:33.503938 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6jtmt_faf2c05e-02b4-4b92-b614-d3de8e443677/pull/0.log" Nov 25 11:33:33 crc kubenswrapper[4821]: I1125 11:33:33.505017 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6jtmt_faf2c05e-02b4-4b92-b614-d3de8e443677/util/0.log" Nov 25 11:33:33 crc kubenswrapper[4821]: I1125 11:33:33.676542 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6jtmt_faf2c05e-02b4-4b92-b614-d3de8e443677/pull/0.log" Nov 25 11:33:33 crc kubenswrapper[4821]: I1125 11:33:33.678628 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6jtmt_faf2c05e-02b4-4b92-b614-d3de8e443677/util/0.log" Nov 25 11:33:33 crc kubenswrapper[4821]: I1125 11:33:33.688673 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6jtmt_faf2c05e-02b4-4b92-b614-d3de8e443677/extract/0.log" Nov 25 11:33:33 crc kubenswrapper[4821]: I1125 11:33:33.829812 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-529qb_410373d3-842e-43b3-a2d8-2293c543a4ac/extract-utilities/0.log" Nov 25 11:33:33 crc kubenswrapper[4821]: I1125 11:33:33.995381 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-529qb_410373d3-842e-43b3-a2d8-2293c543a4ac/extract-utilities/0.log" Nov 25 11:33:34 crc kubenswrapper[4821]: I1125 11:33:34.039834 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-529qb_410373d3-842e-43b3-a2d8-2293c543a4ac/extract-content/0.log" Nov 25 11:33:34 crc kubenswrapper[4821]: I1125 11:33:34.053152 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-529qb_410373d3-842e-43b3-a2d8-2293c543a4ac/extract-content/0.log" Nov 25 11:33:34 crc kubenswrapper[4821]: I1125 11:33:34.219157 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-529qb_410373d3-842e-43b3-a2d8-2293c543a4ac/extract-utilities/0.log" Nov 25 11:33:34 crc kubenswrapper[4821]: I1125 11:33:34.224355 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-529qb_410373d3-842e-43b3-a2d8-2293c543a4ac/extract-content/0.log" Nov 25 11:33:34 crc kubenswrapper[4821]: I1125 11:33:34.438830 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4j8hk_a43c28e0-8883-4567-a86f-bcd311098876/extract-utilities/0.log" Nov 25 11:33:34 crc kubenswrapper[4821]: I1125 11:33:34.479695 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-529qb_410373d3-842e-43b3-a2d8-2293c543a4ac/registry-server/0.log" Nov 25 11:33:34 crc kubenswrapper[4821]: I1125 11:33:34.617652 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4j8hk_a43c28e0-8883-4567-a86f-bcd311098876/extract-content/0.log" Nov 25 11:33:34 crc kubenswrapper[4821]: I1125 11:33:34.630570 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4j8hk_a43c28e0-8883-4567-a86f-bcd311098876/extract-content/0.log" Nov 25 11:33:34 crc kubenswrapper[4821]: I1125 11:33:34.646722 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4j8hk_a43c28e0-8883-4567-a86f-bcd311098876/extract-utilities/0.log" Nov 25 11:33:34 crc kubenswrapper[4821]: I1125 11:33:34.783760 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4j8hk_a43c28e0-8883-4567-a86f-bcd311098876/extract-utilities/0.log" Nov 25 11:33:34 crc kubenswrapper[4821]: I1125 11:33:34.822026 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4j8hk_a43c28e0-8883-4567-a86f-bcd311098876/extract-content/0.log" Nov 25 11:33:34 crc kubenswrapper[4821]: I1125 11:33:34.972970 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6hsmqh_8c67f775-e13f-4c7d-9d2b-115d492961f1/util/0.log" Nov 25 11:33:35 crc kubenswrapper[4821]: I1125 11:33:35.259453 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6hsmqh_8c67f775-e13f-4c7d-9d2b-115d492961f1/util/0.log" Nov 25 11:33:35 crc kubenswrapper[4821]: I1125 11:33:35.284501 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6hsmqh_8c67f775-e13f-4c7d-9d2b-115d492961f1/pull/0.log" Nov 25 11:33:35 crc kubenswrapper[4821]: I1125 11:33:35.291864 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6hsmqh_8c67f775-e13f-4c7d-9d2b-115d492961f1/pull/0.log" Nov 25 11:33:35 crc kubenswrapper[4821]: I1125 11:33:35.445182 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6hsmqh_8c67f775-e13f-4c7d-9d2b-115d492961f1/util/0.log" Nov 25 11:33:35 crc kubenswrapper[4821]: I1125 11:33:35.556117 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6hsmqh_8c67f775-e13f-4c7d-9d2b-115d492961f1/extract/0.log" Nov 25 11:33:35 crc kubenswrapper[4821]: I1125 11:33:35.576892 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6hsmqh_8c67f775-e13f-4c7d-9d2b-115d492961f1/pull/0.log" Nov 25 11:33:35 crc kubenswrapper[4821]: I1125 11:33:35.650336 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4j8hk_a43c28e0-8883-4567-a86f-bcd311098876/registry-server/0.log" Nov 25 11:33:35 crc kubenswrapper[4821]: I1125 11:33:35.765307 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-tdctv_05106897-6fd0-462c-9d26-a832f1385e04/marketplace-operator/0.log" Nov 25 11:33:35 crc kubenswrapper[4821]: I1125 11:33:35.856175 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m99sj_8a0df682-fa69-4ef1-a733-214abe72f1ad/extract-utilities/0.log" Nov 25 11:33:36 crc kubenswrapper[4821]: I1125 11:33:36.066597 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m99sj_8a0df682-fa69-4ef1-a733-214abe72f1ad/extract-content/0.log" Nov 25 11:33:36 crc kubenswrapper[4821]: I1125 11:33:36.069818 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m99sj_8a0df682-fa69-4ef1-a733-214abe72f1ad/extract-content/0.log" Nov 25 11:33:36 crc kubenswrapper[4821]: I1125 11:33:36.071736 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m99sj_8a0df682-fa69-4ef1-a733-214abe72f1ad/extract-utilities/0.log" Nov 25 11:33:36 crc kubenswrapper[4821]: I1125 11:33:36.224309 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m99sj_8a0df682-fa69-4ef1-a733-214abe72f1ad/extract-utilities/0.log" Nov 25 11:33:36 crc kubenswrapper[4821]: I1125 11:33:36.256733 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m99sj_8a0df682-fa69-4ef1-a733-214abe72f1ad/extract-content/0.log" Nov 25 11:33:36 crc kubenswrapper[4821]: I1125 11:33:36.415923 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-vrjfk_7b01ce52-a13b-431c-8597-fbd5dc25183f/extract-utilities/0.log" Nov 25 11:33:36 crc kubenswrapper[4821]: I1125 11:33:36.457342 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m99sj_8a0df682-fa69-4ef1-a733-214abe72f1ad/registry-server/0.log" Nov 25 11:33:36 crc kubenswrapper[4821]: I1125 11:33:36.662318 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-vrjfk_7b01ce52-a13b-431c-8597-fbd5dc25183f/extract-utilities/0.log" Nov 25 11:33:36 crc kubenswrapper[4821]: I1125 11:33:36.692307 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-vrjfk_7b01ce52-a13b-431c-8597-fbd5dc25183f/extract-content/0.log" Nov 25 11:33:36 crc kubenswrapper[4821]: I1125 11:33:36.692550 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-vrjfk_7b01ce52-a13b-431c-8597-fbd5dc25183f/extract-content/0.log" Nov 25 11:33:36 crc kubenswrapper[4821]: I1125 11:33:36.904118 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-vrjfk_7b01ce52-a13b-431c-8597-fbd5dc25183f/extract-content/0.log" Nov 25 11:33:36 crc kubenswrapper[4821]: I1125 11:33:36.925553 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-vrjfk_7b01ce52-a13b-431c-8597-fbd5dc25183f/extract-utilities/0.log" Nov 25 11:33:37 crc kubenswrapper[4821]: I1125 11:33:37.407863 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-vrjfk_7b01ce52-a13b-431c-8597-fbd5dc25183f/registry-server/0.log" Nov 25 11:34:04 crc kubenswrapper[4821]: I1125 11:34:04.714945 4821 patch_prober.go:28] interesting pod/machine-config-daemon-2krbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:34:04 crc kubenswrapper[4821]: I1125 11:34:04.715476 4821 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:34:34 crc kubenswrapper[4821]: I1125 11:34:34.715130 4821 patch_prober.go:28] interesting pod/machine-config-daemon-2krbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:34:34 crc kubenswrapper[4821]: I1125 11:34:34.715804 4821 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:35:04 crc kubenswrapper[4821]: I1125 11:35:04.714442 4821 patch_prober.go:28] interesting pod/machine-config-daemon-2krbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:35:04 crc kubenswrapper[4821]: I1125 11:35:04.715220 4821 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:35:04 crc kubenswrapper[4821]: I1125 11:35:04.715310 4821 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" Nov 25 11:35:04 crc kubenswrapper[4821]: I1125 11:35:04.716687 4821 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f357cf6563c37c0c7cd86102a3e1eea3ba9146a525631fc115bab1d95f4015f3"} pod="openshift-machine-config-operator/machine-config-daemon-2krbf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 11:35:04 crc kubenswrapper[4821]: I1125 11:35:04.716808 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerName="machine-config-daemon" containerID="cri-o://f357cf6563c37c0c7cd86102a3e1eea3ba9146a525631fc115bab1d95f4015f3" gracePeriod=600 Nov 25 11:35:05 crc kubenswrapper[4821]: E1125 11:35:05.380592 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:35:05 crc kubenswrapper[4821]: I1125 11:35:05.563089 4821 generic.go:334] "Generic (PLEG): container finished" podID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerID="f357cf6563c37c0c7cd86102a3e1eea3ba9146a525631fc115bab1d95f4015f3" exitCode=0 Nov 25 11:35:05 crc kubenswrapper[4821]: I1125 11:35:05.563139 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" event={"ID":"5f948b87-ac86-4de6-ad64-c2ef947f84d4","Type":"ContainerDied","Data":"f357cf6563c37c0c7cd86102a3e1eea3ba9146a525631fc115bab1d95f4015f3"} Nov 25 11:35:05 crc kubenswrapper[4821]: I1125 11:35:05.563202 4821 scope.go:117] "RemoveContainer" containerID="52d1a350addd03a9e1e92a65fe40ee197504bf6db7f5b0601ca8860f562470fb" Nov 25 11:35:05 crc kubenswrapper[4821]: I1125 11:35:05.567372 4821 scope.go:117] "RemoveContainer" containerID="f357cf6563c37c0c7cd86102a3e1eea3ba9146a525631fc115bab1d95f4015f3" Nov 25 11:35:05 crc kubenswrapper[4821]: E1125 11:35:05.567799 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:35:18 crc kubenswrapper[4821]: I1125 11:35:18.115445 4821 scope.go:117] "RemoveContainer" containerID="f357cf6563c37c0c7cd86102a3e1eea3ba9146a525631fc115bab1d95f4015f3" Nov 25 11:35:18 crc kubenswrapper[4821]: E1125 11:35:18.116294 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:35:21 crc kubenswrapper[4821]: I1125 11:35:21.706952 4821 generic.go:334] "Generic (PLEG): container finished" podID="d93b17f1-e742-45a7-a5ec-141e52a382c2" containerID="81bc3cd2f14c92f9509d803c926854123e6f0f440a6479ebd4c24da154176cae" exitCode=0 Nov 25 11:35:21 crc kubenswrapper[4821]: I1125 11:35:21.707056 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9zwfm/must-gather-kmj2h" event={"ID":"d93b17f1-e742-45a7-a5ec-141e52a382c2","Type":"ContainerDied","Data":"81bc3cd2f14c92f9509d803c926854123e6f0f440a6479ebd4c24da154176cae"} Nov 25 11:35:21 crc kubenswrapper[4821]: I1125 11:35:21.708102 4821 scope.go:117] "RemoveContainer" containerID="81bc3cd2f14c92f9509d803c926854123e6f0f440a6479ebd4c24da154176cae" Nov 25 11:35:21 crc kubenswrapper[4821]: I1125 11:35:21.821033 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-9zwfm_must-gather-kmj2h_d93b17f1-e742-45a7-a5ec-141e52a382c2/gather/0.log" Nov 25 11:35:29 crc kubenswrapper[4821]: I1125 11:35:29.071898 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-9zwfm/must-gather-kmj2h"] Nov 25 11:35:29 crc kubenswrapper[4821]: I1125 11:35:29.072955 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-9zwfm/must-gather-kmj2h" podUID="d93b17f1-e742-45a7-a5ec-141e52a382c2" containerName="copy" containerID="cri-o://3e1b6937222640bbbef56eb7665a33f444c4d41c3ebe1880942f498bf2797891" gracePeriod=2 Nov 25 11:35:29 crc kubenswrapper[4821]: I1125 11:35:29.086021 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-9zwfm/must-gather-kmj2h"] Nov 25 11:35:29 crc kubenswrapper[4821]: I1125 11:35:29.526987 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-9zwfm_must-gather-kmj2h_d93b17f1-e742-45a7-a5ec-141e52a382c2/copy/0.log" Nov 25 11:35:29 crc kubenswrapper[4821]: I1125 11:35:29.527882 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9zwfm/must-gather-kmj2h" Nov 25 11:35:29 crc kubenswrapper[4821]: I1125 11:35:29.608426 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d93b17f1-e742-45a7-a5ec-141e52a382c2-must-gather-output\") pod \"d93b17f1-e742-45a7-a5ec-141e52a382c2\" (UID: \"d93b17f1-e742-45a7-a5ec-141e52a382c2\") " Nov 25 11:35:29 crc kubenswrapper[4821]: I1125 11:35:29.608568 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wghn6\" (UniqueName: \"kubernetes.io/projected/d93b17f1-e742-45a7-a5ec-141e52a382c2-kube-api-access-wghn6\") pod \"d93b17f1-e742-45a7-a5ec-141e52a382c2\" (UID: \"d93b17f1-e742-45a7-a5ec-141e52a382c2\") " Nov 25 11:35:29 crc kubenswrapper[4821]: I1125 11:35:29.620488 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d93b17f1-e742-45a7-a5ec-141e52a382c2-kube-api-access-wghn6" (OuterVolumeSpecName: "kube-api-access-wghn6") pod "d93b17f1-e742-45a7-a5ec-141e52a382c2" (UID: "d93b17f1-e742-45a7-a5ec-141e52a382c2"). InnerVolumeSpecName "kube-api-access-wghn6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:35:29 crc kubenswrapper[4821]: I1125 11:35:29.712478 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wghn6\" (UniqueName: \"kubernetes.io/projected/d93b17f1-e742-45a7-a5ec-141e52a382c2-kube-api-access-wghn6\") on node \"crc\" DevicePath \"\"" Nov 25 11:35:29 crc kubenswrapper[4821]: I1125 11:35:29.768723 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d93b17f1-e742-45a7-a5ec-141e52a382c2-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "d93b17f1-e742-45a7-a5ec-141e52a382c2" (UID: "d93b17f1-e742-45a7-a5ec-141e52a382c2"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:35:29 crc kubenswrapper[4821]: I1125 11:35:29.810796 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-9zwfm_must-gather-kmj2h_d93b17f1-e742-45a7-a5ec-141e52a382c2/copy/0.log" Nov 25 11:35:29 crc kubenswrapper[4821]: I1125 11:35:29.811425 4821 generic.go:334] "Generic (PLEG): container finished" podID="d93b17f1-e742-45a7-a5ec-141e52a382c2" containerID="3e1b6937222640bbbef56eb7665a33f444c4d41c3ebe1880942f498bf2797891" exitCode=143 Nov 25 11:35:29 crc kubenswrapper[4821]: I1125 11:35:29.811514 4821 scope.go:117] "RemoveContainer" containerID="3e1b6937222640bbbef56eb7665a33f444c4d41c3ebe1880942f498bf2797891" Nov 25 11:35:29 crc kubenswrapper[4821]: I1125 11:35:29.811530 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9zwfm/must-gather-kmj2h" Nov 25 11:35:29 crc kubenswrapper[4821]: I1125 11:35:29.814419 4821 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d93b17f1-e742-45a7-a5ec-141e52a382c2-must-gather-output\") on node \"crc\" DevicePath \"\"" Nov 25 11:35:29 crc kubenswrapper[4821]: I1125 11:35:29.838489 4821 scope.go:117] "RemoveContainer" containerID="81bc3cd2f14c92f9509d803c926854123e6f0f440a6479ebd4c24da154176cae" Nov 25 11:35:29 crc kubenswrapper[4821]: I1125 11:35:29.906697 4821 scope.go:117] "RemoveContainer" containerID="3e1b6937222640bbbef56eb7665a33f444c4d41c3ebe1880942f498bf2797891" Nov 25 11:35:29 crc kubenswrapper[4821]: E1125 11:35:29.907250 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e1b6937222640bbbef56eb7665a33f444c4d41c3ebe1880942f498bf2797891\": container with ID starting with 3e1b6937222640bbbef56eb7665a33f444c4d41c3ebe1880942f498bf2797891 not found: ID does not exist" containerID="3e1b6937222640bbbef56eb7665a33f444c4d41c3ebe1880942f498bf2797891" Nov 25 11:35:29 crc kubenswrapper[4821]: I1125 11:35:29.907302 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e1b6937222640bbbef56eb7665a33f444c4d41c3ebe1880942f498bf2797891"} err="failed to get container status \"3e1b6937222640bbbef56eb7665a33f444c4d41c3ebe1880942f498bf2797891\": rpc error: code = NotFound desc = could not find container \"3e1b6937222640bbbef56eb7665a33f444c4d41c3ebe1880942f498bf2797891\": container with ID starting with 3e1b6937222640bbbef56eb7665a33f444c4d41c3ebe1880942f498bf2797891 not found: ID does not exist" Nov 25 11:35:29 crc kubenswrapper[4821]: I1125 11:35:29.907375 4821 scope.go:117] "RemoveContainer" containerID="81bc3cd2f14c92f9509d803c926854123e6f0f440a6479ebd4c24da154176cae" Nov 25 11:35:29 crc kubenswrapper[4821]: E1125 11:35:29.907966 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"81bc3cd2f14c92f9509d803c926854123e6f0f440a6479ebd4c24da154176cae\": container with ID starting with 81bc3cd2f14c92f9509d803c926854123e6f0f440a6479ebd4c24da154176cae not found: ID does not exist" containerID="81bc3cd2f14c92f9509d803c926854123e6f0f440a6479ebd4c24da154176cae" Nov 25 11:35:29 crc kubenswrapper[4821]: I1125 11:35:29.908014 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81bc3cd2f14c92f9509d803c926854123e6f0f440a6479ebd4c24da154176cae"} err="failed to get container status \"81bc3cd2f14c92f9509d803c926854123e6f0f440a6479ebd4c24da154176cae\": rpc error: code = NotFound desc = could not find container \"81bc3cd2f14c92f9509d803c926854123e6f0f440a6479ebd4c24da154176cae\": container with ID starting with 81bc3cd2f14c92f9509d803c926854123e6f0f440a6479ebd4c24da154176cae not found: ID does not exist" Nov 25 11:35:30 crc kubenswrapper[4821]: I1125 11:35:30.130810 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d93b17f1-e742-45a7-a5ec-141e52a382c2" path="/var/lib/kubelet/pods/d93b17f1-e742-45a7-a5ec-141e52a382c2/volumes" Nov 25 11:35:31 crc kubenswrapper[4821]: I1125 11:35:31.115726 4821 scope.go:117] "RemoveContainer" containerID="f357cf6563c37c0c7cd86102a3e1eea3ba9146a525631fc115bab1d95f4015f3" Nov 25 11:35:31 crc kubenswrapper[4821]: E1125 11:35:31.116595 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:35:46 crc kubenswrapper[4821]: I1125 11:35:46.115113 4821 scope.go:117] "RemoveContainer" containerID="f357cf6563c37c0c7cd86102a3e1eea3ba9146a525631fc115bab1d95f4015f3" Nov 25 11:35:46 crc kubenswrapper[4821]: E1125 11:35:46.115954 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:35:58 crc kubenswrapper[4821]: I1125 11:35:58.114624 4821 scope.go:117] "RemoveContainer" containerID="f357cf6563c37c0c7cd86102a3e1eea3ba9146a525631fc115bab1d95f4015f3" Nov 25 11:35:58 crc kubenswrapper[4821]: E1125 11:35:58.115385 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:36:12 crc kubenswrapper[4821]: I1125 11:36:12.115025 4821 scope.go:117] "RemoveContainer" containerID="f357cf6563c37c0c7cd86102a3e1eea3ba9146a525631fc115bab1d95f4015f3" Nov 25 11:36:12 crc kubenswrapper[4821]: E1125 11:36:12.115774 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:36:23 crc kubenswrapper[4821]: I1125 11:36:23.113928 4821 scope.go:117] "RemoveContainer" containerID="f357cf6563c37c0c7cd86102a3e1eea3ba9146a525631fc115bab1d95f4015f3" Nov 25 11:36:23 crc kubenswrapper[4821]: E1125 11:36:23.114697 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:36:36 crc kubenswrapper[4821]: I1125 11:36:36.114383 4821 scope.go:117] "RemoveContainer" containerID="f357cf6563c37c0c7cd86102a3e1eea3ba9146a525631fc115bab1d95f4015f3" Nov 25 11:36:36 crc kubenswrapper[4821]: E1125 11:36:36.115104 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:36:50 crc kubenswrapper[4821]: I1125 11:36:50.122416 4821 scope.go:117] "RemoveContainer" containerID="f357cf6563c37c0c7cd86102a3e1eea3ba9146a525631fc115bab1d95f4015f3" Nov 25 11:36:50 crc kubenswrapper[4821]: E1125 11:36:50.123262 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:37:04 crc kubenswrapper[4821]: I1125 11:37:04.114397 4821 scope.go:117] "RemoveContainer" containerID="f357cf6563c37c0c7cd86102a3e1eea3ba9146a525631fc115bab1d95f4015f3" Nov 25 11:37:04 crc kubenswrapper[4821]: E1125 11:37:04.115537 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:37:16 crc kubenswrapper[4821]: I1125 11:37:16.113949 4821 scope.go:117] "RemoveContainer" containerID="f357cf6563c37c0c7cd86102a3e1eea3ba9146a525631fc115bab1d95f4015f3" Nov 25 11:37:16 crc kubenswrapper[4821]: E1125 11:37:16.114786 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:37:26 crc kubenswrapper[4821]: I1125 11:37:26.007456 4821 scope.go:117] "RemoveContainer" containerID="3f75c1323e9159aa66657b856a5905d399ef8b6c10e67b18c77e54c8f6ef5ceb" Nov 25 11:37:31 crc kubenswrapper[4821]: I1125 11:37:31.115320 4821 scope.go:117] "RemoveContainer" containerID="f357cf6563c37c0c7cd86102a3e1eea3ba9146a525631fc115bab1d95f4015f3" Nov 25 11:37:31 crc kubenswrapper[4821]: E1125 11:37:31.116207 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:37:44 crc kubenswrapper[4821]: I1125 11:37:44.115412 4821 scope.go:117] "RemoveContainer" containerID="f357cf6563c37c0c7cd86102a3e1eea3ba9146a525631fc115bab1d95f4015f3" Nov 25 11:37:44 crc kubenswrapper[4821]: E1125 11:37:44.116655 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:37:56 crc kubenswrapper[4821]: I1125 11:37:56.998071 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-j4bh5"] Nov 25 11:37:57 crc kubenswrapper[4821]: E1125 11:37:56.999122 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d93b17f1-e742-45a7-a5ec-141e52a382c2" containerName="copy" Nov 25 11:37:57 crc kubenswrapper[4821]: I1125 11:37:56.999134 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="d93b17f1-e742-45a7-a5ec-141e52a382c2" containerName="copy" Nov 25 11:37:57 crc kubenswrapper[4821]: E1125 11:37:56.999181 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdd5fc1c-bf87-4bb1-9834-93e2bef9e773" containerName="container-00" Nov 25 11:37:57 crc kubenswrapper[4821]: I1125 11:37:56.999187 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdd5fc1c-bf87-4bb1-9834-93e2bef9e773" containerName="container-00" Nov 25 11:37:57 crc kubenswrapper[4821]: E1125 11:37:56.999209 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d93b17f1-e742-45a7-a5ec-141e52a382c2" containerName="gather" Nov 25 11:37:57 crc kubenswrapper[4821]: I1125 11:37:56.999215 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="d93b17f1-e742-45a7-a5ec-141e52a382c2" containerName="gather" Nov 25 11:37:57 crc kubenswrapper[4821]: I1125 11:37:57.002205 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="d93b17f1-e742-45a7-a5ec-141e52a382c2" containerName="gather" Nov 25 11:37:57 crc kubenswrapper[4821]: I1125 11:37:57.002238 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="d93b17f1-e742-45a7-a5ec-141e52a382c2" containerName="copy" Nov 25 11:37:57 crc kubenswrapper[4821]: I1125 11:37:57.002263 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="cdd5fc1c-bf87-4bb1-9834-93e2bef9e773" containerName="container-00" Nov 25 11:37:57 crc kubenswrapper[4821]: I1125 11:37:57.003747 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-j4bh5" Nov 25 11:37:57 crc kubenswrapper[4821]: I1125 11:37:57.021351 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-j4bh5"] Nov 25 11:37:57 crc kubenswrapper[4821]: I1125 11:37:57.098857 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f2ad157-aa04-48bb-bdb8-2364f7127c34-utilities\") pod \"redhat-marketplace-j4bh5\" (UID: \"2f2ad157-aa04-48bb-bdb8-2364f7127c34\") " pod="openshift-marketplace/redhat-marketplace-j4bh5" Nov 25 11:37:57 crc kubenswrapper[4821]: I1125 11:37:57.098932 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f2ad157-aa04-48bb-bdb8-2364f7127c34-catalog-content\") pod \"redhat-marketplace-j4bh5\" (UID: \"2f2ad157-aa04-48bb-bdb8-2364f7127c34\") " pod="openshift-marketplace/redhat-marketplace-j4bh5" Nov 25 11:37:57 crc kubenswrapper[4821]: I1125 11:37:57.099361 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m5z46\" (UniqueName: \"kubernetes.io/projected/2f2ad157-aa04-48bb-bdb8-2364f7127c34-kube-api-access-m5z46\") pod \"redhat-marketplace-j4bh5\" (UID: \"2f2ad157-aa04-48bb-bdb8-2364f7127c34\") " pod="openshift-marketplace/redhat-marketplace-j4bh5" Nov 25 11:37:57 crc kubenswrapper[4821]: I1125 11:37:57.114266 4821 scope.go:117] "RemoveContainer" containerID="f357cf6563c37c0c7cd86102a3e1eea3ba9146a525631fc115bab1d95f4015f3" Nov 25 11:37:57 crc kubenswrapper[4821]: E1125 11:37:57.114655 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:37:57 crc kubenswrapper[4821]: I1125 11:37:57.200678 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m5z46\" (UniqueName: \"kubernetes.io/projected/2f2ad157-aa04-48bb-bdb8-2364f7127c34-kube-api-access-m5z46\") pod \"redhat-marketplace-j4bh5\" (UID: \"2f2ad157-aa04-48bb-bdb8-2364f7127c34\") " pod="openshift-marketplace/redhat-marketplace-j4bh5" Nov 25 11:37:57 crc kubenswrapper[4821]: I1125 11:37:57.200828 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f2ad157-aa04-48bb-bdb8-2364f7127c34-utilities\") pod \"redhat-marketplace-j4bh5\" (UID: \"2f2ad157-aa04-48bb-bdb8-2364f7127c34\") " pod="openshift-marketplace/redhat-marketplace-j4bh5" Nov 25 11:37:57 crc kubenswrapper[4821]: I1125 11:37:57.200858 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f2ad157-aa04-48bb-bdb8-2364f7127c34-catalog-content\") pod \"redhat-marketplace-j4bh5\" (UID: \"2f2ad157-aa04-48bb-bdb8-2364f7127c34\") " pod="openshift-marketplace/redhat-marketplace-j4bh5" Nov 25 11:37:57 crc kubenswrapper[4821]: I1125 11:37:57.201518 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f2ad157-aa04-48bb-bdb8-2364f7127c34-catalog-content\") pod \"redhat-marketplace-j4bh5\" (UID: \"2f2ad157-aa04-48bb-bdb8-2364f7127c34\") " pod="openshift-marketplace/redhat-marketplace-j4bh5" Nov 25 11:37:57 crc kubenswrapper[4821]: I1125 11:37:57.201874 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f2ad157-aa04-48bb-bdb8-2364f7127c34-utilities\") pod \"redhat-marketplace-j4bh5\" (UID: \"2f2ad157-aa04-48bb-bdb8-2364f7127c34\") " pod="openshift-marketplace/redhat-marketplace-j4bh5" Nov 25 11:37:57 crc kubenswrapper[4821]: I1125 11:37:57.339558 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m5z46\" (UniqueName: \"kubernetes.io/projected/2f2ad157-aa04-48bb-bdb8-2364f7127c34-kube-api-access-m5z46\") pod \"redhat-marketplace-j4bh5\" (UID: \"2f2ad157-aa04-48bb-bdb8-2364f7127c34\") " pod="openshift-marketplace/redhat-marketplace-j4bh5" Nov 25 11:37:57 crc kubenswrapper[4821]: I1125 11:37:57.634487 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-j4bh5" Nov 25 11:37:58 crc kubenswrapper[4821]: I1125 11:37:58.080032 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-j4bh5"] Nov 25 11:37:58 crc kubenswrapper[4821]: I1125 11:37:58.168874 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j4bh5" event={"ID":"2f2ad157-aa04-48bb-bdb8-2364f7127c34","Type":"ContainerStarted","Data":"cd8c86fc71b9a1f3434a9fb769305407dc1ad0bd2c97c1b9ca6c982dd75f2e4c"} Nov 25 11:37:59 crc kubenswrapper[4821]: I1125 11:37:59.183734 4821 generic.go:334] "Generic (PLEG): container finished" podID="2f2ad157-aa04-48bb-bdb8-2364f7127c34" containerID="b22ebf569a598bc380dfeabcd816c8349000f68d190ebb4e0b4c5fde78bf353f" exitCode=0 Nov 25 11:37:59 crc kubenswrapper[4821]: I1125 11:37:59.183866 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j4bh5" event={"ID":"2f2ad157-aa04-48bb-bdb8-2364f7127c34","Type":"ContainerDied","Data":"b22ebf569a598bc380dfeabcd816c8349000f68d190ebb4e0b4c5fde78bf353f"} Nov 25 11:37:59 crc kubenswrapper[4821]: I1125 11:37:59.186044 4821 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 11:38:00 crc kubenswrapper[4821]: I1125 11:38:00.199035 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j4bh5" event={"ID":"2f2ad157-aa04-48bb-bdb8-2364f7127c34","Type":"ContainerStarted","Data":"6531d282b49e90d86bd9ad77b1847d249afa20bd6711937d08524ae915ff58e4"} Nov 25 11:38:01 crc kubenswrapper[4821]: I1125 11:38:01.210578 4821 generic.go:334] "Generic (PLEG): container finished" podID="2f2ad157-aa04-48bb-bdb8-2364f7127c34" containerID="6531d282b49e90d86bd9ad77b1847d249afa20bd6711937d08524ae915ff58e4" exitCode=0 Nov 25 11:38:01 crc kubenswrapper[4821]: I1125 11:38:01.210681 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j4bh5" event={"ID":"2f2ad157-aa04-48bb-bdb8-2364f7127c34","Type":"ContainerDied","Data":"6531d282b49e90d86bd9ad77b1847d249afa20bd6711937d08524ae915ff58e4"} Nov 25 11:38:02 crc kubenswrapper[4821]: I1125 11:38:02.222863 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j4bh5" event={"ID":"2f2ad157-aa04-48bb-bdb8-2364f7127c34","Type":"ContainerStarted","Data":"cb58faeccc9e182e2c5677e3fa925cac3a35f52db5a961902531657890af3c6b"} Nov 25 11:38:02 crc kubenswrapper[4821]: I1125 11:38:02.253056 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-j4bh5" podStartSLOduration=3.694149623 podStartE2EDuration="6.253033081s" podCreationTimestamp="2025-11-25 11:37:56 +0000 UTC" firstStartedPulling="2025-11-25 11:37:59.185799568 +0000 UTC m=+3949.722119415" lastFinishedPulling="2025-11-25 11:38:01.744683026 +0000 UTC m=+3952.281002873" observedRunningTime="2025-11-25 11:38:02.243881149 +0000 UTC m=+3952.780201006" watchObservedRunningTime="2025-11-25 11:38:02.253033081 +0000 UTC m=+3952.789352928" Nov 25 11:38:02 crc kubenswrapper[4821]: I1125 11:38:02.375075 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-gdrrz/must-gather-xxxft"] Nov 25 11:38:02 crc kubenswrapper[4821]: I1125 11:38:02.377356 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gdrrz/must-gather-xxxft" Nov 25 11:38:02 crc kubenswrapper[4821]: I1125 11:38:02.390322 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-gdrrz"/"openshift-service-ca.crt" Nov 25 11:38:02 crc kubenswrapper[4821]: I1125 11:38:02.390411 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-gdrrz"/"kube-root-ca.crt" Nov 25 11:38:02 crc kubenswrapper[4821]: I1125 11:38:02.398739 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-gdrrz/must-gather-xxxft"] Nov 25 11:38:02 crc kubenswrapper[4821]: I1125 11:38:02.399575 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vnh5p\" (UniqueName: \"kubernetes.io/projected/61c863b6-7476-4c5d-b59c-40735554a04e-kube-api-access-vnh5p\") pod \"must-gather-xxxft\" (UID: \"61c863b6-7476-4c5d-b59c-40735554a04e\") " pod="openshift-must-gather-gdrrz/must-gather-xxxft" Nov 25 11:38:02 crc kubenswrapper[4821]: I1125 11:38:02.399653 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/61c863b6-7476-4c5d-b59c-40735554a04e-must-gather-output\") pod \"must-gather-xxxft\" (UID: \"61c863b6-7476-4c5d-b59c-40735554a04e\") " pod="openshift-must-gather-gdrrz/must-gather-xxxft" Nov 25 11:38:02 crc kubenswrapper[4821]: I1125 11:38:02.502549 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/61c863b6-7476-4c5d-b59c-40735554a04e-must-gather-output\") pod \"must-gather-xxxft\" (UID: \"61c863b6-7476-4c5d-b59c-40735554a04e\") " pod="openshift-must-gather-gdrrz/must-gather-xxxft" Nov 25 11:38:02 crc kubenswrapper[4821]: I1125 11:38:02.502714 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vnh5p\" (UniqueName: \"kubernetes.io/projected/61c863b6-7476-4c5d-b59c-40735554a04e-kube-api-access-vnh5p\") pod \"must-gather-xxxft\" (UID: \"61c863b6-7476-4c5d-b59c-40735554a04e\") " pod="openshift-must-gather-gdrrz/must-gather-xxxft" Nov 25 11:38:02 crc kubenswrapper[4821]: I1125 11:38:02.503000 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/61c863b6-7476-4c5d-b59c-40735554a04e-must-gather-output\") pod \"must-gather-xxxft\" (UID: \"61c863b6-7476-4c5d-b59c-40735554a04e\") " pod="openshift-must-gather-gdrrz/must-gather-xxxft" Nov 25 11:38:02 crc kubenswrapper[4821]: I1125 11:38:02.527850 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vnh5p\" (UniqueName: \"kubernetes.io/projected/61c863b6-7476-4c5d-b59c-40735554a04e-kube-api-access-vnh5p\") pod \"must-gather-xxxft\" (UID: \"61c863b6-7476-4c5d-b59c-40735554a04e\") " pod="openshift-must-gather-gdrrz/must-gather-xxxft" Nov 25 11:38:02 crc kubenswrapper[4821]: I1125 11:38:02.700746 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gdrrz/must-gather-xxxft" Nov 25 11:38:03 crc kubenswrapper[4821]: I1125 11:38:03.210907 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-gdrrz/must-gather-xxxft"] Nov 25 11:38:03 crc kubenswrapper[4821]: W1125 11:38:03.227758 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod61c863b6_7476_4c5d_b59c_40735554a04e.slice/crio-f864f760e039a68a0855e598d1bb8bf9b4be37544591a44a32f81d80e765d9a8 WatchSource:0}: Error finding container f864f760e039a68a0855e598d1bb8bf9b4be37544591a44a32f81d80e765d9a8: Status 404 returned error can't find the container with id f864f760e039a68a0855e598d1bb8bf9b4be37544591a44a32f81d80e765d9a8 Nov 25 11:38:04 crc kubenswrapper[4821]: I1125 11:38:04.239446 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gdrrz/must-gather-xxxft" event={"ID":"61c863b6-7476-4c5d-b59c-40735554a04e","Type":"ContainerStarted","Data":"0e9e48de0f9fa74d65eb2152de4e317c5a1c3e27a8059c7bc775e2fe6c49ca7a"} Nov 25 11:38:04 crc kubenswrapper[4821]: I1125 11:38:04.239506 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gdrrz/must-gather-xxxft" event={"ID":"61c863b6-7476-4c5d-b59c-40735554a04e","Type":"ContainerStarted","Data":"03c70625987368ad0a26a2cc86962944984ace1e191066e2b4ad1293b5d73adc"} Nov 25 11:38:04 crc kubenswrapper[4821]: I1125 11:38:04.239520 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gdrrz/must-gather-xxxft" event={"ID":"61c863b6-7476-4c5d-b59c-40735554a04e","Type":"ContainerStarted","Data":"f864f760e039a68a0855e598d1bb8bf9b4be37544591a44a32f81d80e765d9a8"} Nov 25 11:38:04 crc kubenswrapper[4821]: I1125 11:38:04.261670 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-gdrrz/must-gather-xxxft" podStartSLOduration=2.261653559 podStartE2EDuration="2.261653559s" podCreationTimestamp="2025-11-25 11:38:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:38:04.256194136 +0000 UTC m=+3954.792513983" watchObservedRunningTime="2025-11-25 11:38:04.261653559 +0000 UTC m=+3954.797973406" Nov 25 11:38:06 crc kubenswrapper[4821]: E1125 11:38:06.993092 4821 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.129.56.252:39514->38.129.56.252:37639: write tcp 38.129.56.252:39514->38.129.56.252:37639: write: broken pipe Nov 25 11:38:07 crc kubenswrapper[4821]: I1125 11:38:07.401398 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-gdrrz/crc-debug-mwt2q"] Nov 25 11:38:07 crc kubenswrapper[4821]: I1125 11:38:07.402915 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gdrrz/crc-debug-mwt2q" Nov 25 11:38:07 crc kubenswrapper[4821]: I1125 11:38:07.405151 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-gdrrz"/"default-dockercfg-8kkz2" Nov 25 11:38:07 crc kubenswrapper[4821]: I1125 11:38:07.487390 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4ea200a7-e3dc-42cf-b4a5-51e45bcec116-host\") pod \"crc-debug-mwt2q\" (UID: \"4ea200a7-e3dc-42cf-b4a5-51e45bcec116\") " pod="openshift-must-gather-gdrrz/crc-debug-mwt2q" Nov 25 11:38:07 crc kubenswrapper[4821]: I1125 11:38:07.487729 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8wzh8\" (UniqueName: \"kubernetes.io/projected/4ea200a7-e3dc-42cf-b4a5-51e45bcec116-kube-api-access-8wzh8\") pod \"crc-debug-mwt2q\" (UID: \"4ea200a7-e3dc-42cf-b4a5-51e45bcec116\") " pod="openshift-must-gather-gdrrz/crc-debug-mwt2q" Nov 25 11:38:07 crc kubenswrapper[4821]: I1125 11:38:07.589660 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8wzh8\" (UniqueName: \"kubernetes.io/projected/4ea200a7-e3dc-42cf-b4a5-51e45bcec116-kube-api-access-8wzh8\") pod \"crc-debug-mwt2q\" (UID: \"4ea200a7-e3dc-42cf-b4a5-51e45bcec116\") " pod="openshift-must-gather-gdrrz/crc-debug-mwt2q" Nov 25 11:38:07 crc kubenswrapper[4821]: I1125 11:38:07.589808 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4ea200a7-e3dc-42cf-b4a5-51e45bcec116-host\") pod \"crc-debug-mwt2q\" (UID: \"4ea200a7-e3dc-42cf-b4a5-51e45bcec116\") " pod="openshift-must-gather-gdrrz/crc-debug-mwt2q" Nov 25 11:38:07 crc kubenswrapper[4821]: I1125 11:38:07.589976 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4ea200a7-e3dc-42cf-b4a5-51e45bcec116-host\") pod \"crc-debug-mwt2q\" (UID: \"4ea200a7-e3dc-42cf-b4a5-51e45bcec116\") " pod="openshift-must-gather-gdrrz/crc-debug-mwt2q" Nov 25 11:38:07 crc kubenswrapper[4821]: I1125 11:38:07.632538 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8wzh8\" (UniqueName: \"kubernetes.io/projected/4ea200a7-e3dc-42cf-b4a5-51e45bcec116-kube-api-access-8wzh8\") pod \"crc-debug-mwt2q\" (UID: \"4ea200a7-e3dc-42cf-b4a5-51e45bcec116\") " pod="openshift-must-gather-gdrrz/crc-debug-mwt2q" Nov 25 11:38:07 crc kubenswrapper[4821]: I1125 11:38:07.635503 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-j4bh5" Nov 25 11:38:07 crc kubenswrapper[4821]: I1125 11:38:07.635560 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-j4bh5" Nov 25 11:38:07 crc kubenswrapper[4821]: I1125 11:38:07.687294 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-j4bh5" Nov 25 11:38:07 crc kubenswrapper[4821]: I1125 11:38:07.737603 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gdrrz/crc-debug-mwt2q" Nov 25 11:38:08 crc kubenswrapper[4821]: I1125 11:38:08.276418 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gdrrz/crc-debug-mwt2q" event={"ID":"4ea200a7-e3dc-42cf-b4a5-51e45bcec116","Type":"ContainerStarted","Data":"49bc7c06039be18e3024a8625707ddd6ccb9e5d76c93fa103fc1a4f3f44e7bcf"} Nov 25 11:38:08 crc kubenswrapper[4821]: I1125 11:38:08.276807 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gdrrz/crc-debug-mwt2q" event={"ID":"4ea200a7-e3dc-42cf-b4a5-51e45bcec116","Type":"ContainerStarted","Data":"5fad0c4cbe751b497435cf5c1e5ae0b839a139ee794c804c70f7d872a7e7bed8"} Nov 25 11:38:08 crc kubenswrapper[4821]: I1125 11:38:08.298335 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-gdrrz/crc-debug-mwt2q" podStartSLOduration=1.2983157570000001 podStartE2EDuration="1.298315757s" podCreationTimestamp="2025-11-25 11:38:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:38:08.290304312 +0000 UTC m=+3958.826624179" watchObservedRunningTime="2025-11-25 11:38:08.298315757 +0000 UTC m=+3958.834635614" Nov 25 11:38:08 crc kubenswrapper[4821]: I1125 11:38:08.337499 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-j4bh5" Nov 25 11:38:08 crc kubenswrapper[4821]: I1125 11:38:08.400130 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-j4bh5"] Nov 25 11:38:10 crc kubenswrapper[4821]: I1125 11:38:10.292260 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-j4bh5" podUID="2f2ad157-aa04-48bb-bdb8-2364f7127c34" containerName="registry-server" containerID="cri-o://cb58faeccc9e182e2c5677e3fa925cac3a35f52db5a961902531657890af3c6b" gracePeriod=2 Nov 25 11:38:10 crc kubenswrapper[4821]: I1125 11:38:10.732696 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-j4bh5" Nov 25 11:38:10 crc kubenswrapper[4821]: I1125 11:38:10.847295 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f2ad157-aa04-48bb-bdb8-2364f7127c34-utilities\") pod \"2f2ad157-aa04-48bb-bdb8-2364f7127c34\" (UID: \"2f2ad157-aa04-48bb-bdb8-2364f7127c34\") " Nov 25 11:38:10 crc kubenswrapper[4821]: I1125 11:38:10.847614 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m5z46\" (UniqueName: \"kubernetes.io/projected/2f2ad157-aa04-48bb-bdb8-2364f7127c34-kube-api-access-m5z46\") pod \"2f2ad157-aa04-48bb-bdb8-2364f7127c34\" (UID: \"2f2ad157-aa04-48bb-bdb8-2364f7127c34\") " Nov 25 11:38:10 crc kubenswrapper[4821]: I1125 11:38:10.847780 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f2ad157-aa04-48bb-bdb8-2364f7127c34-catalog-content\") pod \"2f2ad157-aa04-48bb-bdb8-2364f7127c34\" (UID: \"2f2ad157-aa04-48bb-bdb8-2364f7127c34\") " Nov 25 11:38:10 crc kubenswrapper[4821]: I1125 11:38:10.848347 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f2ad157-aa04-48bb-bdb8-2364f7127c34-utilities" (OuterVolumeSpecName: "utilities") pod "2f2ad157-aa04-48bb-bdb8-2364f7127c34" (UID: "2f2ad157-aa04-48bb-bdb8-2364f7127c34"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:38:10 crc kubenswrapper[4821]: I1125 11:38:10.858541 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f2ad157-aa04-48bb-bdb8-2364f7127c34-kube-api-access-m5z46" (OuterVolumeSpecName: "kube-api-access-m5z46") pod "2f2ad157-aa04-48bb-bdb8-2364f7127c34" (UID: "2f2ad157-aa04-48bb-bdb8-2364f7127c34"). InnerVolumeSpecName "kube-api-access-m5z46". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:38:10 crc kubenswrapper[4821]: I1125 11:38:10.872517 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f2ad157-aa04-48bb-bdb8-2364f7127c34-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2f2ad157-aa04-48bb-bdb8-2364f7127c34" (UID: "2f2ad157-aa04-48bb-bdb8-2364f7127c34"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:38:10 crc kubenswrapper[4821]: I1125 11:38:10.950099 4821 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f2ad157-aa04-48bb-bdb8-2364f7127c34-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 11:38:10 crc kubenswrapper[4821]: I1125 11:38:10.950273 4821 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f2ad157-aa04-48bb-bdb8-2364f7127c34-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 11:38:10 crc kubenswrapper[4821]: I1125 11:38:10.950396 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m5z46\" (UniqueName: \"kubernetes.io/projected/2f2ad157-aa04-48bb-bdb8-2364f7127c34-kube-api-access-m5z46\") on node \"crc\" DevicePath \"\"" Nov 25 11:38:11 crc kubenswrapper[4821]: I1125 11:38:11.114484 4821 scope.go:117] "RemoveContainer" containerID="f357cf6563c37c0c7cd86102a3e1eea3ba9146a525631fc115bab1d95f4015f3" Nov 25 11:38:11 crc kubenswrapper[4821]: E1125 11:38:11.114749 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:38:11 crc kubenswrapper[4821]: I1125 11:38:11.303607 4821 generic.go:334] "Generic (PLEG): container finished" podID="2f2ad157-aa04-48bb-bdb8-2364f7127c34" containerID="cb58faeccc9e182e2c5677e3fa925cac3a35f52db5a961902531657890af3c6b" exitCode=0 Nov 25 11:38:11 crc kubenswrapper[4821]: I1125 11:38:11.303673 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-j4bh5" Nov 25 11:38:11 crc kubenswrapper[4821]: I1125 11:38:11.303696 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j4bh5" event={"ID":"2f2ad157-aa04-48bb-bdb8-2364f7127c34","Type":"ContainerDied","Data":"cb58faeccc9e182e2c5677e3fa925cac3a35f52db5a961902531657890af3c6b"} Nov 25 11:38:11 crc kubenswrapper[4821]: I1125 11:38:11.304087 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j4bh5" event={"ID":"2f2ad157-aa04-48bb-bdb8-2364f7127c34","Type":"ContainerDied","Data":"cd8c86fc71b9a1f3434a9fb769305407dc1ad0bd2c97c1b9ca6c982dd75f2e4c"} Nov 25 11:38:11 crc kubenswrapper[4821]: I1125 11:38:11.304112 4821 scope.go:117] "RemoveContainer" containerID="cb58faeccc9e182e2c5677e3fa925cac3a35f52db5a961902531657890af3c6b" Nov 25 11:38:11 crc kubenswrapper[4821]: I1125 11:38:11.341042 4821 scope.go:117] "RemoveContainer" containerID="6531d282b49e90d86bd9ad77b1847d249afa20bd6711937d08524ae915ff58e4" Nov 25 11:38:11 crc kubenswrapper[4821]: I1125 11:38:11.345809 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-j4bh5"] Nov 25 11:38:11 crc kubenswrapper[4821]: I1125 11:38:11.355922 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-j4bh5"] Nov 25 11:38:11 crc kubenswrapper[4821]: I1125 11:38:11.365717 4821 scope.go:117] "RemoveContainer" containerID="b22ebf569a598bc380dfeabcd816c8349000f68d190ebb4e0b4c5fde78bf353f" Nov 25 11:38:11 crc kubenswrapper[4821]: I1125 11:38:11.406397 4821 scope.go:117] "RemoveContainer" containerID="cb58faeccc9e182e2c5677e3fa925cac3a35f52db5a961902531657890af3c6b" Nov 25 11:38:11 crc kubenswrapper[4821]: E1125 11:38:11.406853 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb58faeccc9e182e2c5677e3fa925cac3a35f52db5a961902531657890af3c6b\": container with ID starting with cb58faeccc9e182e2c5677e3fa925cac3a35f52db5a961902531657890af3c6b not found: ID does not exist" containerID="cb58faeccc9e182e2c5677e3fa925cac3a35f52db5a961902531657890af3c6b" Nov 25 11:38:11 crc kubenswrapper[4821]: I1125 11:38:11.406896 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb58faeccc9e182e2c5677e3fa925cac3a35f52db5a961902531657890af3c6b"} err="failed to get container status \"cb58faeccc9e182e2c5677e3fa925cac3a35f52db5a961902531657890af3c6b\": rpc error: code = NotFound desc = could not find container \"cb58faeccc9e182e2c5677e3fa925cac3a35f52db5a961902531657890af3c6b\": container with ID starting with cb58faeccc9e182e2c5677e3fa925cac3a35f52db5a961902531657890af3c6b not found: ID does not exist" Nov 25 11:38:11 crc kubenswrapper[4821]: I1125 11:38:11.406924 4821 scope.go:117] "RemoveContainer" containerID="6531d282b49e90d86bd9ad77b1847d249afa20bd6711937d08524ae915ff58e4" Nov 25 11:38:11 crc kubenswrapper[4821]: E1125 11:38:11.407248 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6531d282b49e90d86bd9ad77b1847d249afa20bd6711937d08524ae915ff58e4\": container with ID starting with 6531d282b49e90d86bd9ad77b1847d249afa20bd6711937d08524ae915ff58e4 not found: ID does not exist" containerID="6531d282b49e90d86bd9ad77b1847d249afa20bd6711937d08524ae915ff58e4" Nov 25 11:38:11 crc kubenswrapper[4821]: I1125 11:38:11.407297 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6531d282b49e90d86bd9ad77b1847d249afa20bd6711937d08524ae915ff58e4"} err="failed to get container status \"6531d282b49e90d86bd9ad77b1847d249afa20bd6711937d08524ae915ff58e4\": rpc error: code = NotFound desc = could not find container \"6531d282b49e90d86bd9ad77b1847d249afa20bd6711937d08524ae915ff58e4\": container with ID starting with 6531d282b49e90d86bd9ad77b1847d249afa20bd6711937d08524ae915ff58e4 not found: ID does not exist" Nov 25 11:38:11 crc kubenswrapper[4821]: I1125 11:38:11.407336 4821 scope.go:117] "RemoveContainer" containerID="b22ebf569a598bc380dfeabcd816c8349000f68d190ebb4e0b4c5fde78bf353f" Nov 25 11:38:11 crc kubenswrapper[4821]: E1125 11:38:11.407826 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b22ebf569a598bc380dfeabcd816c8349000f68d190ebb4e0b4c5fde78bf353f\": container with ID starting with b22ebf569a598bc380dfeabcd816c8349000f68d190ebb4e0b4c5fde78bf353f not found: ID does not exist" containerID="b22ebf569a598bc380dfeabcd816c8349000f68d190ebb4e0b4c5fde78bf353f" Nov 25 11:38:11 crc kubenswrapper[4821]: I1125 11:38:11.407867 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b22ebf569a598bc380dfeabcd816c8349000f68d190ebb4e0b4c5fde78bf353f"} err="failed to get container status \"b22ebf569a598bc380dfeabcd816c8349000f68d190ebb4e0b4c5fde78bf353f\": rpc error: code = NotFound desc = could not find container \"b22ebf569a598bc380dfeabcd816c8349000f68d190ebb4e0b4c5fde78bf353f\": container with ID starting with b22ebf569a598bc380dfeabcd816c8349000f68d190ebb4e0b4c5fde78bf353f not found: ID does not exist" Nov 25 11:38:11 crc kubenswrapper[4821]: I1125 11:38:11.537898 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-4b5tq"] Nov 25 11:38:11 crc kubenswrapper[4821]: E1125 11:38:11.538368 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f2ad157-aa04-48bb-bdb8-2364f7127c34" containerName="extract-content" Nov 25 11:38:11 crc kubenswrapper[4821]: I1125 11:38:11.538389 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f2ad157-aa04-48bb-bdb8-2364f7127c34" containerName="extract-content" Nov 25 11:38:11 crc kubenswrapper[4821]: E1125 11:38:11.538410 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f2ad157-aa04-48bb-bdb8-2364f7127c34" containerName="extract-utilities" Nov 25 11:38:11 crc kubenswrapper[4821]: I1125 11:38:11.538417 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f2ad157-aa04-48bb-bdb8-2364f7127c34" containerName="extract-utilities" Nov 25 11:38:11 crc kubenswrapper[4821]: E1125 11:38:11.538439 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f2ad157-aa04-48bb-bdb8-2364f7127c34" containerName="registry-server" Nov 25 11:38:11 crc kubenswrapper[4821]: I1125 11:38:11.538445 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f2ad157-aa04-48bb-bdb8-2364f7127c34" containerName="registry-server" Nov 25 11:38:11 crc kubenswrapper[4821]: I1125 11:38:11.538674 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f2ad157-aa04-48bb-bdb8-2364f7127c34" containerName="registry-server" Nov 25 11:38:11 crc kubenswrapper[4821]: I1125 11:38:11.540204 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4b5tq" Nov 25 11:38:11 crc kubenswrapper[4821]: I1125 11:38:11.549042 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4b5tq"] Nov 25 11:38:11 crc kubenswrapper[4821]: I1125 11:38:11.578300 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45c9ecf9-7f75-4fbc-983c-5f869ed308ac-catalog-content\") pod \"redhat-operators-4b5tq\" (UID: \"45c9ecf9-7f75-4fbc-983c-5f869ed308ac\") " pod="openshift-marketplace/redhat-operators-4b5tq" Nov 25 11:38:11 crc kubenswrapper[4821]: I1125 11:38:11.578488 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45c9ecf9-7f75-4fbc-983c-5f869ed308ac-utilities\") pod \"redhat-operators-4b5tq\" (UID: \"45c9ecf9-7f75-4fbc-983c-5f869ed308ac\") " pod="openshift-marketplace/redhat-operators-4b5tq" Nov 25 11:38:11 crc kubenswrapper[4821]: I1125 11:38:11.578604 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8bkgx\" (UniqueName: \"kubernetes.io/projected/45c9ecf9-7f75-4fbc-983c-5f869ed308ac-kube-api-access-8bkgx\") pod \"redhat-operators-4b5tq\" (UID: \"45c9ecf9-7f75-4fbc-983c-5f869ed308ac\") " pod="openshift-marketplace/redhat-operators-4b5tq" Nov 25 11:38:11 crc kubenswrapper[4821]: I1125 11:38:11.681134 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45c9ecf9-7f75-4fbc-983c-5f869ed308ac-catalog-content\") pod \"redhat-operators-4b5tq\" (UID: \"45c9ecf9-7f75-4fbc-983c-5f869ed308ac\") " pod="openshift-marketplace/redhat-operators-4b5tq" Nov 25 11:38:11 crc kubenswrapper[4821]: I1125 11:38:11.681588 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45c9ecf9-7f75-4fbc-983c-5f869ed308ac-utilities\") pod \"redhat-operators-4b5tq\" (UID: \"45c9ecf9-7f75-4fbc-983c-5f869ed308ac\") " pod="openshift-marketplace/redhat-operators-4b5tq" Nov 25 11:38:11 crc kubenswrapper[4821]: I1125 11:38:11.681628 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45c9ecf9-7f75-4fbc-983c-5f869ed308ac-catalog-content\") pod \"redhat-operators-4b5tq\" (UID: \"45c9ecf9-7f75-4fbc-983c-5f869ed308ac\") " pod="openshift-marketplace/redhat-operators-4b5tq" Nov 25 11:38:11 crc kubenswrapper[4821]: I1125 11:38:11.681630 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8bkgx\" (UniqueName: \"kubernetes.io/projected/45c9ecf9-7f75-4fbc-983c-5f869ed308ac-kube-api-access-8bkgx\") pod \"redhat-operators-4b5tq\" (UID: \"45c9ecf9-7f75-4fbc-983c-5f869ed308ac\") " pod="openshift-marketplace/redhat-operators-4b5tq" Nov 25 11:38:11 crc kubenswrapper[4821]: I1125 11:38:11.682184 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45c9ecf9-7f75-4fbc-983c-5f869ed308ac-utilities\") pod \"redhat-operators-4b5tq\" (UID: \"45c9ecf9-7f75-4fbc-983c-5f869ed308ac\") " pod="openshift-marketplace/redhat-operators-4b5tq" Nov 25 11:38:11 crc kubenswrapper[4821]: I1125 11:38:11.710545 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8bkgx\" (UniqueName: \"kubernetes.io/projected/45c9ecf9-7f75-4fbc-983c-5f869ed308ac-kube-api-access-8bkgx\") pod \"redhat-operators-4b5tq\" (UID: \"45c9ecf9-7f75-4fbc-983c-5f869ed308ac\") " pod="openshift-marketplace/redhat-operators-4b5tq" Nov 25 11:38:11 crc kubenswrapper[4821]: I1125 11:38:11.911955 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4b5tq" Nov 25 11:38:12 crc kubenswrapper[4821]: I1125 11:38:12.132873 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f2ad157-aa04-48bb-bdb8-2364f7127c34" path="/var/lib/kubelet/pods/2f2ad157-aa04-48bb-bdb8-2364f7127c34/volumes" Nov 25 11:38:12 crc kubenswrapper[4821]: I1125 11:38:12.410935 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4b5tq"] Nov 25 11:38:13 crc kubenswrapper[4821]: I1125 11:38:13.322465 4821 generic.go:334] "Generic (PLEG): container finished" podID="45c9ecf9-7f75-4fbc-983c-5f869ed308ac" containerID="9113c961eed820c7c88e06abe505b706f54b7d1caec003b7d56c1e9e8b6afa93" exitCode=0 Nov 25 11:38:13 crc kubenswrapper[4821]: I1125 11:38:13.322668 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4b5tq" event={"ID":"45c9ecf9-7f75-4fbc-983c-5f869ed308ac","Type":"ContainerDied","Data":"9113c961eed820c7c88e06abe505b706f54b7d1caec003b7d56c1e9e8b6afa93"} Nov 25 11:38:13 crc kubenswrapper[4821]: I1125 11:38:13.322960 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4b5tq" event={"ID":"45c9ecf9-7f75-4fbc-983c-5f869ed308ac","Type":"ContainerStarted","Data":"aff7b12c6cd5eee7348f56a2781605f76c016c4b8a6beddc5b6fab5c74441363"} Nov 25 11:38:25 crc kubenswrapper[4821]: I1125 11:38:25.114103 4821 scope.go:117] "RemoveContainer" containerID="f357cf6563c37c0c7cd86102a3e1eea3ba9146a525631fc115bab1d95f4015f3" Nov 25 11:38:25 crc kubenswrapper[4821]: E1125 11:38:25.114709 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:38:25 crc kubenswrapper[4821]: I1125 11:38:25.444563 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4b5tq" event={"ID":"45c9ecf9-7f75-4fbc-983c-5f869ed308ac","Type":"ContainerStarted","Data":"9fd32911c1f5cdf5c21a8a45179fcb06c2c40da0e8b0e9c92810c4783972527d"} Nov 25 11:38:28 crc kubenswrapper[4821]: I1125 11:38:28.470456 4821 generic.go:334] "Generic (PLEG): container finished" podID="45c9ecf9-7f75-4fbc-983c-5f869ed308ac" containerID="9fd32911c1f5cdf5c21a8a45179fcb06c2c40da0e8b0e9c92810c4783972527d" exitCode=0 Nov 25 11:38:28 crc kubenswrapper[4821]: I1125 11:38:28.470542 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4b5tq" event={"ID":"45c9ecf9-7f75-4fbc-983c-5f869ed308ac","Type":"ContainerDied","Data":"9fd32911c1f5cdf5c21a8a45179fcb06c2c40da0e8b0e9c92810c4783972527d"} Nov 25 11:38:32 crc kubenswrapper[4821]: I1125 11:38:32.502638 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4b5tq" event={"ID":"45c9ecf9-7f75-4fbc-983c-5f869ed308ac","Type":"ContainerStarted","Data":"aed5bb2dac7d475b0526724a67ae978d0a08cf3839fc6cd72e33b520612ab614"} Nov 25 11:38:32 crc kubenswrapper[4821]: I1125 11:38:32.527879 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-4b5tq" podStartSLOduration=3.317743964 podStartE2EDuration="21.527860708s" podCreationTimestamp="2025-11-25 11:38:11 +0000 UTC" firstStartedPulling="2025-11-25 11:38:13.324242198 +0000 UTC m=+3963.860562045" lastFinishedPulling="2025-11-25 11:38:31.534358952 +0000 UTC m=+3982.070678789" observedRunningTime="2025-11-25 11:38:32.521193266 +0000 UTC m=+3983.057513113" watchObservedRunningTime="2025-11-25 11:38:32.527860708 +0000 UTC m=+3983.064180555" Nov 25 11:38:36 crc kubenswrapper[4821]: I1125 11:38:36.114999 4821 scope.go:117] "RemoveContainer" containerID="f357cf6563c37c0c7cd86102a3e1eea3ba9146a525631fc115bab1d95f4015f3" Nov 25 11:38:36 crc kubenswrapper[4821]: E1125 11:38:36.115752 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:38:41 crc kubenswrapper[4821]: I1125 11:38:41.912531 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-4b5tq" Nov 25 11:38:41 crc kubenswrapper[4821]: I1125 11:38:41.913023 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-4b5tq" Nov 25 11:38:41 crc kubenswrapper[4821]: I1125 11:38:41.968836 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-4b5tq" Nov 25 11:38:42 crc kubenswrapper[4821]: I1125 11:38:42.601466 4821 generic.go:334] "Generic (PLEG): container finished" podID="4ea200a7-e3dc-42cf-b4a5-51e45bcec116" containerID="49bc7c06039be18e3024a8625707ddd6ccb9e5d76c93fa103fc1a4f3f44e7bcf" exitCode=0 Nov 25 11:38:42 crc kubenswrapper[4821]: I1125 11:38:42.601551 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gdrrz/crc-debug-mwt2q" event={"ID":"4ea200a7-e3dc-42cf-b4a5-51e45bcec116","Type":"ContainerDied","Data":"49bc7c06039be18e3024a8625707ddd6ccb9e5d76c93fa103fc1a4f3f44e7bcf"} Nov 25 11:38:42 crc kubenswrapper[4821]: I1125 11:38:42.969426 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-4b5tq" Nov 25 11:38:43 crc kubenswrapper[4821]: I1125 11:38:43.025358 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4b5tq"] Nov 25 11:38:43 crc kubenswrapper[4821]: I1125 11:38:43.073379 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vrjfk"] Nov 25 11:38:43 crc kubenswrapper[4821]: I1125 11:38:43.073685 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-vrjfk" podUID="7b01ce52-a13b-431c-8597-fbd5dc25183f" containerName="registry-server" containerID="cri-o://ec86cffebbf2fa03d00d13faa4f868071533d25e0bf365675bf6e468d9d829f5" gracePeriod=2 Nov 25 11:38:43 crc kubenswrapper[4821]: I1125 11:38:43.614643 4821 generic.go:334] "Generic (PLEG): container finished" podID="7b01ce52-a13b-431c-8597-fbd5dc25183f" containerID="ec86cffebbf2fa03d00d13faa4f868071533d25e0bf365675bf6e468d9d829f5" exitCode=0 Nov 25 11:38:43 crc kubenswrapper[4821]: I1125 11:38:43.614731 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vrjfk" event={"ID":"7b01ce52-a13b-431c-8597-fbd5dc25183f","Type":"ContainerDied","Data":"ec86cffebbf2fa03d00d13faa4f868071533d25e0bf365675bf6e468d9d829f5"} Nov 25 11:38:43 crc kubenswrapper[4821]: I1125 11:38:43.615394 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vrjfk" event={"ID":"7b01ce52-a13b-431c-8597-fbd5dc25183f","Type":"ContainerDied","Data":"afda5ad4dc2ed3b17044e732cc977c668a32c8f95a394a1a3ea7d0aec0bef38a"} Nov 25 11:38:43 crc kubenswrapper[4821]: I1125 11:38:43.615418 4821 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="afda5ad4dc2ed3b17044e732cc977c668a32c8f95a394a1a3ea7d0aec0bef38a" Nov 25 11:38:43 crc kubenswrapper[4821]: I1125 11:38:43.700928 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vrjfk" Nov 25 11:38:43 crc kubenswrapper[4821]: I1125 11:38:43.707695 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gdrrz/crc-debug-mwt2q" Nov 25 11:38:43 crc kubenswrapper[4821]: I1125 11:38:43.760313 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-gdrrz/crc-debug-mwt2q"] Nov 25 11:38:43 crc kubenswrapper[4821]: I1125 11:38:43.769689 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-gdrrz/crc-debug-mwt2q"] Nov 25 11:38:43 crc kubenswrapper[4821]: I1125 11:38:43.813806 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s8kj9\" (UniqueName: \"kubernetes.io/projected/7b01ce52-a13b-431c-8597-fbd5dc25183f-kube-api-access-s8kj9\") pod \"7b01ce52-a13b-431c-8597-fbd5dc25183f\" (UID: \"7b01ce52-a13b-431c-8597-fbd5dc25183f\") " Nov 25 11:38:43 crc kubenswrapper[4821]: I1125 11:38:43.813924 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b01ce52-a13b-431c-8597-fbd5dc25183f-catalog-content\") pod \"7b01ce52-a13b-431c-8597-fbd5dc25183f\" (UID: \"7b01ce52-a13b-431c-8597-fbd5dc25183f\") " Nov 25 11:38:43 crc kubenswrapper[4821]: I1125 11:38:43.814081 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8wzh8\" (UniqueName: \"kubernetes.io/projected/4ea200a7-e3dc-42cf-b4a5-51e45bcec116-kube-api-access-8wzh8\") pod \"4ea200a7-e3dc-42cf-b4a5-51e45bcec116\" (UID: \"4ea200a7-e3dc-42cf-b4a5-51e45bcec116\") " Nov 25 11:38:43 crc kubenswrapper[4821]: I1125 11:38:43.814115 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4ea200a7-e3dc-42cf-b4a5-51e45bcec116-host\") pod \"4ea200a7-e3dc-42cf-b4a5-51e45bcec116\" (UID: \"4ea200a7-e3dc-42cf-b4a5-51e45bcec116\") " Nov 25 11:38:43 crc kubenswrapper[4821]: I1125 11:38:43.814189 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b01ce52-a13b-431c-8597-fbd5dc25183f-utilities\") pod \"7b01ce52-a13b-431c-8597-fbd5dc25183f\" (UID: \"7b01ce52-a13b-431c-8597-fbd5dc25183f\") " Nov 25 11:38:43 crc kubenswrapper[4821]: I1125 11:38:43.814721 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4ea200a7-e3dc-42cf-b4a5-51e45bcec116-host" (OuterVolumeSpecName: "host") pod "4ea200a7-e3dc-42cf-b4a5-51e45bcec116" (UID: "4ea200a7-e3dc-42cf-b4a5-51e45bcec116"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 11:38:43 crc kubenswrapper[4821]: I1125 11:38:43.814953 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b01ce52-a13b-431c-8597-fbd5dc25183f-utilities" (OuterVolumeSpecName: "utilities") pod "7b01ce52-a13b-431c-8597-fbd5dc25183f" (UID: "7b01ce52-a13b-431c-8597-fbd5dc25183f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:38:43 crc kubenswrapper[4821]: I1125 11:38:43.820653 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ea200a7-e3dc-42cf-b4a5-51e45bcec116-kube-api-access-8wzh8" (OuterVolumeSpecName: "kube-api-access-8wzh8") pod "4ea200a7-e3dc-42cf-b4a5-51e45bcec116" (UID: "4ea200a7-e3dc-42cf-b4a5-51e45bcec116"). InnerVolumeSpecName "kube-api-access-8wzh8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:38:43 crc kubenswrapper[4821]: I1125 11:38:43.822426 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b01ce52-a13b-431c-8597-fbd5dc25183f-kube-api-access-s8kj9" (OuterVolumeSpecName: "kube-api-access-s8kj9") pod "7b01ce52-a13b-431c-8597-fbd5dc25183f" (UID: "7b01ce52-a13b-431c-8597-fbd5dc25183f"). InnerVolumeSpecName "kube-api-access-s8kj9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:38:43 crc kubenswrapper[4821]: I1125 11:38:43.898091 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b01ce52-a13b-431c-8597-fbd5dc25183f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7b01ce52-a13b-431c-8597-fbd5dc25183f" (UID: "7b01ce52-a13b-431c-8597-fbd5dc25183f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:38:43 crc kubenswrapper[4821]: I1125 11:38:43.916859 4821 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b01ce52-a13b-431c-8597-fbd5dc25183f-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 11:38:43 crc kubenswrapper[4821]: I1125 11:38:43.916905 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s8kj9\" (UniqueName: \"kubernetes.io/projected/7b01ce52-a13b-431c-8597-fbd5dc25183f-kube-api-access-s8kj9\") on node \"crc\" DevicePath \"\"" Nov 25 11:38:43 crc kubenswrapper[4821]: I1125 11:38:43.916920 4821 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b01ce52-a13b-431c-8597-fbd5dc25183f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 11:38:43 crc kubenswrapper[4821]: I1125 11:38:43.916931 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8wzh8\" (UniqueName: \"kubernetes.io/projected/4ea200a7-e3dc-42cf-b4a5-51e45bcec116-kube-api-access-8wzh8\") on node \"crc\" DevicePath \"\"" Nov 25 11:38:43 crc kubenswrapper[4821]: I1125 11:38:43.916946 4821 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4ea200a7-e3dc-42cf-b4a5-51e45bcec116-host\") on node \"crc\" DevicePath \"\"" Nov 25 11:38:44 crc kubenswrapper[4821]: I1125 11:38:44.124469 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ea200a7-e3dc-42cf-b4a5-51e45bcec116" path="/var/lib/kubelet/pods/4ea200a7-e3dc-42cf-b4a5-51e45bcec116/volumes" Nov 25 11:38:44 crc kubenswrapper[4821]: I1125 11:38:44.623625 4821 scope.go:117] "RemoveContainer" containerID="49bc7c06039be18e3024a8625707ddd6ccb9e5d76c93fa103fc1a4f3f44e7bcf" Nov 25 11:38:44 crc kubenswrapper[4821]: I1125 11:38:44.623667 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gdrrz/crc-debug-mwt2q" Nov 25 11:38:44 crc kubenswrapper[4821]: I1125 11:38:44.623670 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vrjfk" Nov 25 11:38:44 crc kubenswrapper[4821]: I1125 11:38:44.655002 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vrjfk"] Nov 25 11:38:44 crc kubenswrapper[4821]: I1125 11:38:44.663571 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-vrjfk"] Nov 25 11:38:45 crc kubenswrapper[4821]: I1125 11:38:45.034783 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-gdrrz/crc-debug-8v7gq"] Nov 25 11:38:45 crc kubenswrapper[4821]: E1125 11:38:45.035214 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ea200a7-e3dc-42cf-b4a5-51e45bcec116" containerName="container-00" Nov 25 11:38:45 crc kubenswrapper[4821]: I1125 11:38:45.035229 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ea200a7-e3dc-42cf-b4a5-51e45bcec116" containerName="container-00" Nov 25 11:38:45 crc kubenswrapper[4821]: E1125 11:38:45.035242 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b01ce52-a13b-431c-8597-fbd5dc25183f" containerName="extract-utilities" Nov 25 11:38:45 crc kubenswrapper[4821]: I1125 11:38:45.035251 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b01ce52-a13b-431c-8597-fbd5dc25183f" containerName="extract-utilities" Nov 25 11:38:45 crc kubenswrapper[4821]: E1125 11:38:45.035264 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b01ce52-a13b-431c-8597-fbd5dc25183f" containerName="registry-server" Nov 25 11:38:45 crc kubenswrapper[4821]: I1125 11:38:45.035271 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b01ce52-a13b-431c-8597-fbd5dc25183f" containerName="registry-server" Nov 25 11:38:45 crc kubenswrapper[4821]: E1125 11:38:45.035318 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b01ce52-a13b-431c-8597-fbd5dc25183f" containerName="extract-content" Nov 25 11:38:45 crc kubenswrapper[4821]: I1125 11:38:45.035332 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b01ce52-a13b-431c-8597-fbd5dc25183f" containerName="extract-content" Nov 25 11:38:45 crc kubenswrapper[4821]: I1125 11:38:45.035511 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ea200a7-e3dc-42cf-b4a5-51e45bcec116" containerName="container-00" Nov 25 11:38:45 crc kubenswrapper[4821]: I1125 11:38:45.035534 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b01ce52-a13b-431c-8597-fbd5dc25183f" containerName="registry-server" Nov 25 11:38:45 crc kubenswrapper[4821]: I1125 11:38:45.036183 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gdrrz/crc-debug-8v7gq" Nov 25 11:38:45 crc kubenswrapper[4821]: I1125 11:38:45.038262 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-gdrrz"/"default-dockercfg-8kkz2" Nov 25 11:38:45 crc kubenswrapper[4821]: I1125 11:38:45.138675 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94z4w\" (UniqueName: \"kubernetes.io/projected/28842e6f-7224-499d-a17b-7c861f395431-kube-api-access-94z4w\") pod \"crc-debug-8v7gq\" (UID: \"28842e6f-7224-499d-a17b-7c861f395431\") " pod="openshift-must-gather-gdrrz/crc-debug-8v7gq" Nov 25 11:38:45 crc kubenswrapper[4821]: I1125 11:38:45.138742 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/28842e6f-7224-499d-a17b-7c861f395431-host\") pod \"crc-debug-8v7gq\" (UID: \"28842e6f-7224-499d-a17b-7c861f395431\") " pod="openshift-must-gather-gdrrz/crc-debug-8v7gq" Nov 25 11:38:45 crc kubenswrapper[4821]: I1125 11:38:45.241202 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-94z4w\" (UniqueName: \"kubernetes.io/projected/28842e6f-7224-499d-a17b-7c861f395431-kube-api-access-94z4w\") pod \"crc-debug-8v7gq\" (UID: \"28842e6f-7224-499d-a17b-7c861f395431\") " pod="openshift-must-gather-gdrrz/crc-debug-8v7gq" Nov 25 11:38:45 crc kubenswrapper[4821]: I1125 11:38:45.241274 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/28842e6f-7224-499d-a17b-7c861f395431-host\") pod \"crc-debug-8v7gq\" (UID: \"28842e6f-7224-499d-a17b-7c861f395431\") " pod="openshift-must-gather-gdrrz/crc-debug-8v7gq" Nov 25 11:38:45 crc kubenswrapper[4821]: I1125 11:38:45.241661 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/28842e6f-7224-499d-a17b-7c861f395431-host\") pod \"crc-debug-8v7gq\" (UID: \"28842e6f-7224-499d-a17b-7c861f395431\") " pod="openshift-must-gather-gdrrz/crc-debug-8v7gq" Nov 25 11:38:45 crc kubenswrapper[4821]: I1125 11:38:45.276175 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-94z4w\" (UniqueName: \"kubernetes.io/projected/28842e6f-7224-499d-a17b-7c861f395431-kube-api-access-94z4w\") pod \"crc-debug-8v7gq\" (UID: \"28842e6f-7224-499d-a17b-7c861f395431\") " pod="openshift-must-gather-gdrrz/crc-debug-8v7gq" Nov 25 11:38:45 crc kubenswrapper[4821]: I1125 11:38:45.354011 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gdrrz/crc-debug-8v7gq" Nov 25 11:38:45 crc kubenswrapper[4821]: I1125 11:38:45.634502 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gdrrz/crc-debug-8v7gq" event={"ID":"28842e6f-7224-499d-a17b-7c861f395431","Type":"ContainerStarted","Data":"36bb26e26cd399ef6bf6902393af0477032407fbe4ef817b3f3b26e7561f6a88"} Nov 25 11:38:45 crc kubenswrapper[4821]: I1125 11:38:45.634759 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gdrrz/crc-debug-8v7gq" event={"ID":"28842e6f-7224-499d-a17b-7c861f395431","Type":"ContainerStarted","Data":"ec6fcc18a3cd7f2d133187fb8fc9ce52125b70b3e10dfd42020f8b83a8e9da69"} Nov 25 11:38:45 crc kubenswrapper[4821]: I1125 11:38:45.653222 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-gdrrz/crc-debug-8v7gq" podStartSLOduration=0.653155469 podStartE2EDuration="653.155469ms" podCreationTimestamp="2025-11-25 11:38:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:38:45.652541684 +0000 UTC m=+3996.188861551" watchObservedRunningTime="2025-11-25 11:38:45.653155469 +0000 UTC m=+3996.189521147" Nov 25 11:38:46 crc kubenswrapper[4821]: I1125 11:38:46.134312 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b01ce52-a13b-431c-8597-fbd5dc25183f" path="/var/lib/kubelet/pods/7b01ce52-a13b-431c-8597-fbd5dc25183f/volumes" Nov 25 11:38:46 crc kubenswrapper[4821]: I1125 11:38:46.644509 4821 generic.go:334] "Generic (PLEG): container finished" podID="28842e6f-7224-499d-a17b-7c861f395431" containerID="36bb26e26cd399ef6bf6902393af0477032407fbe4ef817b3f3b26e7561f6a88" exitCode=0 Nov 25 11:38:46 crc kubenswrapper[4821]: I1125 11:38:46.644557 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gdrrz/crc-debug-8v7gq" event={"ID":"28842e6f-7224-499d-a17b-7c861f395431","Type":"ContainerDied","Data":"36bb26e26cd399ef6bf6902393af0477032407fbe4ef817b3f3b26e7561f6a88"} Nov 25 11:38:47 crc kubenswrapper[4821]: I1125 11:38:47.748899 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gdrrz/crc-debug-8v7gq" Nov 25 11:38:47 crc kubenswrapper[4821]: I1125 11:38:47.789526 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-gdrrz/crc-debug-8v7gq"] Nov 25 11:38:47 crc kubenswrapper[4821]: I1125 11:38:47.800517 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-gdrrz/crc-debug-8v7gq"] Nov 25 11:38:47 crc kubenswrapper[4821]: I1125 11:38:47.891000 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/28842e6f-7224-499d-a17b-7c861f395431-host\") pod \"28842e6f-7224-499d-a17b-7c861f395431\" (UID: \"28842e6f-7224-499d-a17b-7c861f395431\") " Nov 25 11:38:47 crc kubenswrapper[4821]: I1125 11:38:47.891127 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-94z4w\" (UniqueName: \"kubernetes.io/projected/28842e6f-7224-499d-a17b-7c861f395431-kube-api-access-94z4w\") pod \"28842e6f-7224-499d-a17b-7c861f395431\" (UID: \"28842e6f-7224-499d-a17b-7c861f395431\") " Nov 25 11:38:47 crc kubenswrapper[4821]: I1125 11:38:47.891147 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/28842e6f-7224-499d-a17b-7c861f395431-host" (OuterVolumeSpecName: "host") pod "28842e6f-7224-499d-a17b-7c861f395431" (UID: "28842e6f-7224-499d-a17b-7c861f395431"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 11:38:47 crc kubenswrapper[4821]: I1125 11:38:47.891786 4821 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/28842e6f-7224-499d-a17b-7c861f395431-host\") on node \"crc\" DevicePath \"\"" Nov 25 11:38:47 crc kubenswrapper[4821]: I1125 11:38:47.897061 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28842e6f-7224-499d-a17b-7c861f395431-kube-api-access-94z4w" (OuterVolumeSpecName: "kube-api-access-94z4w") pod "28842e6f-7224-499d-a17b-7c861f395431" (UID: "28842e6f-7224-499d-a17b-7c861f395431"). InnerVolumeSpecName "kube-api-access-94z4w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:38:47 crc kubenswrapper[4821]: I1125 11:38:47.993993 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-94z4w\" (UniqueName: \"kubernetes.io/projected/28842e6f-7224-499d-a17b-7c861f395431-kube-api-access-94z4w\") on node \"crc\" DevicePath \"\"" Nov 25 11:38:48 crc kubenswrapper[4821]: I1125 11:38:48.123996 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="28842e6f-7224-499d-a17b-7c861f395431" path="/var/lib/kubelet/pods/28842e6f-7224-499d-a17b-7c861f395431/volumes" Nov 25 11:38:48 crc kubenswrapper[4821]: I1125 11:38:48.663046 4821 scope.go:117] "RemoveContainer" containerID="36bb26e26cd399ef6bf6902393af0477032407fbe4ef817b3f3b26e7561f6a88" Nov 25 11:38:48 crc kubenswrapper[4821]: I1125 11:38:48.663090 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gdrrz/crc-debug-8v7gq" Nov 25 11:38:49 crc kubenswrapper[4821]: I1125 11:38:49.076936 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-gdrrz/crc-debug-tmgr2"] Nov 25 11:38:49 crc kubenswrapper[4821]: E1125 11:38:49.077435 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28842e6f-7224-499d-a17b-7c861f395431" containerName="container-00" Nov 25 11:38:49 crc kubenswrapper[4821]: I1125 11:38:49.077456 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="28842e6f-7224-499d-a17b-7c861f395431" containerName="container-00" Nov 25 11:38:49 crc kubenswrapper[4821]: I1125 11:38:49.077693 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="28842e6f-7224-499d-a17b-7c861f395431" containerName="container-00" Nov 25 11:38:49 crc kubenswrapper[4821]: I1125 11:38:49.078375 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gdrrz/crc-debug-tmgr2" Nov 25 11:38:49 crc kubenswrapper[4821]: I1125 11:38:49.080874 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-gdrrz"/"default-dockercfg-8kkz2" Nov 25 11:38:49 crc kubenswrapper[4821]: I1125 11:38:49.116364 4821 scope.go:117] "RemoveContainer" containerID="f357cf6563c37c0c7cd86102a3e1eea3ba9146a525631fc115bab1d95f4015f3" Nov 25 11:38:49 crc kubenswrapper[4821]: E1125 11:38:49.116615 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:38:49 crc kubenswrapper[4821]: I1125 11:38:49.217884 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/eb3eec6b-7841-485f-b708-d2f9fdf38b44-host\") pod \"crc-debug-tmgr2\" (UID: \"eb3eec6b-7841-485f-b708-d2f9fdf38b44\") " pod="openshift-must-gather-gdrrz/crc-debug-tmgr2" Nov 25 11:38:49 crc kubenswrapper[4821]: I1125 11:38:49.219484 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crld9\" (UniqueName: \"kubernetes.io/projected/eb3eec6b-7841-485f-b708-d2f9fdf38b44-kube-api-access-crld9\") pod \"crc-debug-tmgr2\" (UID: \"eb3eec6b-7841-485f-b708-d2f9fdf38b44\") " pod="openshift-must-gather-gdrrz/crc-debug-tmgr2" Nov 25 11:38:49 crc kubenswrapper[4821]: I1125 11:38:49.322053 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/eb3eec6b-7841-485f-b708-d2f9fdf38b44-host\") pod \"crc-debug-tmgr2\" (UID: \"eb3eec6b-7841-485f-b708-d2f9fdf38b44\") " pod="openshift-must-gather-gdrrz/crc-debug-tmgr2" Nov 25 11:38:49 crc kubenswrapper[4821]: I1125 11:38:49.322281 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crld9\" (UniqueName: \"kubernetes.io/projected/eb3eec6b-7841-485f-b708-d2f9fdf38b44-kube-api-access-crld9\") pod \"crc-debug-tmgr2\" (UID: \"eb3eec6b-7841-485f-b708-d2f9fdf38b44\") " pod="openshift-must-gather-gdrrz/crc-debug-tmgr2" Nov 25 11:38:49 crc kubenswrapper[4821]: I1125 11:38:49.322419 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/eb3eec6b-7841-485f-b708-d2f9fdf38b44-host\") pod \"crc-debug-tmgr2\" (UID: \"eb3eec6b-7841-485f-b708-d2f9fdf38b44\") " pod="openshift-must-gather-gdrrz/crc-debug-tmgr2" Nov 25 11:38:49 crc kubenswrapper[4821]: I1125 11:38:49.347395 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crld9\" (UniqueName: \"kubernetes.io/projected/eb3eec6b-7841-485f-b708-d2f9fdf38b44-kube-api-access-crld9\") pod \"crc-debug-tmgr2\" (UID: \"eb3eec6b-7841-485f-b708-d2f9fdf38b44\") " pod="openshift-must-gather-gdrrz/crc-debug-tmgr2" Nov 25 11:38:49 crc kubenswrapper[4821]: I1125 11:38:49.415304 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gdrrz/crc-debug-tmgr2" Nov 25 11:38:49 crc kubenswrapper[4821]: W1125 11:38:49.445868 4821 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeb3eec6b_7841_485f_b708_d2f9fdf38b44.slice/crio-e4321dc81a3cbc709b745a3bc42c2cf101ba7f3d8625197c40d5956e974044c1 WatchSource:0}: Error finding container e4321dc81a3cbc709b745a3bc42c2cf101ba7f3d8625197c40d5956e974044c1: Status 404 returned error can't find the container with id e4321dc81a3cbc709b745a3bc42c2cf101ba7f3d8625197c40d5956e974044c1 Nov 25 11:38:49 crc kubenswrapper[4821]: I1125 11:38:49.672416 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gdrrz/crc-debug-tmgr2" event={"ID":"eb3eec6b-7841-485f-b708-d2f9fdf38b44","Type":"ContainerStarted","Data":"e4321dc81a3cbc709b745a3bc42c2cf101ba7f3d8625197c40d5956e974044c1"} Nov 25 11:38:50 crc kubenswrapper[4821]: I1125 11:38:50.683500 4821 generic.go:334] "Generic (PLEG): container finished" podID="eb3eec6b-7841-485f-b708-d2f9fdf38b44" containerID="1ba5757b3d8f9e2d9825908000e6f78a56648899ecee86cdfcd0e38150a8ee48" exitCode=0 Nov 25 11:38:50 crc kubenswrapper[4821]: I1125 11:38:50.683557 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gdrrz/crc-debug-tmgr2" event={"ID":"eb3eec6b-7841-485f-b708-d2f9fdf38b44","Type":"ContainerDied","Data":"1ba5757b3d8f9e2d9825908000e6f78a56648899ecee86cdfcd0e38150a8ee48"} Nov 25 11:38:50 crc kubenswrapper[4821]: I1125 11:38:50.724001 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-gdrrz/crc-debug-tmgr2"] Nov 25 11:38:50 crc kubenswrapper[4821]: I1125 11:38:50.733684 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-gdrrz/crc-debug-tmgr2"] Nov 25 11:38:51 crc kubenswrapper[4821]: I1125 11:38:51.794542 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gdrrz/crc-debug-tmgr2" Nov 25 11:38:51 crc kubenswrapper[4821]: I1125 11:38:51.866944 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/eb3eec6b-7841-485f-b708-d2f9fdf38b44-host\") pod \"eb3eec6b-7841-485f-b708-d2f9fdf38b44\" (UID: \"eb3eec6b-7841-485f-b708-d2f9fdf38b44\") " Nov 25 11:38:51 crc kubenswrapper[4821]: I1125 11:38:51.867067 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eb3eec6b-7841-485f-b708-d2f9fdf38b44-host" (OuterVolumeSpecName: "host") pod "eb3eec6b-7841-485f-b708-d2f9fdf38b44" (UID: "eb3eec6b-7841-485f-b708-d2f9fdf38b44"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 11:38:51 crc kubenswrapper[4821]: I1125 11:38:51.867138 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-crld9\" (UniqueName: \"kubernetes.io/projected/eb3eec6b-7841-485f-b708-d2f9fdf38b44-kube-api-access-crld9\") pod \"eb3eec6b-7841-485f-b708-d2f9fdf38b44\" (UID: \"eb3eec6b-7841-485f-b708-d2f9fdf38b44\") " Nov 25 11:38:51 crc kubenswrapper[4821]: I1125 11:38:51.867608 4821 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/eb3eec6b-7841-485f-b708-d2f9fdf38b44-host\") on node \"crc\" DevicePath \"\"" Nov 25 11:38:52 crc kubenswrapper[4821]: I1125 11:38:52.334659 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb3eec6b-7841-485f-b708-d2f9fdf38b44-kube-api-access-crld9" (OuterVolumeSpecName: "kube-api-access-crld9") pod "eb3eec6b-7841-485f-b708-d2f9fdf38b44" (UID: "eb3eec6b-7841-485f-b708-d2f9fdf38b44"). InnerVolumeSpecName "kube-api-access-crld9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:38:52 crc kubenswrapper[4821]: I1125 11:38:52.376721 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-crld9\" (UniqueName: \"kubernetes.io/projected/eb3eec6b-7841-485f-b708-d2f9fdf38b44-kube-api-access-crld9\") on node \"crc\" DevicePath \"\"" Nov 25 11:38:52 crc kubenswrapper[4821]: I1125 11:38:52.705923 4821 scope.go:117] "RemoveContainer" containerID="1ba5757b3d8f9e2d9825908000e6f78a56648899ecee86cdfcd0e38150a8ee48" Nov 25 11:38:52 crc kubenswrapper[4821]: I1125 11:38:52.705973 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gdrrz/crc-debug-tmgr2" Nov 25 11:38:54 crc kubenswrapper[4821]: I1125 11:38:54.126524 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb3eec6b-7841-485f-b708-d2f9fdf38b44" path="/var/lib/kubelet/pods/eb3eec6b-7841-485f-b708-d2f9fdf38b44/volumes" Nov 25 11:39:00 crc kubenswrapper[4821]: I1125 11:39:00.120768 4821 scope.go:117] "RemoveContainer" containerID="f357cf6563c37c0c7cd86102a3e1eea3ba9146a525631fc115bab1d95f4015f3" Nov 25 11:39:00 crc kubenswrapper[4821]: E1125 11:39:00.121683 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:39:13 crc kubenswrapper[4821]: I1125 11:39:13.114805 4821 scope.go:117] "RemoveContainer" containerID="f357cf6563c37c0c7cd86102a3e1eea3ba9146a525631fc115bab1d95f4015f3" Nov 25 11:39:13 crc kubenswrapper[4821]: E1125 11:39:13.115518 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:39:18 crc kubenswrapper[4821]: I1125 11:39:18.387840 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-67b899d6f8-pscf9_b5bf27f5-8585-42d9-9d58-3c6e7f3a34d5/barbican-api/0.log" Nov 25 11:39:18 crc kubenswrapper[4821]: I1125 11:39:18.563261 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-67b899d6f8-pscf9_b5bf27f5-8585-42d9-9d58-3c6e7f3a34d5/barbican-api-log/0.log" Nov 25 11:39:18 crc kubenswrapper[4821]: I1125 11:39:18.579119 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-f75f66468-2k6lr_1e6fee50-9d84-4965-a222-a55f6dcfc4cc/barbican-keystone-listener/0.log" Nov 25 11:39:18 crc kubenswrapper[4821]: I1125 11:39:18.655146 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-f75f66468-2k6lr_1e6fee50-9d84-4965-a222-a55f6dcfc4cc/barbican-keystone-listener-log/0.log" Nov 25 11:39:18 crc kubenswrapper[4821]: I1125 11:39:18.769785 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5c9c8c87dc-mqk8l_62fc74a6-526c-4953-8439-b4af8ce5f39b/barbican-worker/0.log" Nov 25 11:39:18 crc kubenswrapper[4821]: I1125 11:39:18.813720 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5c9c8c87dc-mqk8l_62fc74a6-526c-4953-8439-b4af8ce5f39b/barbican-worker-log/0.log" Nov 25 11:39:19 crc kubenswrapper[4821]: I1125 11:39:19.011939 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_08ee9d9a-c488-4289-9d8f-a87d338992d6/ceilometer-central-agent/0.log" Nov 25 11:39:19 crc kubenswrapper[4821]: I1125 11:39:19.016241 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-pkx5r_0b93320f-a41f-4c0e-874f-b4f945f14c90/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 11:39:19 crc kubenswrapper[4821]: I1125 11:39:19.102748 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_08ee9d9a-c488-4289-9d8f-a87d338992d6/ceilometer-notification-agent/0.log" Nov 25 11:39:19 crc kubenswrapper[4821]: I1125 11:39:19.194183 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_08ee9d9a-c488-4289-9d8f-a87d338992d6/proxy-httpd/0.log" Nov 25 11:39:19 crc kubenswrapper[4821]: I1125 11:39:19.198535 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_08ee9d9a-c488-4289-9d8f-a87d338992d6/sg-core/0.log" Nov 25 11:39:19 crc kubenswrapper[4821]: I1125 11:39:19.325763 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_4fd0ccca-05dd-43e3-80b1-1c312d130056/cinder-api/0.log" Nov 25 11:39:19 crc kubenswrapper[4821]: I1125 11:39:19.422121 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_4fd0ccca-05dd-43e3-80b1-1c312d130056/cinder-api-log/0.log" Nov 25 11:39:19 crc kubenswrapper[4821]: I1125 11:39:19.552435 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_7b0c571f-6107-4ecd-8ffb-cd7b992d05a5/cinder-scheduler/0.log" Nov 25 11:39:19 crc kubenswrapper[4821]: I1125 11:39:19.576102 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_7b0c571f-6107-4ecd-8ffb-cd7b992d05a5/probe/0.log" Nov 25 11:39:19 crc kubenswrapper[4821]: I1125 11:39:19.675820 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-sg8cc_5a15c7fc-4220-46ae-8af3-490c28f8b4ff/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 11:39:19 crc kubenswrapper[4821]: I1125 11:39:19.785900 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-d8lr8_c20e6196-b767-4fbb-9d59-a88080a3aa11/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 11:39:19 crc kubenswrapper[4821]: I1125 11:39:19.893922 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-78c64bc9c5-8r5jj_bda8a37b-363b-446a-8cbb-a708cb3f716f/init/0.log" Nov 25 11:39:20 crc kubenswrapper[4821]: I1125 11:39:20.086297 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-78c64bc9c5-8r5jj_bda8a37b-363b-446a-8cbb-a708cb3f716f/init/0.log" Nov 25 11:39:20 crc kubenswrapper[4821]: I1125 11:39:20.098634 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-78c64bc9c5-8r5jj_bda8a37b-363b-446a-8cbb-a708cb3f716f/dnsmasq-dns/0.log" Nov 25 11:39:20 crc kubenswrapper[4821]: I1125 11:39:20.136539 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-tfcn9_fcf6a791-0724-4ac2-8881-c7f8804f6540/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 11:39:20 crc kubenswrapper[4821]: I1125 11:39:20.272495 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_fada5a81-0157-452d-9ade-dd6ce6303918/glance-log/0.log" Nov 25 11:39:20 crc kubenswrapper[4821]: I1125 11:39:20.322519 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_fada5a81-0157-452d-9ade-dd6ce6303918/glance-httpd/0.log" Nov 25 11:39:20 crc kubenswrapper[4821]: I1125 11:39:20.487151 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_1816fd90-5883-4de2-9124-64e326d1743f/glance-httpd/0.log" Nov 25 11:39:20 crc kubenswrapper[4821]: I1125 11:39:20.543095 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_1816fd90-5883-4de2-9124-64e326d1743f/glance-log/0.log" Nov 25 11:39:20 crc kubenswrapper[4821]: I1125 11:39:20.746940 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-67f765d856-ncc6t_93b4abfa-c14a-4df9-a6d1-9cdeaec918dd/horizon/0.log" Nov 25 11:39:20 crc kubenswrapper[4821]: I1125 11:39:20.844791 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-48x26_98d0e490-2704-4aa6-baa5-32ec701a5b68/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 11:39:20 crc kubenswrapper[4821]: I1125 11:39:20.990631 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-x5qj5_9418fa73-031c-4c0a-bfbd-ffb206714075/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 11:39:21 crc kubenswrapper[4821]: I1125 11:39:21.066762 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-67f765d856-ncc6t_93b4abfa-c14a-4df9-a6d1-9cdeaec918dd/horizon-log/0.log" Nov 25 11:39:21 crc kubenswrapper[4821]: I1125 11:39:21.358778 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29401141-j95bv_0f34ab84-5b6a-423d-b622-9b90c288695e/keystone-cron/0.log" Nov 25 11:39:21 crc kubenswrapper[4821]: I1125 11:39:21.431772 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-5fb467464c-hlrl9_eabb0cb5-e9c3-474f-80ba-df6ecf373c3f/keystone-api/0.log" Nov 25 11:39:21 crc kubenswrapper[4821]: I1125 11:39:21.547137 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_7c286dad-cd09-4085-a4e0-4b274f30d38d/kube-state-metrics/0.log" Nov 25 11:39:21 crc kubenswrapper[4821]: I1125 11:39:21.620869 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-9w8v8_0ee4d7f9-162b-4882-b189-13c53bcdc274/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 11:39:22 crc kubenswrapper[4821]: I1125 11:39:22.001019 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-55575b4d57-6t4kz_a8419a57-35ff-4340-b2d7-fd11724948fe/neutron-api/0.log" Nov 25 11:39:22 crc kubenswrapper[4821]: I1125 11:39:22.019251 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-55575b4d57-6t4kz_a8419a57-35ff-4340-b2d7-fd11724948fe/neutron-httpd/0.log" Nov 25 11:39:22 crc kubenswrapper[4821]: I1125 11:39:22.196580 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-225z5_0f26a3b0-ba05-4484-8281-215ff2f09d25/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 11:39:23 crc kubenswrapper[4821]: I1125 11:39:23.169427 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_2bd244d2-4836-4caf-a6f6-1a00aac5dbb3/nova-api-log/0.log" Nov 25 11:39:23 crc kubenswrapper[4821]: I1125 11:39:23.396275 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_e5dfcd15-af77-4be7-93db-e26f3da1d7d2/nova-cell0-conductor-conductor/0.log" Nov 25 11:39:23 crc kubenswrapper[4821]: I1125 11:39:23.593140 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_fb540939-5683-495a-b8df-d7a000fedaa2/nova-cell1-conductor-conductor/0.log" Nov 25 11:39:23 crc kubenswrapper[4821]: I1125 11:39:23.737412 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_5b8402cb-db96-4ea4-8f95-a2ea84f202f5/nova-cell1-novncproxy-novncproxy/0.log" Nov 25 11:39:23 crc kubenswrapper[4821]: I1125 11:39:23.775544 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_2bd244d2-4836-4caf-a6f6-1a00aac5dbb3/nova-api-api/0.log" Nov 25 11:39:23 crc kubenswrapper[4821]: I1125 11:39:23.880337 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-mt9vk_40682cac-f449-4465-9ad3-3bfa28b3800d/nova-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 11:39:24 crc kubenswrapper[4821]: I1125 11:39:24.427262 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_39dff25e-5059-4eaf-bf98-93aea0eeb258/nova-metadata-log/0.log" Nov 25 11:39:24 crc kubenswrapper[4821]: I1125 11:39:24.543901 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_3bdff849-753c-42b5-9e06-2e37dfe42666/nova-scheduler-scheduler/0.log" Nov 25 11:39:24 crc kubenswrapper[4821]: I1125 11:39:24.637995 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_41c8c049-b466-4c19-944e-50a639491190/mysql-bootstrap/0.log" Nov 25 11:39:24 crc kubenswrapper[4821]: I1125 11:39:24.852926 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_41c8c049-b466-4c19-944e-50a639491190/galera/0.log" Nov 25 11:39:24 crc kubenswrapper[4821]: I1125 11:39:24.857674 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_41c8c049-b466-4c19-944e-50a639491190/mysql-bootstrap/0.log" Nov 25 11:39:25 crc kubenswrapper[4821]: I1125 11:39:25.034574 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_ecb5eaf2-4178-43ec-bc21-0e05de8fc237/mysql-bootstrap/0.log" Nov 25 11:39:25 crc kubenswrapper[4821]: I1125 11:39:25.262976 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_ecb5eaf2-4178-43ec-bc21-0e05de8fc237/galera/0.log" Nov 25 11:39:25 crc kubenswrapper[4821]: I1125 11:39:25.273317 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_ecb5eaf2-4178-43ec-bc21-0e05de8fc237/mysql-bootstrap/0.log" Nov 25 11:39:25 crc kubenswrapper[4821]: I1125 11:39:25.474329 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_d969bbb6-564d-4563-98c6-5e2ec965c2dc/openstackclient/0.log" Nov 25 11:39:25 crc kubenswrapper[4821]: I1125 11:39:25.524502 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ltxc6_47eaf58d-190b-4e75-9aac-11d7e51df259/ovn-controller/0.log" Nov 25 11:39:25 crc kubenswrapper[4821]: I1125 11:39:25.745702 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-kkntf_832f5849-e58a-407b-9f45-55c52cfbcd02/openstack-network-exporter/0.log" Nov 25 11:39:25 crc kubenswrapper[4821]: I1125 11:39:25.795415 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_39dff25e-5059-4eaf-bf98-93aea0eeb258/nova-metadata-metadata/0.log" Nov 25 11:39:25 crc kubenswrapper[4821]: I1125 11:39:25.912239 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-6mbr2_72b307b1-f36e-424f-91ef-f7402b6d7535/ovsdb-server-init/0.log" Nov 25 11:39:26 crc kubenswrapper[4821]: I1125 11:39:26.107428 4821 scope.go:117] "RemoveContainer" containerID="ec86cffebbf2fa03d00d13faa4f868071533d25e0bf365675bf6e468d9d829f5" Nov 25 11:39:26 crc kubenswrapper[4821]: I1125 11:39:26.114769 4821 scope.go:117] "RemoveContainer" containerID="f357cf6563c37c0c7cd86102a3e1eea3ba9146a525631fc115bab1d95f4015f3" Nov 25 11:39:26 crc kubenswrapper[4821]: E1125 11:39:26.115250 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:39:26 crc kubenswrapper[4821]: I1125 11:39:26.134561 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-6mbr2_72b307b1-f36e-424f-91ef-f7402b6d7535/ovsdb-server/0.log" Nov 25 11:39:26 crc kubenswrapper[4821]: I1125 11:39:26.138564 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-6mbr2_72b307b1-f36e-424f-91ef-f7402b6d7535/ovs-vswitchd/0.log" Nov 25 11:39:26 crc kubenswrapper[4821]: I1125 11:39:26.146658 4821 scope.go:117] "RemoveContainer" containerID="7f00fac22e0cf3ffaf5716d44747e9ea8239f22a2bae85ae8bef7e9428ce8a0f" Nov 25 11:39:26 crc kubenswrapper[4821]: I1125 11:39:26.153013 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-6mbr2_72b307b1-f36e-424f-91ef-f7402b6d7535/ovsdb-server-init/0.log" Nov 25 11:39:26 crc kubenswrapper[4821]: I1125 11:39:26.182374 4821 scope.go:117] "RemoveContainer" containerID="02730237ea8ba401377da3e688f1268c42752d3dc386c2cbd3fd7481401320dd" Nov 25 11:39:26 crc kubenswrapper[4821]: I1125 11:39:26.363273 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-9vb7k_df834005-0241-4897-9112-05186899309e/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 11:39:26 crc kubenswrapper[4821]: I1125 11:39:26.435069 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_006cceec-8fb3-44a5-b3e1-50a01bc4b334/openstack-network-exporter/0.log" Nov 25 11:39:26 crc kubenswrapper[4821]: I1125 11:39:26.499439 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_006cceec-8fb3-44a5-b3e1-50a01bc4b334/ovn-northd/0.log" Nov 25 11:39:26 crc kubenswrapper[4821]: I1125 11:39:26.626458 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_30500d7b-4e4b-4839-8a9d-0b05a0ebd1fb/openstack-network-exporter/0.log" Nov 25 11:39:26 crc kubenswrapper[4821]: I1125 11:39:26.715128 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_30500d7b-4e4b-4839-8a9d-0b05a0ebd1fb/ovsdbserver-nb/0.log" Nov 25 11:39:26 crc kubenswrapper[4821]: I1125 11:39:26.825541 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_9aed573a-ad60-479a-ba27-8666d7dc4d9e/openstack-network-exporter/0.log" Nov 25 11:39:26 crc kubenswrapper[4821]: I1125 11:39:26.842965 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_9aed573a-ad60-479a-ba27-8666d7dc4d9e/ovsdbserver-sb/0.log" Nov 25 11:39:27 crc kubenswrapper[4821]: I1125 11:39:27.061836 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-f8667547d-9v5xh_b755bd74-3de7-49ed-857a-a5f23cf84f30/placement-api/0.log" Nov 25 11:39:27 crc kubenswrapper[4821]: I1125 11:39:27.146724 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-f8667547d-9v5xh_b755bd74-3de7-49ed-857a-a5f23cf84f30/placement-log/0.log" Nov 25 11:39:27 crc kubenswrapper[4821]: I1125 11:39:27.185000 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_3556cc6b-7d73-4cbc-9c2e-378c6ba606e4/setup-container/0.log" Nov 25 11:39:27 crc kubenswrapper[4821]: I1125 11:39:27.364509 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_3556cc6b-7d73-4cbc-9c2e-378c6ba606e4/setup-container/0.log" Nov 25 11:39:27 crc kubenswrapper[4821]: I1125 11:39:27.422737 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_3556cc6b-7d73-4cbc-9c2e-378c6ba606e4/rabbitmq/0.log" Nov 25 11:39:27 crc kubenswrapper[4821]: I1125 11:39:27.424514 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_50c27d16-2f67-4f90-83f0-83e16161de50/setup-container/0.log" Nov 25 11:39:27 crc kubenswrapper[4821]: I1125 11:39:27.638218 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_50c27d16-2f67-4f90-83f0-83e16161de50/rabbitmq/0.log" Nov 25 11:39:27 crc kubenswrapper[4821]: I1125 11:39:27.668887 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_50c27d16-2f67-4f90-83f0-83e16161de50/setup-container/0.log" Nov 25 11:39:27 crc kubenswrapper[4821]: I1125 11:39:27.872303 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-98mzz_0d9e3bea-dfb1-479d-a1b2-a17b2334cd86/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 11:39:27 crc kubenswrapper[4821]: I1125 11:39:27.962348 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-9tzcw_74c5a130-3a40-4920-8e84-e90e9ee4b39f/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 11:39:28 crc kubenswrapper[4821]: I1125 11:39:28.077867 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-rdcx6_a15db0db-c211-40aa-bb3e-8f222669b005/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 11:39:28 crc kubenswrapper[4821]: I1125 11:39:28.231026 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-j8sp8_3ad8bc38-c088-4632-8864-00d472c78fbf/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 11:39:28 crc kubenswrapper[4821]: I1125 11:39:28.277762 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-dxdpz_e1247bbd-54fc-45ce-8a90-0b8a5df02c47/ssh-known-hosts-edpm-deployment/0.log" Nov 25 11:39:28 crc kubenswrapper[4821]: I1125 11:39:28.524834 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-ff8c8c76c-ctf8b_db085977-6941-483f-a800-411ff71612af/proxy-server/0.log" Nov 25 11:39:28 crc kubenswrapper[4821]: I1125 11:39:28.618560 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-ff8c8c76c-ctf8b_db085977-6941-483f-a800-411ff71612af/proxy-httpd/0.log" Nov 25 11:39:28 crc kubenswrapper[4821]: I1125 11:39:28.778515 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-gmrpw_1aa8233d-1fe4-44ec-acc6-0d5548c70d59/swift-ring-rebalance/0.log" Nov 25 11:39:28 crc kubenswrapper[4821]: I1125 11:39:28.779975 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0ffb0da2-affb-4a95-87f3-eac442dc754d/account-auditor/0.log" Nov 25 11:39:28 crc kubenswrapper[4821]: I1125 11:39:28.878808 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0ffb0da2-affb-4a95-87f3-eac442dc754d/account-reaper/0.log" Nov 25 11:39:28 crc kubenswrapper[4821]: I1125 11:39:28.970937 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0ffb0da2-affb-4a95-87f3-eac442dc754d/account-server/0.log" Nov 25 11:39:29 crc kubenswrapper[4821]: I1125 11:39:29.006173 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0ffb0da2-affb-4a95-87f3-eac442dc754d/container-auditor/0.log" Nov 25 11:39:29 crc kubenswrapper[4821]: I1125 11:39:29.020587 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0ffb0da2-affb-4a95-87f3-eac442dc754d/account-replicator/0.log" Nov 25 11:39:29 crc kubenswrapper[4821]: I1125 11:39:29.139316 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0ffb0da2-affb-4a95-87f3-eac442dc754d/container-replicator/0.log" Nov 25 11:39:29 crc kubenswrapper[4821]: I1125 11:39:29.176525 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0ffb0da2-affb-4a95-87f3-eac442dc754d/container-updater/0.log" Nov 25 11:39:29 crc kubenswrapper[4821]: I1125 11:39:29.187451 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0ffb0da2-affb-4a95-87f3-eac442dc754d/container-server/0.log" Nov 25 11:39:29 crc kubenswrapper[4821]: I1125 11:39:29.283183 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0ffb0da2-affb-4a95-87f3-eac442dc754d/object-auditor/0.log" Nov 25 11:39:29 crc kubenswrapper[4821]: I1125 11:39:29.333568 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0ffb0da2-affb-4a95-87f3-eac442dc754d/object-expirer/0.log" Nov 25 11:39:29 crc kubenswrapper[4821]: I1125 11:39:29.383837 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0ffb0da2-affb-4a95-87f3-eac442dc754d/object-server/0.log" Nov 25 11:39:29 crc kubenswrapper[4821]: I1125 11:39:29.417794 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0ffb0da2-affb-4a95-87f3-eac442dc754d/object-replicator/0.log" Nov 25 11:39:29 crc kubenswrapper[4821]: I1125 11:39:29.528018 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0ffb0da2-affb-4a95-87f3-eac442dc754d/rsync/0.log" Nov 25 11:39:29 crc kubenswrapper[4821]: I1125 11:39:29.537330 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0ffb0da2-affb-4a95-87f3-eac442dc754d/object-updater/0.log" Nov 25 11:39:29 crc kubenswrapper[4821]: I1125 11:39:29.666677 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_0ffb0da2-affb-4a95-87f3-eac442dc754d/swift-recon-cron/0.log" Nov 25 11:39:29 crc kubenswrapper[4821]: I1125 11:39:29.815697 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-w75cf_814ef062-0cc8-40bf-ac77-c87010466f7d/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 11:39:29 crc kubenswrapper[4821]: I1125 11:39:29.917411 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_d8aea755-bf2f-4443-8196-46936ce3e2fd/tempest-tests-tempest-tests-runner/0.log" Nov 25 11:39:30 crc kubenswrapper[4821]: I1125 11:39:30.015030 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_adcf552f-2365-451b-b4e2-12b58631195e/test-operator-logs-container/0.log" Nov 25 11:39:30 crc kubenswrapper[4821]: I1125 11:39:30.094583 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-x5tw6_d77aa250-b203-4596-913c-7878a4d8b517/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 11:39:38 crc kubenswrapper[4821]: I1125 11:39:38.106744 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_8b9e8e7c-b055-47d3-874a-227d5f02432a/memcached/0.log" Nov 25 11:39:39 crc kubenswrapper[4821]: I1125 11:39:39.114680 4821 scope.go:117] "RemoveContainer" containerID="f357cf6563c37c0c7cd86102a3e1eea3ba9146a525631fc115bab1d95f4015f3" Nov 25 11:39:39 crc kubenswrapper[4821]: E1125 11:39:39.116121 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:39:41 crc kubenswrapper[4821]: I1125 11:39:41.514998 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-fxjm4"] Nov 25 11:39:41 crc kubenswrapper[4821]: E1125 11:39:41.515963 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb3eec6b-7841-485f-b708-d2f9fdf38b44" containerName="container-00" Nov 25 11:39:41 crc kubenswrapper[4821]: I1125 11:39:41.515981 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb3eec6b-7841-485f-b708-d2f9fdf38b44" containerName="container-00" Nov 25 11:39:41 crc kubenswrapper[4821]: I1125 11:39:41.516255 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb3eec6b-7841-485f-b708-d2f9fdf38b44" containerName="container-00" Nov 25 11:39:41 crc kubenswrapper[4821]: I1125 11:39:41.517968 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fxjm4" Nov 25 11:39:41 crc kubenswrapper[4821]: I1125 11:39:41.537614 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fxjm4"] Nov 25 11:39:41 crc kubenswrapper[4821]: I1125 11:39:41.606234 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4h74\" (UniqueName: \"kubernetes.io/projected/2b797317-945f-4546-9548-38cb57d0cd2b-kube-api-access-m4h74\") pod \"community-operators-fxjm4\" (UID: \"2b797317-945f-4546-9548-38cb57d0cd2b\") " pod="openshift-marketplace/community-operators-fxjm4" Nov 25 11:39:41 crc kubenswrapper[4821]: I1125 11:39:41.606361 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b797317-945f-4546-9548-38cb57d0cd2b-catalog-content\") pod \"community-operators-fxjm4\" (UID: \"2b797317-945f-4546-9548-38cb57d0cd2b\") " pod="openshift-marketplace/community-operators-fxjm4" Nov 25 11:39:41 crc kubenswrapper[4821]: I1125 11:39:41.606439 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b797317-945f-4546-9548-38cb57d0cd2b-utilities\") pod \"community-operators-fxjm4\" (UID: \"2b797317-945f-4546-9548-38cb57d0cd2b\") " pod="openshift-marketplace/community-operators-fxjm4" Nov 25 11:39:41 crc kubenswrapper[4821]: I1125 11:39:41.707950 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4h74\" (UniqueName: \"kubernetes.io/projected/2b797317-945f-4546-9548-38cb57d0cd2b-kube-api-access-m4h74\") pod \"community-operators-fxjm4\" (UID: \"2b797317-945f-4546-9548-38cb57d0cd2b\") " pod="openshift-marketplace/community-operators-fxjm4" Nov 25 11:39:41 crc kubenswrapper[4821]: I1125 11:39:41.708074 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b797317-945f-4546-9548-38cb57d0cd2b-catalog-content\") pod \"community-operators-fxjm4\" (UID: \"2b797317-945f-4546-9548-38cb57d0cd2b\") " pod="openshift-marketplace/community-operators-fxjm4" Nov 25 11:39:41 crc kubenswrapper[4821]: I1125 11:39:41.708154 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b797317-945f-4546-9548-38cb57d0cd2b-utilities\") pod \"community-operators-fxjm4\" (UID: \"2b797317-945f-4546-9548-38cb57d0cd2b\") " pod="openshift-marketplace/community-operators-fxjm4" Nov 25 11:39:41 crc kubenswrapper[4821]: I1125 11:39:41.708658 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b797317-945f-4546-9548-38cb57d0cd2b-utilities\") pod \"community-operators-fxjm4\" (UID: \"2b797317-945f-4546-9548-38cb57d0cd2b\") " pod="openshift-marketplace/community-operators-fxjm4" Nov 25 11:39:41 crc kubenswrapper[4821]: I1125 11:39:41.709232 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b797317-945f-4546-9548-38cb57d0cd2b-catalog-content\") pod \"community-operators-fxjm4\" (UID: \"2b797317-945f-4546-9548-38cb57d0cd2b\") " pod="openshift-marketplace/community-operators-fxjm4" Nov 25 11:39:41 crc kubenswrapper[4821]: I1125 11:39:41.739064 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4h74\" (UniqueName: \"kubernetes.io/projected/2b797317-945f-4546-9548-38cb57d0cd2b-kube-api-access-m4h74\") pod \"community-operators-fxjm4\" (UID: \"2b797317-945f-4546-9548-38cb57d0cd2b\") " pod="openshift-marketplace/community-operators-fxjm4" Nov 25 11:39:41 crc kubenswrapper[4821]: I1125 11:39:41.848337 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fxjm4" Nov 25 11:39:42 crc kubenswrapper[4821]: I1125 11:39:42.373505 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fxjm4"] Nov 25 11:39:43 crc kubenswrapper[4821]: I1125 11:39:43.127566 4821 generic.go:334] "Generic (PLEG): container finished" podID="2b797317-945f-4546-9548-38cb57d0cd2b" containerID="2f6465c536ea2481228e8d66d3f0b4b3b4f58f7826454a2ac7d0322f1f454769" exitCode=0 Nov 25 11:39:43 crc kubenswrapper[4821]: I1125 11:39:43.127643 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fxjm4" event={"ID":"2b797317-945f-4546-9548-38cb57d0cd2b","Type":"ContainerDied","Data":"2f6465c536ea2481228e8d66d3f0b4b3b4f58f7826454a2ac7d0322f1f454769"} Nov 25 11:39:43 crc kubenswrapper[4821]: I1125 11:39:43.127926 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fxjm4" event={"ID":"2b797317-945f-4546-9548-38cb57d0cd2b","Type":"ContainerStarted","Data":"895f0202174442ff51c5734883d68107636e06e1f818ac1647c8a3fd8b833cd4"} Nov 25 11:39:45 crc kubenswrapper[4821]: I1125 11:39:45.147194 4821 generic.go:334] "Generic (PLEG): container finished" podID="2b797317-945f-4546-9548-38cb57d0cd2b" containerID="c93e01280b0473234c8b9b160c00b940efc946108332ccd8ea27e00a16de0270" exitCode=0 Nov 25 11:39:45 crc kubenswrapper[4821]: I1125 11:39:45.147296 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fxjm4" event={"ID":"2b797317-945f-4546-9548-38cb57d0cd2b","Type":"ContainerDied","Data":"c93e01280b0473234c8b9b160c00b940efc946108332ccd8ea27e00a16de0270"} Nov 25 11:39:46 crc kubenswrapper[4821]: I1125 11:39:46.160102 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fxjm4" event={"ID":"2b797317-945f-4546-9548-38cb57d0cd2b","Type":"ContainerStarted","Data":"63e2a6459c1a27e136d97791b6991df3d41b1aeefe7a84689713c4613134c539"} Nov 25 11:39:46 crc kubenswrapper[4821]: I1125 11:39:46.183340 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-fxjm4" podStartSLOduration=2.562264785 podStartE2EDuration="5.18332617s" podCreationTimestamp="2025-11-25 11:39:41 +0000 UTC" firstStartedPulling="2025-11-25 11:39:43.129684843 +0000 UTC m=+4053.666004690" lastFinishedPulling="2025-11-25 11:39:45.750746228 +0000 UTC m=+4056.287066075" observedRunningTime="2025-11-25 11:39:46.181351572 +0000 UTC m=+4056.717671419" watchObservedRunningTime="2025-11-25 11:39:46.18332617 +0000 UTC m=+4056.719646017" Nov 25 11:39:51 crc kubenswrapper[4821]: I1125 11:39:51.849072 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-fxjm4" Nov 25 11:39:51 crc kubenswrapper[4821]: I1125 11:39:51.851332 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-fxjm4" Nov 25 11:39:51 crc kubenswrapper[4821]: I1125 11:39:51.903313 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-fxjm4" Nov 25 11:39:52 crc kubenswrapper[4821]: I1125 11:39:52.114137 4821 scope.go:117] "RemoveContainer" containerID="f357cf6563c37c0c7cd86102a3e1eea3ba9146a525631fc115bab1d95f4015f3" Nov 25 11:39:52 crc kubenswrapper[4821]: E1125 11:39:52.114404 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:39:52 crc kubenswrapper[4821]: I1125 11:39:52.309489 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-fxjm4" Nov 25 11:39:52 crc kubenswrapper[4821]: I1125 11:39:52.372635 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fxjm4"] Nov 25 11:39:54 crc kubenswrapper[4821]: I1125 11:39:54.270445 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-fxjm4" podUID="2b797317-945f-4546-9548-38cb57d0cd2b" containerName="registry-server" containerID="cri-o://63e2a6459c1a27e136d97791b6991df3d41b1aeefe7a84689713c4613134c539" gracePeriod=2 Nov 25 11:39:55 crc kubenswrapper[4821]: I1125 11:39:55.280920 4821 generic.go:334] "Generic (PLEG): container finished" podID="2b797317-945f-4546-9548-38cb57d0cd2b" containerID="63e2a6459c1a27e136d97791b6991df3d41b1aeefe7a84689713c4613134c539" exitCode=0 Nov 25 11:39:55 crc kubenswrapper[4821]: I1125 11:39:55.281004 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fxjm4" event={"ID":"2b797317-945f-4546-9548-38cb57d0cd2b","Type":"ContainerDied","Data":"63e2a6459c1a27e136d97791b6991df3d41b1aeefe7a84689713c4613134c539"} Nov 25 11:39:55 crc kubenswrapper[4821]: I1125 11:39:55.281307 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fxjm4" event={"ID":"2b797317-945f-4546-9548-38cb57d0cd2b","Type":"ContainerDied","Data":"895f0202174442ff51c5734883d68107636e06e1f818ac1647c8a3fd8b833cd4"} Nov 25 11:39:55 crc kubenswrapper[4821]: I1125 11:39:55.281327 4821 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="895f0202174442ff51c5734883d68107636e06e1f818ac1647c8a3fd8b833cd4" Nov 25 11:39:55 crc kubenswrapper[4821]: I1125 11:39:55.328221 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fxjm4" Nov 25 11:39:55 crc kubenswrapper[4821]: I1125 11:39:55.413870 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m4h74\" (UniqueName: \"kubernetes.io/projected/2b797317-945f-4546-9548-38cb57d0cd2b-kube-api-access-m4h74\") pod \"2b797317-945f-4546-9548-38cb57d0cd2b\" (UID: \"2b797317-945f-4546-9548-38cb57d0cd2b\") " Nov 25 11:39:55 crc kubenswrapper[4821]: I1125 11:39:55.414350 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b797317-945f-4546-9548-38cb57d0cd2b-utilities\") pod \"2b797317-945f-4546-9548-38cb57d0cd2b\" (UID: \"2b797317-945f-4546-9548-38cb57d0cd2b\") " Nov 25 11:39:55 crc kubenswrapper[4821]: I1125 11:39:55.414425 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b797317-945f-4546-9548-38cb57d0cd2b-catalog-content\") pod \"2b797317-945f-4546-9548-38cb57d0cd2b\" (UID: \"2b797317-945f-4546-9548-38cb57d0cd2b\") " Nov 25 11:39:55 crc kubenswrapper[4821]: I1125 11:39:55.417209 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b797317-945f-4546-9548-38cb57d0cd2b-utilities" (OuterVolumeSpecName: "utilities") pod "2b797317-945f-4546-9548-38cb57d0cd2b" (UID: "2b797317-945f-4546-9548-38cb57d0cd2b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:39:55 crc kubenswrapper[4821]: I1125 11:39:55.434552 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b797317-945f-4546-9548-38cb57d0cd2b-kube-api-access-m4h74" (OuterVolumeSpecName: "kube-api-access-m4h74") pod "2b797317-945f-4546-9548-38cb57d0cd2b" (UID: "2b797317-945f-4546-9548-38cb57d0cd2b"). InnerVolumeSpecName "kube-api-access-m4h74". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:39:55 crc kubenswrapper[4821]: I1125 11:39:55.518244 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m4h74\" (UniqueName: \"kubernetes.io/projected/2b797317-945f-4546-9548-38cb57d0cd2b-kube-api-access-m4h74\") on node \"crc\" DevicePath \"\"" Nov 25 11:39:55 crc kubenswrapper[4821]: I1125 11:39:55.518298 4821 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b797317-945f-4546-9548-38cb57d0cd2b-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 11:39:55 crc kubenswrapper[4821]: I1125 11:39:55.667144 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-86dc4d89c8-pnq4j_db28ec64-d8be-4b56-a427-e4cceb256fd0/kube-rbac-proxy/0.log" Nov 25 11:39:55 crc kubenswrapper[4821]: I1125 11:39:55.752587 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-86dc4d89c8-pnq4j_db28ec64-d8be-4b56-a427-e4cceb256fd0/manager/0.log" Nov 25 11:39:55 crc kubenswrapper[4821]: I1125 11:39:55.861737 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-79856dc55c-p4bwc_8f16d64f-ea14-4853-b30b-2df294ce6cc3/kube-rbac-proxy/0.log" Nov 25 11:39:55 crc kubenswrapper[4821]: I1125 11:39:55.968472 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-79856dc55c-p4bwc_8f16d64f-ea14-4853-b30b-2df294ce6cc3/manager/0.log" Nov 25 11:39:56 crc kubenswrapper[4821]: I1125 11:39:56.013837 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-7d695c9b56-7sx9m_2236607e-2a7b-4858-b5b9-c43e47c9375d/kube-rbac-proxy/0.log" Nov 25 11:39:56 crc kubenswrapper[4821]: I1125 11:39:56.052193 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-7d695c9b56-7sx9m_2236607e-2a7b-4858-b5b9-c43e47c9375d/manager/0.log" Nov 25 11:39:56 crc kubenswrapper[4821]: I1125 11:39:56.175063 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f381f3655dcf4f93a6a93fe2921d660ff39f878b90de9b88a6ab8f46514g6tz_bd4ce1de-05e0-49f4-ad3a-7baae72f3e09/util/0.log" Nov 25 11:39:56 crc kubenswrapper[4821]: I1125 11:39:56.308240 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fxjm4" Nov 25 11:39:56 crc kubenswrapper[4821]: I1125 11:39:56.419983 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f381f3655dcf4f93a6a93fe2921d660ff39f878b90de9b88a6ab8f46514g6tz_bd4ce1de-05e0-49f4-ad3a-7baae72f3e09/pull/0.log" Nov 25 11:39:56 crc kubenswrapper[4821]: I1125 11:39:56.424632 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f381f3655dcf4f93a6a93fe2921d660ff39f878b90de9b88a6ab8f46514g6tz_bd4ce1de-05e0-49f4-ad3a-7baae72f3e09/pull/0.log" Nov 25 11:39:56 crc kubenswrapper[4821]: I1125 11:39:56.588474 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f381f3655dcf4f93a6a93fe2921d660ff39f878b90de9b88a6ab8f46514g6tz_bd4ce1de-05e0-49f4-ad3a-7baae72f3e09/extract/0.log" Nov 25 11:39:56 crc kubenswrapper[4821]: I1125 11:39:56.889653 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-68b95954c9-qgs92_a3dc5167-4997-44bf-ace3-706829fd0f17/kube-rbac-proxy/0.log" Nov 25 11:39:56 crc kubenswrapper[4821]: I1125 11:39:56.969298 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f381f3655dcf4f93a6a93fe2921d660ff39f878b90de9b88a6ab8f46514g6tz_bd4ce1de-05e0-49f4-ad3a-7baae72f3e09/util/0.log" Nov 25 11:39:56 crc kubenswrapper[4821]: I1125 11:39:56.982643 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f381f3655dcf4f93a6a93fe2921d660ff39f878b90de9b88a6ab8f46514g6tz_bd4ce1de-05e0-49f4-ad3a-7baae72f3e09/pull/0.log" Nov 25 11:39:56 crc kubenswrapper[4821]: I1125 11:39:56.989847 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f381f3655dcf4f93a6a93fe2921d660ff39f878b90de9b88a6ab8f46514g6tz_bd4ce1de-05e0-49f4-ad3a-7baae72f3e09/util/0.log" Nov 25 11:39:57 crc kubenswrapper[4821]: I1125 11:39:57.130235 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-68b95954c9-qgs92_a3dc5167-4997-44bf-ace3-706829fd0f17/manager/0.log" Nov 25 11:39:57 crc kubenswrapper[4821]: I1125 11:39:57.301829 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-774b86978c-s49f6_b35e42b0-2a6f-454f-b8e2-4f7cd9b9217f/kube-rbac-proxy/0.log" Nov 25 11:39:57 crc kubenswrapper[4821]: I1125 11:39:57.312741 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-774b86978c-s49f6_b35e42b0-2a6f-454f-b8e2-4f7cd9b9217f/manager/0.log" Nov 25 11:39:57 crc kubenswrapper[4821]: I1125 11:39:57.495828 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c9694994-2qzrg_b7895321-27ec-4c75-aa5e-581fe46e0789/kube-rbac-proxy/0.log" Nov 25 11:39:57 crc kubenswrapper[4821]: I1125 11:39:57.538744 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c9694994-2qzrg_b7895321-27ec-4c75-aa5e-581fe46e0789/manager/0.log" Nov 25 11:39:57 crc kubenswrapper[4821]: I1125 11:39:57.566450 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-d5cc86f4b-8cqfp_9db84cac-a8df-4359-8051-ebc064a97b4d/kube-rbac-proxy/0.log" Nov 25 11:39:57 crc kubenswrapper[4821]: I1125 11:39:57.778064 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5bfcdc958c-fqfwh_1976151f-2b7a-4cde-95d3-9020b17d938c/kube-rbac-proxy/0.log" Nov 25 11:39:57 crc kubenswrapper[4821]: I1125 11:39:57.815969 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-d5cc86f4b-8cqfp_9db84cac-a8df-4359-8051-ebc064a97b4d/manager/0.log" Nov 25 11:39:57 crc kubenswrapper[4821]: I1125 11:39:57.854925 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5bfcdc958c-fqfwh_1976151f-2b7a-4cde-95d3-9020b17d938c/manager/0.log" Nov 25 11:39:57 crc kubenswrapper[4821]: I1125 11:39:57.954443 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b797317-945f-4546-9548-38cb57d0cd2b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2b797317-945f-4546-9548-38cb57d0cd2b" (UID: "2b797317-945f-4546-9548-38cb57d0cd2b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:39:57 crc kubenswrapper[4821]: I1125 11:39:57.963746 4821 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b797317-945f-4546-9548-38cb57d0cd2b-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 11:39:57 crc kubenswrapper[4821]: I1125 11:39:57.971670 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-748dc6576f-f8tvv_7982734e-b887-4cea-9724-3a37f9369ac5/kube-rbac-proxy/0.log" Nov 25 11:39:58 crc kubenswrapper[4821]: I1125 11:39:58.063652 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-748dc6576f-f8tvv_7982734e-b887-4cea-9724-3a37f9369ac5/manager/0.log" Nov 25 11:39:58 crc kubenswrapper[4821]: I1125 11:39:58.142889 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fxjm4"] Nov 25 11:39:58 crc kubenswrapper[4821]: I1125 11:39:58.153071 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-fxjm4"] Nov 25 11:39:58 crc kubenswrapper[4821]: I1125 11:39:58.208722 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-58bb8d67cc-vnf9x_5fefd5f2-18cb-4c63-848f-2fc9d63bb828/manager/0.log" Nov 25 11:39:58 crc kubenswrapper[4821]: I1125 11:39:58.245738 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-58bb8d67cc-vnf9x_5fefd5f2-18cb-4c63-848f-2fc9d63bb828/kube-rbac-proxy/0.log" Nov 25 11:39:58 crc kubenswrapper[4821]: I1125 11:39:58.329292 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-cb6c4fdb7-fw96v_d42d0991-9e6f-46d0-9236-00e85c26a603/kube-rbac-proxy/0.log" Nov 25 11:39:58 crc kubenswrapper[4821]: I1125 11:39:58.450404 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-cb6c4fdb7-fw96v_d42d0991-9e6f-46d0-9236-00e85c26a603/manager/0.log" Nov 25 11:39:58 crc kubenswrapper[4821]: I1125 11:39:58.507187 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-7c57c8bbc4-hwp4x_c0376a5f-ed42-4399-8f41-2d2157a654fd/kube-rbac-proxy/0.log" Nov 25 11:39:58 crc kubenswrapper[4821]: I1125 11:39:58.594556 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-7c57c8bbc4-hwp4x_c0376a5f-ed42-4399-8f41-2d2157a654fd/manager/0.log" Nov 25 11:39:58 crc kubenswrapper[4821]: I1125 11:39:58.687178 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-79556f57fc-4ftbp_1f69797f-4204-4989-a00b-d9722d44d3c6/kube-rbac-proxy/0.log" Nov 25 11:39:58 crc kubenswrapper[4821]: I1125 11:39:58.797286 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-79556f57fc-4ftbp_1f69797f-4204-4989-a00b-d9722d44d3c6/manager/0.log" Nov 25 11:39:58 crc kubenswrapper[4821]: I1125 11:39:58.895840 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-fd75fd47d-5ncqq_db1d55a6-1285-429d-b9dd-b2e100e602c6/manager/0.log" Nov 25 11:39:58 crc kubenswrapper[4821]: I1125 11:39:58.906567 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-fd75fd47d-5ncqq_db1d55a6-1285-429d-b9dd-b2e100e602c6/kube-rbac-proxy/0.log" Nov 25 11:39:59 crc kubenswrapper[4821]: I1125 11:39:59.068655 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-544b9bb9-lwp9z_14385dd9-9b06-473b-919d-9d2f9207ffe1/kube-rbac-proxy/0.log" Nov 25 11:39:59 crc kubenswrapper[4821]: I1125 11:39:59.068970 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-544b9bb9-lwp9z_14385dd9-9b06-473b-919d-9d2f9207ffe1/manager/0.log" Nov 25 11:39:59 crc kubenswrapper[4821]: I1125 11:39:59.458557 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-7f8f598fc5-8gsxh_1875cb8b-6a5c-48dc-bf01-07eef1d129fe/operator/0.log" Nov 25 11:39:59 crc kubenswrapper[4821]: I1125 11:39:59.472074 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-tzjrl_afdcf8ff-544a-4625-b72e-d1a880434014/registry-server/0.log" Nov 25 11:39:59 crc kubenswrapper[4821]: I1125 11:39:59.580524 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-66cf5c67ff-qxgcp_2b44040a-e224-4468-a528-9f96a98e3891/kube-rbac-proxy/0.log" Nov 25 11:39:59 crc kubenswrapper[4821]: I1125 11:39:59.768534 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5db546f9d9-66kf4_876c088c-7ee5-4b61-a798-eb36ac1e7e21/kube-rbac-proxy/0.log" Nov 25 11:39:59 crc kubenswrapper[4821]: I1125 11:39:59.789577 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-66cf5c67ff-qxgcp_2b44040a-e224-4468-a528-9f96a98e3891/manager/0.log" Nov 25 11:39:59 crc kubenswrapper[4821]: I1125 11:39:59.875655 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5db546f9d9-66kf4_876c088c-7ee5-4b61-a798-eb36ac1e7e21/manager/0.log" Nov 25 11:40:00 crc kubenswrapper[4821]: I1125 11:40:00.018257 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-fprdh_5cea2a2e-a19f-42fc-ac84-aadfa677ec4c/operator/0.log" Nov 25 11:40:00 crc kubenswrapper[4821]: I1125 11:40:00.126046 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6fdc4fcf86-75q4l_c77cb0f4-1dff-466c-a8bf-6a987bc87ced/kube-rbac-proxy/0.log" Nov 25 11:40:00 crc kubenswrapper[4821]: I1125 11:40:00.128564 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b797317-945f-4546-9548-38cb57d0cd2b" path="/var/lib/kubelet/pods/2b797317-945f-4546-9548-38cb57d0cd2b/volumes" Nov 25 11:40:00 crc kubenswrapper[4821]: I1125 11:40:00.220747 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6fdc4fcf86-75q4l_c77cb0f4-1dff-466c-a8bf-6a987bc87ced/manager/0.log" Nov 25 11:40:00 crc kubenswrapper[4821]: I1125 11:40:00.285317 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-8dbc5685d-gcl85_a64e88ce-fe1a-4cb1-984a-c6d8b5e5ffd9/kube-rbac-proxy/0.log" Nov 25 11:40:00 crc kubenswrapper[4821]: I1125 11:40:00.443632 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-64b4f4bf8b-vzb6r_40e2ab14-9d05-4569-8a3a-69a34586c504/manager/0.log" Nov 25 11:40:00 crc kubenswrapper[4821]: I1125 11:40:00.456325 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-8dbc5685d-gcl85_a64e88ce-fe1a-4cb1-984a-c6d8b5e5ffd9/manager/0.log" Nov 25 11:40:00 crc kubenswrapper[4821]: I1125 11:40:00.488868 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cb74df96-lppg4_8fef5c2e-676a-49b6-a10c-94abe226600a/kube-rbac-proxy/0.log" Nov 25 11:40:00 crc kubenswrapper[4821]: I1125 11:40:00.503687 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cb74df96-lppg4_8fef5c2e-676a-49b6-a10c-94abe226600a/manager/0.log" Nov 25 11:40:00 crc kubenswrapper[4821]: I1125 11:40:00.645720 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-864885998-vnmtq_505faacd-65f5-4078-9867-9ae648652b50/kube-rbac-proxy/0.log" Nov 25 11:40:00 crc kubenswrapper[4821]: I1125 11:40:00.666794 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-864885998-vnmtq_505faacd-65f5-4078-9867-9ae648652b50/manager/0.log" Nov 25 11:40:03 crc kubenswrapper[4821]: I1125 11:40:03.114060 4821 scope.go:117] "RemoveContainer" containerID="f357cf6563c37c0c7cd86102a3e1eea3ba9146a525631fc115bab1d95f4015f3" Nov 25 11:40:03 crc kubenswrapper[4821]: E1125 11:40:03.114809 4821 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2krbf_openshift-machine-config-operator(5f948b87-ac86-4de6-ad64-c2ef947f84d4)\"" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" Nov 25 11:40:17 crc kubenswrapper[4821]: I1125 11:40:17.114651 4821 scope.go:117] "RemoveContainer" containerID="f357cf6563c37c0c7cd86102a3e1eea3ba9146a525631fc115bab1d95f4015f3" Nov 25 11:40:17 crc kubenswrapper[4821]: I1125 11:40:17.479406 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" event={"ID":"5f948b87-ac86-4de6-ad64-c2ef947f84d4","Type":"ContainerStarted","Data":"b90a0504e63aa3c9d24899ca1bb330848b4b0bc00ba7e2363f96654d9d7aca46"} Nov 25 11:40:17 crc kubenswrapper[4821]: I1125 11:40:17.671452 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-2stv6_4a74ce9e-b145-4e40-ad4e-67ecdf722195/control-plane-machine-set-operator/0.log" Nov 25 11:40:17 crc kubenswrapper[4821]: I1125 11:40:17.861422 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-zj4s2_24708c13-efe2-435b-acde-b29b48871f16/kube-rbac-proxy/0.log" Nov 25 11:40:17 crc kubenswrapper[4821]: I1125 11:40:17.875113 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-zj4s2_24708c13-efe2-435b-acde-b29b48871f16/machine-api-operator/0.log" Nov 25 11:40:29 crc kubenswrapper[4821]: I1125 11:40:29.141747 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-crm7d_c7f95e74-b964-455f-bacd-60e1fceb87fb/cert-manager-controller/0.log" Nov 25 11:40:29 crc kubenswrapper[4821]: I1125 11:40:29.359832 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-g7gr9_baa81eb8-c73e-41d1-bd01-9958477aa031/cert-manager-cainjector/0.log" Nov 25 11:40:29 crc kubenswrapper[4821]: I1125 11:40:29.417920 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-4qh5c_4096527e-0fc1-4991-ad0c-f47a84cd6e84/cert-manager-webhook/0.log" Nov 25 11:40:42 crc kubenswrapper[4821]: I1125 11:40:42.163068 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-5874bd7bc5-7mnfd_dff35e42-91ef-41f1-81b9-c12191388f31/nmstate-console-plugin/0.log" Nov 25 11:40:42 crc kubenswrapper[4821]: I1125 11:40:42.397463 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-89r7d_e29f6151-8a55-4b3e-8a15-8f8ec4b5ac66/nmstate-handler/0.log" Nov 25 11:40:42 crc kubenswrapper[4821]: I1125 11:40:42.462995 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-79cx4_221a1f74-bcbb-4788-83d1-4c5c92d3fedd/kube-rbac-proxy/0.log" Nov 25 11:40:42 crc kubenswrapper[4821]: I1125 11:40:42.495527 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-79cx4_221a1f74-bcbb-4788-83d1-4c5c92d3fedd/nmstate-metrics/0.log" Nov 25 11:40:42 crc kubenswrapper[4821]: I1125 11:40:42.650467 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-557fdffb88-d8wjl_7218bb0e-8630-47d7-a259-3d4f79f1a470/nmstate-operator/0.log" Nov 25 11:40:42 crc kubenswrapper[4821]: I1125 11:40:42.714484 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6b89b748d8-c9bsx_de41e739-6f86-44cc-80a2-99bb874d259f/nmstate-webhook/0.log" Nov 25 11:40:57 crc kubenswrapper[4821]: I1125 11:40:57.085125 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-5knnl_902ced61-7b2c-4a77-b8fc-ea9a226e05ba/kube-rbac-proxy/0.log" Nov 25 11:40:57 crc kubenswrapper[4821]: I1125 11:40:57.153538 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-5knnl_902ced61-7b2c-4a77-b8fc-ea9a226e05ba/controller/0.log" Nov 25 11:40:57 crc kubenswrapper[4821]: I1125 11:40:57.282926 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7knkm_0a268005-d45e-4cb1-b04a-5bedb97ea5ee/cp-frr-files/0.log" Nov 25 11:40:57 crc kubenswrapper[4821]: I1125 11:40:57.414734 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7knkm_0a268005-d45e-4cb1-b04a-5bedb97ea5ee/cp-frr-files/0.log" Nov 25 11:40:57 crc kubenswrapper[4821]: I1125 11:40:57.452879 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7knkm_0a268005-d45e-4cb1-b04a-5bedb97ea5ee/cp-reloader/0.log" Nov 25 11:40:57 crc kubenswrapper[4821]: I1125 11:40:57.454713 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7knkm_0a268005-d45e-4cb1-b04a-5bedb97ea5ee/cp-metrics/0.log" Nov 25 11:40:57 crc kubenswrapper[4821]: I1125 11:40:57.479509 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7knkm_0a268005-d45e-4cb1-b04a-5bedb97ea5ee/cp-reloader/0.log" Nov 25 11:40:57 crc kubenswrapper[4821]: I1125 11:40:57.730513 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7knkm_0a268005-d45e-4cb1-b04a-5bedb97ea5ee/cp-reloader/0.log" Nov 25 11:40:57 crc kubenswrapper[4821]: I1125 11:40:57.735223 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7knkm_0a268005-d45e-4cb1-b04a-5bedb97ea5ee/cp-metrics/0.log" Nov 25 11:40:57 crc kubenswrapper[4821]: I1125 11:40:57.773870 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7knkm_0a268005-d45e-4cb1-b04a-5bedb97ea5ee/cp-frr-files/0.log" Nov 25 11:40:57 crc kubenswrapper[4821]: I1125 11:40:57.779724 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7knkm_0a268005-d45e-4cb1-b04a-5bedb97ea5ee/cp-metrics/0.log" Nov 25 11:40:57 crc kubenswrapper[4821]: I1125 11:40:57.911336 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7knkm_0a268005-d45e-4cb1-b04a-5bedb97ea5ee/cp-reloader/0.log" Nov 25 11:40:57 crc kubenswrapper[4821]: I1125 11:40:57.940698 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7knkm_0a268005-d45e-4cb1-b04a-5bedb97ea5ee/cp-frr-files/0.log" Nov 25 11:40:57 crc kubenswrapper[4821]: I1125 11:40:57.949652 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7knkm_0a268005-d45e-4cb1-b04a-5bedb97ea5ee/controller/0.log" Nov 25 11:40:57 crc kubenswrapper[4821]: I1125 11:40:57.952309 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7knkm_0a268005-d45e-4cb1-b04a-5bedb97ea5ee/cp-metrics/0.log" Nov 25 11:40:58 crc kubenswrapper[4821]: I1125 11:40:58.111092 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7knkm_0a268005-d45e-4cb1-b04a-5bedb97ea5ee/kube-rbac-proxy/0.log" Nov 25 11:40:58 crc kubenswrapper[4821]: I1125 11:40:58.132522 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7knkm_0a268005-d45e-4cb1-b04a-5bedb97ea5ee/kube-rbac-proxy-frr/0.log" Nov 25 11:40:58 crc kubenswrapper[4821]: I1125 11:40:58.154066 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7knkm_0a268005-d45e-4cb1-b04a-5bedb97ea5ee/frr-metrics/0.log" Nov 25 11:40:58 crc kubenswrapper[4821]: I1125 11:40:58.312016 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7knkm_0a268005-d45e-4cb1-b04a-5bedb97ea5ee/reloader/0.log" Nov 25 11:40:58 crc kubenswrapper[4821]: I1125 11:40:58.351354 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-6998585d5-bclxv_1144ecf2-7246-4c05-a7ca-f7b7213092fd/frr-k8s-webhook-server/0.log" Nov 25 11:40:58 crc kubenswrapper[4821]: I1125 11:40:58.581434 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-644d88f8f5-gtq55_b1969fcd-bedc-42dc-966d-9cd5b8770fd9/manager/0.log" Nov 25 11:40:58 crc kubenswrapper[4821]: I1125 11:40:58.783011 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-5qbb4_c8b12b33-c9ca-48dd-9f7f-96b5f1c40f86/kube-rbac-proxy/0.log" Nov 25 11:40:58 crc kubenswrapper[4821]: I1125 11:40:58.846147 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-666cb5dd59-l9nfz_83c8e5d6-48b9-47ff-8c0a-aad92bbe511a/webhook-server/0.log" Nov 25 11:40:59 crc kubenswrapper[4821]: I1125 11:40:59.410074 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-5qbb4_c8b12b33-c9ca-48dd-9f7f-96b5f1c40f86/speaker/0.log" Nov 25 11:40:59 crc kubenswrapper[4821]: I1125 11:40:59.548686 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7knkm_0a268005-d45e-4cb1-b04a-5bedb97ea5ee/frr/0.log" Nov 25 11:41:11 crc kubenswrapper[4821]: I1125 11:41:11.373355 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6jtmt_faf2c05e-02b4-4b92-b614-d3de8e443677/util/0.log" Nov 25 11:41:11 crc kubenswrapper[4821]: I1125 11:41:11.564701 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6jtmt_faf2c05e-02b4-4b92-b614-d3de8e443677/util/0.log" Nov 25 11:41:11 crc kubenswrapper[4821]: I1125 11:41:11.589177 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6jtmt_faf2c05e-02b4-4b92-b614-d3de8e443677/pull/0.log" Nov 25 11:41:11 crc kubenswrapper[4821]: I1125 11:41:11.611452 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6jtmt_faf2c05e-02b4-4b92-b614-d3de8e443677/pull/0.log" Nov 25 11:41:11 crc kubenswrapper[4821]: I1125 11:41:11.766577 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6jtmt_faf2c05e-02b4-4b92-b614-d3de8e443677/util/0.log" Nov 25 11:41:11 crc kubenswrapper[4821]: I1125 11:41:11.786828 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6jtmt_faf2c05e-02b4-4b92-b614-d3de8e443677/extract/0.log" Nov 25 11:41:11 crc kubenswrapper[4821]: I1125 11:41:11.788626 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6jtmt_faf2c05e-02b4-4b92-b614-d3de8e443677/pull/0.log" Nov 25 11:41:11 crc kubenswrapper[4821]: I1125 11:41:11.947818 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-529qb_410373d3-842e-43b3-a2d8-2293c543a4ac/extract-utilities/0.log" Nov 25 11:41:12 crc kubenswrapper[4821]: I1125 11:41:12.120459 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-529qb_410373d3-842e-43b3-a2d8-2293c543a4ac/extract-content/0.log" Nov 25 11:41:12 crc kubenswrapper[4821]: I1125 11:41:12.127327 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-529qb_410373d3-842e-43b3-a2d8-2293c543a4ac/extract-content/0.log" Nov 25 11:41:12 crc kubenswrapper[4821]: I1125 11:41:12.130436 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-529qb_410373d3-842e-43b3-a2d8-2293c543a4ac/extract-utilities/0.log" Nov 25 11:41:12 crc kubenswrapper[4821]: I1125 11:41:12.313509 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-529qb_410373d3-842e-43b3-a2d8-2293c543a4ac/extract-utilities/0.log" Nov 25 11:41:12 crc kubenswrapper[4821]: I1125 11:41:12.315126 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-529qb_410373d3-842e-43b3-a2d8-2293c543a4ac/extract-content/0.log" Nov 25 11:41:12 crc kubenswrapper[4821]: I1125 11:41:12.533778 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4j8hk_a43c28e0-8883-4567-a86f-bcd311098876/extract-utilities/0.log" Nov 25 11:41:12 crc kubenswrapper[4821]: I1125 11:41:12.565540 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-529qb_410373d3-842e-43b3-a2d8-2293c543a4ac/registry-server/0.log" Nov 25 11:41:12 crc kubenswrapper[4821]: I1125 11:41:12.737001 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4j8hk_a43c28e0-8883-4567-a86f-bcd311098876/extract-content/0.log" Nov 25 11:41:12 crc kubenswrapper[4821]: I1125 11:41:12.753349 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4j8hk_a43c28e0-8883-4567-a86f-bcd311098876/extract-utilities/0.log" Nov 25 11:41:12 crc kubenswrapper[4821]: I1125 11:41:12.786783 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4j8hk_a43c28e0-8883-4567-a86f-bcd311098876/extract-content/0.log" Nov 25 11:41:12 crc kubenswrapper[4821]: I1125 11:41:12.899646 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4j8hk_a43c28e0-8883-4567-a86f-bcd311098876/extract-utilities/0.log" Nov 25 11:41:12 crc kubenswrapper[4821]: I1125 11:41:12.911106 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4j8hk_a43c28e0-8883-4567-a86f-bcd311098876/extract-content/0.log" Nov 25 11:41:13 crc kubenswrapper[4821]: I1125 11:41:13.099563 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6hsmqh_8c67f775-e13f-4c7d-9d2b-115d492961f1/util/0.log" Nov 25 11:41:13 crc kubenswrapper[4821]: I1125 11:41:13.314698 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6hsmqh_8c67f775-e13f-4c7d-9d2b-115d492961f1/pull/0.log" Nov 25 11:41:13 crc kubenswrapper[4821]: I1125 11:41:13.318300 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6hsmqh_8c67f775-e13f-4c7d-9d2b-115d492961f1/pull/0.log" Nov 25 11:41:13 crc kubenswrapper[4821]: I1125 11:41:13.392691 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6hsmqh_8c67f775-e13f-4c7d-9d2b-115d492961f1/util/0.log" Nov 25 11:41:13 crc kubenswrapper[4821]: I1125 11:41:13.748403 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4j8hk_a43c28e0-8883-4567-a86f-bcd311098876/registry-server/0.log" Nov 25 11:41:14 crc kubenswrapper[4821]: I1125 11:41:14.091107 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6hsmqh_8c67f775-e13f-4c7d-9d2b-115d492961f1/util/0.log" Nov 25 11:41:14 crc kubenswrapper[4821]: I1125 11:41:14.104411 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6hsmqh_8c67f775-e13f-4c7d-9d2b-115d492961f1/extract/0.log" Nov 25 11:41:14 crc kubenswrapper[4821]: I1125 11:41:14.127021 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6hsmqh_8c67f775-e13f-4c7d-9d2b-115d492961f1/pull/0.log" Nov 25 11:41:14 crc kubenswrapper[4821]: I1125 11:41:14.273295 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-tdctv_05106897-6fd0-462c-9d26-a832f1385e04/marketplace-operator/0.log" Nov 25 11:41:14 crc kubenswrapper[4821]: I1125 11:41:14.335890 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m99sj_8a0df682-fa69-4ef1-a733-214abe72f1ad/extract-utilities/0.log" Nov 25 11:41:14 crc kubenswrapper[4821]: I1125 11:41:14.521098 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m99sj_8a0df682-fa69-4ef1-a733-214abe72f1ad/extract-content/0.log" Nov 25 11:41:14 crc kubenswrapper[4821]: I1125 11:41:14.521908 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m99sj_8a0df682-fa69-4ef1-a733-214abe72f1ad/extract-utilities/0.log" Nov 25 11:41:14 crc kubenswrapper[4821]: I1125 11:41:14.522258 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m99sj_8a0df682-fa69-4ef1-a733-214abe72f1ad/extract-content/0.log" Nov 25 11:41:14 crc kubenswrapper[4821]: I1125 11:41:14.705852 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m99sj_8a0df682-fa69-4ef1-a733-214abe72f1ad/extract-utilities/0.log" Nov 25 11:41:14 crc kubenswrapper[4821]: I1125 11:41:14.735487 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m99sj_8a0df682-fa69-4ef1-a733-214abe72f1ad/extract-content/0.log" Nov 25 11:41:14 crc kubenswrapper[4821]: I1125 11:41:14.837577 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m99sj_8a0df682-fa69-4ef1-a733-214abe72f1ad/registry-server/0.log" Nov 25 11:41:14 crc kubenswrapper[4821]: I1125 11:41:14.931973 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-4b5tq_45c9ecf9-7f75-4fbc-983c-5f869ed308ac/extract-utilities/0.log" Nov 25 11:41:15 crc kubenswrapper[4821]: I1125 11:41:15.491554 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-4b5tq_45c9ecf9-7f75-4fbc-983c-5f869ed308ac/extract-content/0.log" Nov 25 11:41:15 crc kubenswrapper[4821]: I1125 11:41:15.510362 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-4b5tq_45c9ecf9-7f75-4fbc-983c-5f869ed308ac/extract-utilities/0.log" Nov 25 11:41:15 crc kubenswrapper[4821]: I1125 11:41:15.523903 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-4b5tq_45c9ecf9-7f75-4fbc-983c-5f869ed308ac/extract-content/0.log" Nov 25 11:41:15 crc kubenswrapper[4821]: I1125 11:41:15.666577 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-4b5tq_45c9ecf9-7f75-4fbc-983c-5f869ed308ac/extract-content/0.log" Nov 25 11:41:15 crc kubenswrapper[4821]: I1125 11:41:15.668823 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-4b5tq_45c9ecf9-7f75-4fbc-983c-5f869ed308ac/extract-utilities/0.log" Nov 25 11:41:15 crc kubenswrapper[4821]: I1125 11:41:15.820140 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-4b5tq_45c9ecf9-7f75-4fbc-983c-5f869ed308ac/registry-server/0.log" Nov 25 11:41:35 crc kubenswrapper[4821]: E1125 11:41:35.499637 4821 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.129.56.252:48830->38.129.56.252:37639: write tcp 38.129.56.252:48830->38.129.56.252:37639: write: broken pipe Nov 25 11:42:34 crc kubenswrapper[4821]: I1125 11:42:34.714439 4821 patch_prober.go:28] interesting pod/machine-config-daemon-2krbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:42:34 crc kubenswrapper[4821]: I1125 11:42:34.715101 4821 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:42:59 crc kubenswrapper[4821]: I1125 11:42:59.921259 4821 generic.go:334] "Generic (PLEG): container finished" podID="61c863b6-7476-4c5d-b59c-40735554a04e" containerID="03c70625987368ad0a26a2cc86962944984ace1e191066e2b4ad1293b5d73adc" exitCode=0 Nov 25 11:42:59 crc kubenswrapper[4821]: I1125 11:42:59.921328 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gdrrz/must-gather-xxxft" event={"ID":"61c863b6-7476-4c5d-b59c-40735554a04e","Type":"ContainerDied","Data":"03c70625987368ad0a26a2cc86962944984ace1e191066e2b4ad1293b5d73adc"} Nov 25 11:42:59 crc kubenswrapper[4821]: I1125 11:42:59.923641 4821 scope.go:117] "RemoveContainer" containerID="03c70625987368ad0a26a2cc86962944984ace1e191066e2b4ad1293b5d73adc" Nov 25 11:43:00 crc kubenswrapper[4821]: I1125 11:43:00.132569 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-gdrrz_must-gather-xxxft_61c863b6-7476-4c5d-b59c-40735554a04e/gather/0.log" Nov 25 11:43:04 crc kubenswrapper[4821]: I1125 11:43:04.714447 4821 patch_prober.go:28] interesting pod/machine-config-daemon-2krbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:43:04 crc kubenswrapper[4821]: I1125 11:43:04.715717 4821 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:43:09 crc kubenswrapper[4821]: I1125 11:43:09.926282 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-gdrrz/must-gather-xxxft"] Nov 25 11:43:09 crc kubenswrapper[4821]: I1125 11:43:09.927018 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-gdrrz/must-gather-xxxft" podUID="61c863b6-7476-4c5d-b59c-40735554a04e" containerName="copy" containerID="cri-o://0e9e48de0f9fa74d65eb2152de4e317c5a1c3e27a8059c7bc775e2fe6c49ca7a" gracePeriod=2 Nov 25 11:43:09 crc kubenswrapper[4821]: I1125 11:43:09.937493 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-gdrrz/must-gather-xxxft"] Nov 25 11:43:10 crc kubenswrapper[4821]: I1125 11:43:10.062578 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-gdrrz_must-gather-xxxft_61c863b6-7476-4c5d-b59c-40735554a04e/copy/0.log" Nov 25 11:43:10 crc kubenswrapper[4821]: I1125 11:43:10.063396 4821 generic.go:334] "Generic (PLEG): container finished" podID="61c863b6-7476-4c5d-b59c-40735554a04e" containerID="0e9e48de0f9fa74d65eb2152de4e317c5a1c3e27a8059c7bc775e2fe6c49ca7a" exitCode=143 Nov 25 11:43:10 crc kubenswrapper[4821]: I1125 11:43:10.700372 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-gdrrz_must-gather-xxxft_61c863b6-7476-4c5d-b59c-40735554a04e/copy/0.log" Nov 25 11:43:10 crc kubenswrapper[4821]: I1125 11:43:10.701053 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gdrrz/must-gather-xxxft" Nov 25 11:43:10 crc kubenswrapper[4821]: I1125 11:43:10.818431 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vnh5p\" (UniqueName: \"kubernetes.io/projected/61c863b6-7476-4c5d-b59c-40735554a04e-kube-api-access-vnh5p\") pod \"61c863b6-7476-4c5d-b59c-40735554a04e\" (UID: \"61c863b6-7476-4c5d-b59c-40735554a04e\") " Nov 25 11:43:10 crc kubenswrapper[4821]: I1125 11:43:10.818483 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/61c863b6-7476-4c5d-b59c-40735554a04e-must-gather-output\") pod \"61c863b6-7476-4c5d-b59c-40735554a04e\" (UID: \"61c863b6-7476-4c5d-b59c-40735554a04e\") " Nov 25 11:43:10 crc kubenswrapper[4821]: I1125 11:43:10.847481 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61c863b6-7476-4c5d-b59c-40735554a04e-kube-api-access-vnh5p" (OuterVolumeSpecName: "kube-api-access-vnh5p") pod "61c863b6-7476-4c5d-b59c-40735554a04e" (UID: "61c863b6-7476-4c5d-b59c-40735554a04e"). InnerVolumeSpecName "kube-api-access-vnh5p". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:43:10 crc kubenswrapper[4821]: I1125 11:43:10.920482 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vnh5p\" (UniqueName: \"kubernetes.io/projected/61c863b6-7476-4c5d-b59c-40735554a04e-kube-api-access-vnh5p\") on node \"crc\" DevicePath \"\"" Nov 25 11:43:10 crc kubenswrapper[4821]: I1125 11:43:10.962798 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/61c863b6-7476-4c5d-b59c-40735554a04e-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "61c863b6-7476-4c5d-b59c-40735554a04e" (UID: "61c863b6-7476-4c5d-b59c-40735554a04e"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:43:11 crc kubenswrapper[4821]: I1125 11:43:11.022378 4821 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/61c863b6-7476-4c5d-b59c-40735554a04e-must-gather-output\") on node \"crc\" DevicePath \"\"" Nov 25 11:43:11 crc kubenswrapper[4821]: I1125 11:43:11.073070 4821 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-gdrrz_must-gather-xxxft_61c863b6-7476-4c5d-b59c-40735554a04e/copy/0.log" Nov 25 11:43:11 crc kubenswrapper[4821]: I1125 11:43:11.074494 4821 scope.go:117] "RemoveContainer" containerID="0e9e48de0f9fa74d65eb2152de4e317c5a1c3e27a8059c7bc775e2fe6c49ca7a" Nov 25 11:43:11 crc kubenswrapper[4821]: I1125 11:43:11.074670 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gdrrz/must-gather-xxxft" Nov 25 11:43:11 crc kubenswrapper[4821]: I1125 11:43:11.109360 4821 scope.go:117] "RemoveContainer" containerID="03c70625987368ad0a26a2cc86962944984ace1e191066e2b4ad1293b5d73adc" Nov 25 11:43:12 crc kubenswrapper[4821]: I1125 11:43:12.124704 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61c863b6-7476-4c5d-b59c-40735554a04e" path="/var/lib/kubelet/pods/61c863b6-7476-4c5d-b59c-40735554a04e/volumes" Nov 25 11:43:12 crc kubenswrapper[4821]: I1125 11:43:12.935474 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5d8zv"] Nov 25 11:43:12 crc kubenswrapper[4821]: E1125 11:43:12.936268 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b797317-945f-4546-9548-38cb57d0cd2b" containerName="extract-utilities" Nov 25 11:43:12 crc kubenswrapper[4821]: I1125 11:43:12.936305 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b797317-945f-4546-9548-38cb57d0cd2b" containerName="extract-utilities" Nov 25 11:43:12 crc kubenswrapper[4821]: E1125 11:43:12.936337 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b797317-945f-4546-9548-38cb57d0cd2b" containerName="extract-content" Nov 25 11:43:12 crc kubenswrapper[4821]: I1125 11:43:12.936351 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b797317-945f-4546-9548-38cb57d0cd2b" containerName="extract-content" Nov 25 11:43:12 crc kubenswrapper[4821]: E1125 11:43:12.936371 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61c863b6-7476-4c5d-b59c-40735554a04e" containerName="copy" Nov 25 11:43:12 crc kubenswrapper[4821]: I1125 11:43:12.936385 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="61c863b6-7476-4c5d-b59c-40735554a04e" containerName="copy" Nov 25 11:43:12 crc kubenswrapper[4821]: E1125 11:43:12.936413 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b797317-945f-4546-9548-38cb57d0cd2b" containerName="registry-server" Nov 25 11:43:12 crc kubenswrapper[4821]: I1125 11:43:12.936427 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b797317-945f-4546-9548-38cb57d0cd2b" containerName="registry-server" Nov 25 11:43:12 crc kubenswrapper[4821]: E1125 11:43:12.936462 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61c863b6-7476-4c5d-b59c-40735554a04e" containerName="gather" Nov 25 11:43:12 crc kubenswrapper[4821]: I1125 11:43:12.936475 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="61c863b6-7476-4c5d-b59c-40735554a04e" containerName="gather" Nov 25 11:43:12 crc kubenswrapper[4821]: I1125 11:43:12.936810 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="61c863b6-7476-4c5d-b59c-40735554a04e" containerName="gather" Nov 25 11:43:12 crc kubenswrapper[4821]: I1125 11:43:12.936848 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b797317-945f-4546-9548-38cb57d0cd2b" containerName="registry-server" Nov 25 11:43:12 crc kubenswrapper[4821]: I1125 11:43:12.936889 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="61c863b6-7476-4c5d-b59c-40735554a04e" containerName="copy" Nov 25 11:43:12 crc kubenswrapper[4821]: I1125 11:43:12.940045 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5d8zv" Nov 25 11:43:12 crc kubenswrapper[4821]: I1125 11:43:12.946907 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5d8zv"] Nov 25 11:43:13 crc kubenswrapper[4821]: I1125 11:43:13.002011 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2nqn\" (UniqueName: \"kubernetes.io/projected/5c6f0100-2dec-45c5-95e4-229b499d8d7f-kube-api-access-n2nqn\") pod \"certified-operators-5d8zv\" (UID: \"5c6f0100-2dec-45c5-95e4-229b499d8d7f\") " pod="openshift-marketplace/certified-operators-5d8zv" Nov 25 11:43:13 crc kubenswrapper[4821]: I1125 11:43:13.002498 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c6f0100-2dec-45c5-95e4-229b499d8d7f-catalog-content\") pod \"certified-operators-5d8zv\" (UID: \"5c6f0100-2dec-45c5-95e4-229b499d8d7f\") " pod="openshift-marketplace/certified-operators-5d8zv" Nov 25 11:43:13 crc kubenswrapper[4821]: I1125 11:43:13.002535 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c6f0100-2dec-45c5-95e4-229b499d8d7f-utilities\") pod \"certified-operators-5d8zv\" (UID: \"5c6f0100-2dec-45c5-95e4-229b499d8d7f\") " pod="openshift-marketplace/certified-operators-5d8zv" Nov 25 11:43:13 crc kubenswrapper[4821]: I1125 11:43:13.103721 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n2nqn\" (UniqueName: \"kubernetes.io/projected/5c6f0100-2dec-45c5-95e4-229b499d8d7f-kube-api-access-n2nqn\") pod \"certified-operators-5d8zv\" (UID: \"5c6f0100-2dec-45c5-95e4-229b499d8d7f\") " pod="openshift-marketplace/certified-operators-5d8zv" Nov 25 11:43:13 crc kubenswrapper[4821]: I1125 11:43:13.103905 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c6f0100-2dec-45c5-95e4-229b499d8d7f-catalog-content\") pod \"certified-operators-5d8zv\" (UID: \"5c6f0100-2dec-45c5-95e4-229b499d8d7f\") " pod="openshift-marketplace/certified-operators-5d8zv" Nov 25 11:43:13 crc kubenswrapper[4821]: I1125 11:43:13.103943 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c6f0100-2dec-45c5-95e4-229b499d8d7f-utilities\") pod \"certified-operators-5d8zv\" (UID: \"5c6f0100-2dec-45c5-95e4-229b499d8d7f\") " pod="openshift-marketplace/certified-operators-5d8zv" Nov 25 11:43:13 crc kubenswrapper[4821]: I1125 11:43:13.104427 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c6f0100-2dec-45c5-95e4-229b499d8d7f-utilities\") pod \"certified-operators-5d8zv\" (UID: \"5c6f0100-2dec-45c5-95e4-229b499d8d7f\") " pod="openshift-marketplace/certified-operators-5d8zv" Nov 25 11:43:13 crc kubenswrapper[4821]: I1125 11:43:13.104952 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c6f0100-2dec-45c5-95e4-229b499d8d7f-catalog-content\") pod \"certified-operators-5d8zv\" (UID: \"5c6f0100-2dec-45c5-95e4-229b499d8d7f\") " pod="openshift-marketplace/certified-operators-5d8zv" Nov 25 11:43:13 crc kubenswrapper[4821]: I1125 11:43:13.134816 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2nqn\" (UniqueName: \"kubernetes.io/projected/5c6f0100-2dec-45c5-95e4-229b499d8d7f-kube-api-access-n2nqn\") pod \"certified-operators-5d8zv\" (UID: \"5c6f0100-2dec-45c5-95e4-229b499d8d7f\") " pod="openshift-marketplace/certified-operators-5d8zv" Nov 25 11:43:13 crc kubenswrapper[4821]: I1125 11:43:13.265478 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5d8zv" Nov 25 11:43:13 crc kubenswrapper[4821]: I1125 11:43:13.833441 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5d8zv"] Nov 25 11:43:14 crc kubenswrapper[4821]: I1125 11:43:14.104577 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5d8zv" event={"ID":"5c6f0100-2dec-45c5-95e4-229b499d8d7f","Type":"ContainerStarted","Data":"f9144ee5813a9b07d8298e4c1defe9186645343a7779e23beeffba33d5ebfab9"} Nov 25 11:43:15 crc kubenswrapper[4821]: I1125 11:43:15.117649 4821 generic.go:334] "Generic (PLEG): container finished" podID="5c6f0100-2dec-45c5-95e4-229b499d8d7f" containerID="788f950095ec9f86466843e0f119d29c31392508319701233925306ac58877ca" exitCode=0 Nov 25 11:43:15 crc kubenswrapper[4821]: I1125 11:43:15.117695 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5d8zv" event={"ID":"5c6f0100-2dec-45c5-95e4-229b499d8d7f","Type":"ContainerDied","Data":"788f950095ec9f86466843e0f119d29c31392508319701233925306ac58877ca"} Nov 25 11:43:15 crc kubenswrapper[4821]: I1125 11:43:15.120155 4821 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 11:43:17 crc kubenswrapper[4821]: I1125 11:43:17.140671 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5d8zv" event={"ID":"5c6f0100-2dec-45c5-95e4-229b499d8d7f","Type":"ContainerStarted","Data":"74b87bed043ebfc35a1be0af710ca469f5665f97a15e7052481eadc20277ffa3"} Nov 25 11:43:19 crc kubenswrapper[4821]: I1125 11:43:19.165326 4821 generic.go:334] "Generic (PLEG): container finished" podID="5c6f0100-2dec-45c5-95e4-229b499d8d7f" containerID="74b87bed043ebfc35a1be0af710ca469f5665f97a15e7052481eadc20277ffa3" exitCode=0 Nov 25 11:43:19 crc kubenswrapper[4821]: I1125 11:43:19.165400 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5d8zv" event={"ID":"5c6f0100-2dec-45c5-95e4-229b499d8d7f","Type":"ContainerDied","Data":"74b87bed043ebfc35a1be0af710ca469f5665f97a15e7052481eadc20277ffa3"} Nov 25 11:43:20 crc kubenswrapper[4821]: I1125 11:43:20.198629 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5d8zv" event={"ID":"5c6f0100-2dec-45c5-95e4-229b499d8d7f","Type":"ContainerStarted","Data":"6531262721fafd583de42eba831a2206e88918cef7c8550397b317e3a7ecb33e"} Nov 25 11:43:20 crc kubenswrapper[4821]: I1125 11:43:20.231612 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5d8zv" podStartSLOduration=3.6356290639999997 podStartE2EDuration="8.231579396s" podCreationTimestamp="2025-11-25 11:43:12 +0000 UTC" firstStartedPulling="2025-11-25 11:43:15.119963574 +0000 UTC m=+4265.656283421" lastFinishedPulling="2025-11-25 11:43:19.715913906 +0000 UTC m=+4270.252233753" observedRunningTime="2025-11-25 11:43:20.225340863 +0000 UTC m=+4270.761660710" watchObservedRunningTime="2025-11-25 11:43:20.231579396 +0000 UTC m=+4270.767899243" Nov 25 11:43:23 crc kubenswrapper[4821]: I1125 11:43:23.266663 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5d8zv" Nov 25 11:43:23 crc kubenswrapper[4821]: I1125 11:43:23.267194 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5d8zv" Nov 25 11:43:23 crc kubenswrapper[4821]: I1125 11:43:23.327131 4821 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5d8zv" Nov 25 11:43:24 crc kubenswrapper[4821]: I1125 11:43:24.299336 4821 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5d8zv" Nov 25 11:43:24 crc kubenswrapper[4821]: I1125 11:43:24.362371 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5d8zv"] Nov 25 11:43:26 crc kubenswrapper[4821]: I1125 11:43:26.252574 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-5d8zv" podUID="5c6f0100-2dec-45c5-95e4-229b499d8d7f" containerName="registry-server" containerID="cri-o://6531262721fafd583de42eba831a2206e88918cef7c8550397b317e3a7ecb33e" gracePeriod=2 Nov 25 11:43:26 crc kubenswrapper[4821]: I1125 11:43:26.692812 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5d8zv" Nov 25 11:43:26 crc kubenswrapper[4821]: I1125 11:43:26.812280 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n2nqn\" (UniqueName: \"kubernetes.io/projected/5c6f0100-2dec-45c5-95e4-229b499d8d7f-kube-api-access-n2nqn\") pod \"5c6f0100-2dec-45c5-95e4-229b499d8d7f\" (UID: \"5c6f0100-2dec-45c5-95e4-229b499d8d7f\") " Nov 25 11:43:26 crc kubenswrapper[4821]: I1125 11:43:26.812576 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c6f0100-2dec-45c5-95e4-229b499d8d7f-utilities\") pod \"5c6f0100-2dec-45c5-95e4-229b499d8d7f\" (UID: \"5c6f0100-2dec-45c5-95e4-229b499d8d7f\") " Nov 25 11:43:26 crc kubenswrapper[4821]: I1125 11:43:26.812657 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c6f0100-2dec-45c5-95e4-229b499d8d7f-catalog-content\") pod \"5c6f0100-2dec-45c5-95e4-229b499d8d7f\" (UID: \"5c6f0100-2dec-45c5-95e4-229b499d8d7f\") " Nov 25 11:43:26 crc kubenswrapper[4821]: I1125 11:43:26.814333 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c6f0100-2dec-45c5-95e4-229b499d8d7f-utilities" (OuterVolumeSpecName: "utilities") pod "5c6f0100-2dec-45c5-95e4-229b499d8d7f" (UID: "5c6f0100-2dec-45c5-95e4-229b499d8d7f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:43:26 crc kubenswrapper[4821]: I1125 11:43:26.820323 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c6f0100-2dec-45c5-95e4-229b499d8d7f-kube-api-access-n2nqn" (OuterVolumeSpecName: "kube-api-access-n2nqn") pod "5c6f0100-2dec-45c5-95e4-229b499d8d7f" (UID: "5c6f0100-2dec-45c5-95e4-229b499d8d7f"). InnerVolumeSpecName "kube-api-access-n2nqn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:43:26 crc kubenswrapper[4821]: I1125 11:43:26.872192 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c6f0100-2dec-45c5-95e4-229b499d8d7f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5c6f0100-2dec-45c5-95e4-229b499d8d7f" (UID: "5c6f0100-2dec-45c5-95e4-229b499d8d7f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:43:26 crc kubenswrapper[4821]: I1125 11:43:26.915352 4821 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c6f0100-2dec-45c5-95e4-229b499d8d7f-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 11:43:26 crc kubenswrapper[4821]: I1125 11:43:26.915395 4821 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c6f0100-2dec-45c5-95e4-229b499d8d7f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 11:43:26 crc kubenswrapper[4821]: I1125 11:43:26.915409 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n2nqn\" (UniqueName: \"kubernetes.io/projected/5c6f0100-2dec-45c5-95e4-229b499d8d7f-kube-api-access-n2nqn\") on node \"crc\" DevicePath \"\"" Nov 25 11:43:27 crc kubenswrapper[4821]: I1125 11:43:27.263804 4821 generic.go:334] "Generic (PLEG): container finished" podID="5c6f0100-2dec-45c5-95e4-229b499d8d7f" containerID="6531262721fafd583de42eba831a2206e88918cef7c8550397b317e3a7ecb33e" exitCode=0 Nov 25 11:43:27 crc kubenswrapper[4821]: I1125 11:43:27.263849 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5d8zv" event={"ID":"5c6f0100-2dec-45c5-95e4-229b499d8d7f","Type":"ContainerDied","Data":"6531262721fafd583de42eba831a2206e88918cef7c8550397b317e3a7ecb33e"} Nov 25 11:43:27 crc kubenswrapper[4821]: I1125 11:43:27.263873 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5d8zv" event={"ID":"5c6f0100-2dec-45c5-95e4-229b499d8d7f","Type":"ContainerDied","Data":"f9144ee5813a9b07d8298e4c1defe9186645343a7779e23beeffba33d5ebfab9"} Nov 25 11:43:27 crc kubenswrapper[4821]: I1125 11:43:27.263890 4821 scope.go:117] "RemoveContainer" containerID="6531262721fafd583de42eba831a2206e88918cef7c8550397b317e3a7ecb33e" Nov 25 11:43:27 crc kubenswrapper[4821]: I1125 11:43:27.264021 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5d8zv" Nov 25 11:43:27 crc kubenswrapper[4821]: I1125 11:43:27.284826 4821 scope.go:117] "RemoveContainer" containerID="74b87bed043ebfc35a1be0af710ca469f5665f97a15e7052481eadc20277ffa3" Nov 25 11:43:27 crc kubenswrapper[4821]: I1125 11:43:27.302626 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5d8zv"] Nov 25 11:43:27 crc kubenswrapper[4821]: I1125 11:43:27.312445 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-5d8zv"] Nov 25 11:43:27 crc kubenswrapper[4821]: I1125 11:43:27.319592 4821 scope.go:117] "RemoveContainer" containerID="788f950095ec9f86466843e0f119d29c31392508319701233925306ac58877ca" Nov 25 11:43:27 crc kubenswrapper[4821]: I1125 11:43:27.356753 4821 scope.go:117] "RemoveContainer" containerID="6531262721fafd583de42eba831a2206e88918cef7c8550397b317e3a7ecb33e" Nov 25 11:43:27 crc kubenswrapper[4821]: E1125 11:43:27.357331 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6531262721fafd583de42eba831a2206e88918cef7c8550397b317e3a7ecb33e\": container with ID starting with 6531262721fafd583de42eba831a2206e88918cef7c8550397b317e3a7ecb33e not found: ID does not exist" containerID="6531262721fafd583de42eba831a2206e88918cef7c8550397b317e3a7ecb33e" Nov 25 11:43:27 crc kubenswrapper[4821]: I1125 11:43:27.357381 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6531262721fafd583de42eba831a2206e88918cef7c8550397b317e3a7ecb33e"} err="failed to get container status \"6531262721fafd583de42eba831a2206e88918cef7c8550397b317e3a7ecb33e\": rpc error: code = NotFound desc = could not find container \"6531262721fafd583de42eba831a2206e88918cef7c8550397b317e3a7ecb33e\": container with ID starting with 6531262721fafd583de42eba831a2206e88918cef7c8550397b317e3a7ecb33e not found: ID does not exist" Nov 25 11:43:27 crc kubenswrapper[4821]: I1125 11:43:27.357416 4821 scope.go:117] "RemoveContainer" containerID="74b87bed043ebfc35a1be0af710ca469f5665f97a15e7052481eadc20277ffa3" Nov 25 11:43:27 crc kubenswrapper[4821]: E1125 11:43:27.357785 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"74b87bed043ebfc35a1be0af710ca469f5665f97a15e7052481eadc20277ffa3\": container with ID starting with 74b87bed043ebfc35a1be0af710ca469f5665f97a15e7052481eadc20277ffa3 not found: ID does not exist" containerID="74b87bed043ebfc35a1be0af710ca469f5665f97a15e7052481eadc20277ffa3" Nov 25 11:43:27 crc kubenswrapper[4821]: I1125 11:43:27.357804 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74b87bed043ebfc35a1be0af710ca469f5665f97a15e7052481eadc20277ffa3"} err="failed to get container status \"74b87bed043ebfc35a1be0af710ca469f5665f97a15e7052481eadc20277ffa3\": rpc error: code = NotFound desc = could not find container \"74b87bed043ebfc35a1be0af710ca469f5665f97a15e7052481eadc20277ffa3\": container with ID starting with 74b87bed043ebfc35a1be0af710ca469f5665f97a15e7052481eadc20277ffa3 not found: ID does not exist" Nov 25 11:43:27 crc kubenswrapper[4821]: I1125 11:43:27.357820 4821 scope.go:117] "RemoveContainer" containerID="788f950095ec9f86466843e0f119d29c31392508319701233925306ac58877ca" Nov 25 11:43:27 crc kubenswrapper[4821]: E1125 11:43:27.358083 4821 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"788f950095ec9f86466843e0f119d29c31392508319701233925306ac58877ca\": container with ID starting with 788f950095ec9f86466843e0f119d29c31392508319701233925306ac58877ca not found: ID does not exist" containerID="788f950095ec9f86466843e0f119d29c31392508319701233925306ac58877ca" Nov 25 11:43:27 crc kubenswrapper[4821]: I1125 11:43:27.358103 4821 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"788f950095ec9f86466843e0f119d29c31392508319701233925306ac58877ca"} err="failed to get container status \"788f950095ec9f86466843e0f119d29c31392508319701233925306ac58877ca\": rpc error: code = NotFound desc = could not find container \"788f950095ec9f86466843e0f119d29c31392508319701233925306ac58877ca\": container with ID starting with 788f950095ec9f86466843e0f119d29c31392508319701233925306ac58877ca not found: ID does not exist" Nov 25 11:43:28 crc kubenswrapper[4821]: I1125 11:43:28.128063 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c6f0100-2dec-45c5-95e4-229b499d8d7f" path="/var/lib/kubelet/pods/5c6f0100-2dec-45c5-95e4-229b499d8d7f/volumes" Nov 25 11:43:34 crc kubenswrapper[4821]: I1125 11:43:34.715346 4821 patch_prober.go:28] interesting pod/machine-config-daemon-2krbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:43:34 crc kubenswrapper[4821]: I1125 11:43:34.717269 4821 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:43:34 crc kubenswrapper[4821]: I1125 11:43:34.717394 4821 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" Nov 25 11:43:34 crc kubenswrapper[4821]: I1125 11:43:34.718260 4821 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b90a0504e63aa3c9d24899ca1bb330848b4b0bc00ba7e2363f96654d9d7aca46"} pod="openshift-machine-config-operator/machine-config-daemon-2krbf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 11:43:34 crc kubenswrapper[4821]: I1125 11:43:34.718409 4821 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerName="machine-config-daemon" containerID="cri-o://b90a0504e63aa3c9d24899ca1bb330848b4b0bc00ba7e2363f96654d9d7aca46" gracePeriod=600 Nov 25 11:43:35 crc kubenswrapper[4821]: I1125 11:43:35.336049 4821 generic.go:334] "Generic (PLEG): container finished" podID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerID="b90a0504e63aa3c9d24899ca1bb330848b4b0bc00ba7e2363f96654d9d7aca46" exitCode=0 Nov 25 11:43:35 crc kubenswrapper[4821]: I1125 11:43:35.336119 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" event={"ID":"5f948b87-ac86-4de6-ad64-c2ef947f84d4","Type":"ContainerDied","Data":"b90a0504e63aa3c9d24899ca1bb330848b4b0bc00ba7e2363f96654d9d7aca46"} Nov 25 11:43:35 crc kubenswrapper[4821]: I1125 11:43:35.336417 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" event={"ID":"5f948b87-ac86-4de6-ad64-c2ef947f84d4","Type":"ContainerStarted","Data":"0fc1b943bb602130a92a3b0d44e272cfc3f53ea63aab63a8b7b42a4de0c85c13"} Nov 25 11:43:35 crc kubenswrapper[4821]: I1125 11:43:35.336443 4821 scope.go:117] "RemoveContainer" containerID="f357cf6563c37c0c7cd86102a3e1eea3ba9146a525631fc115bab1d95f4015f3" Nov 25 11:45:00 crc kubenswrapper[4821]: I1125 11:45:00.170541 4821 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401185-48pk5"] Nov 25 11:45:00 crc kubenswrapper[4821]: E1125 11:45:00.172803 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c6f0100-2dec-45c5-95e4-229b499d8d7f" containerName="extract-content" Nov 25 11:45:00 crc kubenswrapper[4821]: I1125 11:45:00.172833 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c6f0100-2dec-45c5-95e4-229b499d8d7f" containerName="extract-content" Nov 25 11:45:00 crc kubenswrapper[4821]: E1125 11:45:00.172884 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c6f0100-2dec-45c5-95e4-229b499d8d7f" containerName="registry-server" Nov 25 11:45:00 crc kubenswrapper[4821]: I1125 11:45:00.172894 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c6f0100-2dec-45c5-95e4-229b499d8d7f" containerName="registry-server" Nov 25 11:45:00 crc kubenswrapper[4821]: E1125 11:45:00.172931 4821 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c6f0100-2dec-45c5-95e4-229b499d8d7f" containerName="extract-utilities" Nov 25 11:45:00 crc kubenswrapper[4821]: I1125 11:45:00.172940 4821 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c6f0100-2dec-45c5-95e4-229b499d8d7f" containerName="extract-utilities" Nov 25 11:45:00 crc kubenswrapper[4821]: I1125 11:45:00.173442 4821 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c6f0100-2dec-45c5-95e4-229b499d8d7f" containerName="registry-server" Nov 25 11:45:00 crc kubenswrapper[4821]: I1125 11:45:00.174510 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401185-48pk5" Nov 25 11:45:00 crc kubenswrapper[4821]: I1125 11:45:00.184853 4821 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 25 11:45:00 crc kubenswrapper[4821]: I1125 11:45:00.184909 4821 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 25 11:45:00 crc kubenswrapper[4821]: I1125 11:45:00.185334 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/257fbc9c-5c96-4feb-bd63-b490a99c655a-config-volume\") pod \"collect-profiles-29401185-48pk5\" (UID: \"257fbc9c-5c96-4feb-bd63-b490a99c655a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401185-48pk5" Nov 25 11:45:00 crc kubenswrapper[4821]: I1125 11:45:00.185500 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qt22t\" (UniqueName: \"kubernetes.io/projected/257fbc9c-5c96-4feb-bd63-b490a99c655a-kube-api-access-qt22t\") pod \"collect-profiles-29401185-48pk5\" (UID: \"257fbc9c-5c96-4feb-bd63-b490a99c655a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401185-48pk5" Nov 25 11:45:00 crc kubenswrapper[4821]: I1125 11:45:00.185576 4821 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/257fbc9c-5c96-4feb-bd63-b490a99c655a-secret-volume\") pod \"collect-profiles-29401185-48pk5\" (UID: \"257fbc9c-5c96-4feb-bd63-b490a99c655a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401185-48pk5" Nov 25 11:45:00 crc kubenswrapper[4821]: I1125 11:45:00.187449 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401185-48pk5"] Nov 25 11:45:00 crc kubenswrapper[4821]: I1125 11:45:00.287591 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qt22t\" (UniqueName: \"kubernetes.io/projected/257fbc9c-5c96-4feb-bd63-b490a99c655a-kube-api-access-qt22t\") pod \"collect-profiles-29401185-48pk5\" (UID: \"257fbc9c-5c96-4feb-bd63-b490a99c655a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401185-48pk5" Nov 25 11:45:00 crc kubenswrapper[4821]: I1125 11:45:00.287649 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/257fbc9c-5c96-4feb-bd63-b490a99c655a-secret-volume\") pod \"collect-profiles-29401185-48pk5\" (UID: \"257fbc9c-5c96-4feb-bd63-b490a99c655a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401185-48pk5" Nov 25 11:45:00 crc kubenswrapper[4821]: I1125 11:45:00.287759 4821 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/257fbc9c-5c96-4feb-bd63-b490a99c655a-config-volume\") pod \"collect-profiles-29401185-48pk5\" (UID: \"257fbc9c-5c96-4feb-bd63-b490a99c655a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401185-48pk5" Nov 25 11:45:00 crc kubenswrapper[4821]: I1125 11:45:00.290138 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/257fbc9c-5c96-4feb-bd63-b490a99c655a-config-volume\") pod \"collect-profiles-29401185-48pk5\" (UID: \"257fbc9c-5c96-4feb-bd63-b490a99c655a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401185-48pk5" Nov 25 11:45:00 crc kubenswrapper[4821]: I1125 11:45:00.296366 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/257fbc9c-5c96-4feb-bd63-b490a99c655a-secret-volume\") pod \"collect-profiles-29401185-48pk5\" (UID: \"257fbc9c-5c96-4feb-bd63-b490a99c655a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401185-48pk5" Nov 25 11:45:00 crc kubenswrapper[4821]: I1125 11:45:00.306146 4821 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qt22t\" (UniqueName: \"kubernetes.io/projected/257fbc9c-5c96-4feb-bd63-b490a99c655a-kube-api-access-qt22t\") pod \"collect-profiles-29401185-48pk5\" (UID: \"257fbc9c-5c96-4feb-bd63-b490a99c655a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401185-48pk5" Nov 25 11:45:00 crc kubenswrapper[4821]: I1125 11:45:00.511725 4821 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401185-48pk5" Nov 25 11:45:00 crc kubenswrapper[4821]: I1125 11:45:00.969095 4821 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401185-48pk5"] Nov 25 11:45:01 crc kubenswrapper[4821]: I1125 11:45:01.194427 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401185-48pk5" event={"ID":"257fbc9c-5c96-4feb-bd63-b490a99c655a","Type":"ContainerStarted","Data":"e53c04e1d3dbecb214a8b83442d51db2ad4ae6b5f0455c9e81177cfc33b5aab0"} Nov 25 11:45:01 crc kubenswrapper[4821]: I1125 11:45:01.194511 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401185-48pk5" event={"ID":"257fbc9c-5c96-4feb-bd63-b490a99c655a","Type":"ContainerStarted","Data":"73791b6b893245d401aaadd0541043b8f688cc493abd7fe4249bd27d254b83af"} Nov 25 11:45:01 crc kubenswrapper[4821]: I1125 11:45:01.219308 4821 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29401185-48pk5" podStartSLOduration=1.219287164 podStartE2EDuration="1.219287164s" podCreationTimestamp="2025-11-25 11:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:45:01.213621716 +0000 UTC m=+4371.749941563" watchObservedRunningTime="2025-11-25 11:45:01.219287164 +0000 UTC m=+4371.755607011" Nov 25 11:45:02 crc kubenswrapper[4821]: I1125 11:45:02.204757 4821 generic.go:334] "Generic (PLEG): container finished" podID="257fbc9c-5c96-4feb-bd63-b490a99c655a" containerID="e53c04e1d3dbecb214a8b83442d51db2ad4ae6b5f0455c9e81177cfc33b5aab0" exitCode=0 Nov 25 11:45:02 crc kubenswrapper[4821]: I1125 11:45:02.204882 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401185-48pk5" event={"ID":"257fbc9c-5c96-4feb-bd63-b490a99c655a","Type":"ContainerDied","Data":"e53c04e1d3dbecb214a8b83442d51db2ad4ae6b5f0455c9e81177cfc33b5aab0"} Nov 25 11:45:03 crc kubenswrapper[4821]: I1125 11:45:03.637580 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401185-48pk5" Nov 25 11:45:03 crc kubenswrapper[4821]: I1125 11:45:03.656849 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/257fbc9c-5c96-4feb-bd63-b490a99c655a-secret-volume\") pod \"257fbc9c-5c96-4feb-bd63-b490a99c655a\" (UID: \"257fbc9c-5c96-4feb-bd63-b490a99c655a\") " Nov 25 11:45:03 crc kubenswrapper[4821]: I1125 11:45:03.657035 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/257fbc9c-5c96-4feb-bd63-b490a99c655a-config-volume\") pod \"257fbc9c-5c96-4feb-bd63-b490a99c655a\" (UID: \"257fbc9c-5c96-4feb-bd63-b490a99c655a\") " Nov 25 11:45:03 crc kubenswrapper[4821]: I1125 11:45:03.657059 4821 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qt22t\" (UniqueName: \"kubernetes.io/projected/257fbc9c-5c96-4feb-bd63-b490a99c655a-kube-api-access-qt22t\") pod \"257fbc9c-5c96-4feb-bd63-b490a99c655a\" (UID: \"257fbc9c-5c96-4feb-bd63-b490a99c655a\") " Nov 25 11:45:03 crc kubenswrapper[4821]: I1125 11:45:03.657711 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/257fbc9c-5c96-4feb-bd63-b490a99c655a-config-volume" (OuterVolumeSpecName: "config-volume") pod "257fbc9c-5c96-4feb-bd63-b490a99c655a" (UID: "257fbc9c-5c96-4feb-bd63-b490a99c655a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:45:03 crc kubenswrapper[4821]: I1125 11:45:03.664437 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/257fbc9c-5c96-4feb-bd63-b490a99c655a-kube-api-access-qt22t" (OuterVolumeSpecName: "kube-api-access-qt22t") pod "257fbc9c-5c96-4feb-bd63-b490a99c655a" (UID: "257fbc9c-5c96-4feb-bd63-b490a99c655a"). InnerVolumeSpecName "kube-api-access-qt22t". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:45:03 crc kubenswrapper[4821]: I1125 11:45:03.666364 4821 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/257fbc9c-5c96-4feb-bd63-b490a99c655a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "257fbc9c-5c96-4feb-bd63-b490a99c655a" (UID: "257fbc9c-5c96-4feb-bd63-b490a99c655a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:45:03 crc kubenswrapper[4821]: I1125 11:45:03.758884 4821 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/257fbc9c-5c96-4feb-bd63-b490a99c655a-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 25 11:45:03 crc kubenswrapper[4821]: I1125 11:45:03.758922 4821 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/257fbc9c-5c96-4feb-bd63-b490a99c655a-config-volume\") on node \"crc\" DevicePath \"\"" Nov 25 11:45:03 crc kubenswrapper[4821]: I1125 11:45:03.758934 4821 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qt22t\" (UniqueName: \"kubernetes.io/projected/257fbc9c-5c96-4feb-bd63-b490a99c655a-kube-api-access-qt22t\") on node \"crc\" DevicePath \"\"" Nov 25 11:45:04 crc kubenswrapper[4821]: I1125 11:45:04.221697 4821 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401185-48pk5" event={"ID":"257fbc9c-5c96-4feb-bd63-b490a99c655a","Type":"ContainerDied","Data":"73791b6b893245d401aaadd0541043b8f688cc493abd7fe4249bd27d254b83af"} Nov 25 11:45:04 crc kubenswrapper[4821]: I1125 11:45:04.221735 4821 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="73791b6b893245d401aaadd0541043b8f688cc493abd7fe4249bd27d254b83af" Nov 25 11:45:04 crc kubenswrapper[4821]: I1125 11:45:04.221785 4821 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401185-48pk5" Nov 25 11:45:04 crc kubenswrapper[4821]: I1125 11:45:04.285680 4821 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401140-cqkws"] Nov 25 11:45:04 crc kubenswrapper[4821]: I1125 11:45:04.296455 4821 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401140-cqkws"] Nov 25 11:45:06 crc kubenswrapper[4821]: I1125 11:45:06.130232 4821 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d1e4062-40fb-4409-823a-3e1b5023dea0" path="/var/lib/kubelet/pods/6d1e4062-40fb-4409-823a-3e1b5023dea0/volumes" Nov 25 11:45:26 crc kubenswrapper[4821]: I1125 11:45:26.438873 4821 scope.go:117] "RemoveContainer" containerID="68356fa839f71c30317b5cc58463ef2ee7d91f459aee6a8ebe563974d0d98aab" Nov 25 11:46:04 crc kubenswrapper[4821]: I1125 11:46:04.714205 4821 patch_prober.go:28] interesting pod/machine-config-daemon-2krbf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:46:04 crc kubenswrapper[4821]: I1125 11:46:04.714774 4821 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2krbf" podUID="5f948b87-ac86-4de6-ad64-c2ef947f84d4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515111313407024441 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015111313410017350 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015111302303016472 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015111302303015442 5ustar corecore